CodeScout
RL-trained code search agents (1.7B, 4B, 14B) that outperform 2β18Γ larger models using only a Unix terminal. π arxiv.org/abs/2603.17829
Text Generation β’ 15B β’ Updated β’ 82 β’ 2Note π CodeScout-14B β strongest model, SOTA on SWE-Bench Verified/Pro/Lite
OpenHands/CodeScout-4B
Text Generation β’ 4B β’ Updated β’ 88Note β‘ CodeScout-4B β outperforms 8Γ larger Qwen3-32B across all benchmarks
OpenHands/CodeScout-1.7B
Text Generation β’ 2B β’ Updated β’ 83 β’ 1Note π¬ CodeScout-1.7B β post-RL checkpoint, outperforms 8Γ larger Qwen3-14B
OpenHands/CodeScout-1.7B-RFT
Text Generation β’ 2B β’ Updated β’ 81 β’ 1Note π¦ CodeScout-1.7B-RFT β pre-RL (rejection fine-tuned) checkpoint
OpenHands/CodeScout_Training_Rollouts
Viewer β’ Updated β’ 54.8k β’ 13 β’ 1Note ποΈ Training rollouts from SWE-Smith environments
OpenHands/CodeScout_Eval_Rollouts
Viewer β’ Updated β’ 12.7k β’ 12Note π Evaluation trajectories on SWE-Bench Verified, Pro, and Lite
OpenHands/SWE-smith-py-code-search
Viewer β’ Updated β’ 39.3k β’ 7Note π SWE-Smith code search localization targets
OpenHands/SWE-Gym-code-search
Viewer β’ Updated β’ 2.32k β’ 8Note π SWE-Gym code search localization targets
OpenHands/SWE-rebench-code-search
Viewer β’ Updated β’ 17.6k β’ 8Note π SWE-rebench code search localization targets
OpenHands/SWE-bench_Verified-locagent
Viewer β’ Updated β’ 500 β’ 11Note π― SWE-Bench Verified β localization ground truth
OpenHands/SWE-bench_Lite-locagent
Viewer β’ Updated β’ 300 β’ 8Note π― SWE-Bench Lite β localization ground truth
OpenHands/SWE-bench_Pro-locagent
Viewer β’ Updated β’ 264 β’ 8Note π― SWE-Bench Pro β localization ground truth