|
12 | 12 | [](https://huggingface.co/papers/2601.11655) |
13 | 13 | [](https://deepsoftwareanalytics.github.io/Awesome-Issue-Resolution/tables/) |
14 | 14 | [](https://github.com/DeepSoftwareAnalytics/Awesome-Issue-Resolution/graphs/contributors) |
15 | | - |
| 15 | + |
16 | 16 |
|
17 | 17 | [**📖 Documentation Website**](https://deepsoftwareanalytics.github.io/Awesome-Issue-Resolution/) | [**📄 Full Paper**](https://deepsoftwareanalytics.github.io/Awesome-Issue-Resolution/paper/) | [**📋 Tables & Resources**](https://deepsoftwareanalytics.github.io/Awesome-Issue-Resolution/tables/) |
18 | 18 |
|
|
32 | 32 |
|
33 | 33 | ## 📖 Abstract |
34 | 34 |
|
35 | | -Based on a systematic review of **183 papers and online resources**, this survey establishes a holistic theoretical framework for Issue Resolution in software engineering. We examine how **Large Language Models (LLMs)** are transforming the automation of GitHub issue resolution. Beyond the theoretical analysis, we have curated a comprehensive collection of datasets and model training resources, which are continuously synchronized with our GitHub repository and project documentation website. |
| 35 | +Based on a systematic review of **186 papers and online resources**, this survey establishes a holistic theoretical framework for Issue Resolution in software engineering. We examine how **Large Language Models (LLMs)** are transforming the automation of GitHub issue resolution. Beyond the theoretical analysis, we have curated a comprehensive collection of datasets and model training resources, which are continuously synchronized with our GitHub repository and project documentation website. |
36 | 36 |
|
37 | 37 | <!-- START EXPLORE --> |
38 | 38 | **🔍 Explore This Survey:** |
@@ -67,7 +67,7 @@ Based on a systematic review of **183 papers and online resources**, this survey |
67 | 67 | ## 📚 Complete Paper List |
68 | 68 |
|
69 | 69 |
|
70 | | -> **Total: 183 works** across 14 categories |
| 70 | +> **Total: 186 works** across 14 categories |
71 | 71 |
|
72 | 72 |
|
73 | 73 | ### 📊 Evaluation Datasets |
@@ -98,6 +98,7 @@ Based on a systematic review of **183 papers and online resources**, this survey |
98 | 98 | - **SWE-fficiency**: SWE-fficiency: Can Language Models Optimize Real-World Repositories on Real Workloads? (2025) [](https://arxiv.org/abs/2511.06090) |
99 | 99 | - **SWE-Compass**: SWE-Compass: Towards Unified Evaluation of Agentic Coding Abilities for Large Language Models (2025) [](https://arxiv.org/abs/2511.05459) |
100 | 100 | - **SWE-EVO**: SWE-EVO: Benchmarking Coding Agents in Long-Horizon Software Evolution Scenarios (2025) [](https://arxiv.org/abs/2512.18470) |
| 101 | +- **SWE Context Bench**: SWE Context Bench: A Benchmark for Context Learning in Coding (2026) [](https://arxiv.org/pdf/2602.08316) |
101 | 102 |
|
102 | 103 | ### 🎯 Training Datasets |
103 | 104 |
|
@@ -251,6 +252,7 @@ Based on a systematic review of **183 papers and online resources**, this survey |
251 | 252 | - **SWE-Lego**: SWE-Lego: Pushing the Limits of Supervised Fine-tuning for Software Issue Resolving (2026) [](https://arxiv.org/abs/2601.01426) |
252 | 253 | - **Agentic Rubrics**: Agentic Rubrics as Contextual Verifiers for SWE Agents (2026) [](https://arxiv.org/abs/2601.04171) |
253 | 254 | - **CGM**: Code Graph Model (CGM): A Graph-Integrated Large Language Model for Repository-Level Software Engineering Tasks (2025) [](https://arxiv.org/abs/2505.16901) [](https://github.com/codefuse-ai/CodeFuse-CGM) [](https://huggingface.co/codefuse-ai/CodeFuse-CGM-72B) |
| 255 | +- **SWE-Replay**: SWE-Replay: Efficient Test-Time Scaling for Software Engineering Agents (2026) [](https://arxiv.org/abs/2601.22129) |
254 | 256 |
|
255 | 257 | ### 🎮 Reinforcement Learning (RL) |
256 | 258 |
|
@@ -293,6 +295,7 @@ Based on a systematic review of **183 papers and online resources**, this survey |
293 | 295 | - **LongCat-Flash-Think**: Introducing LongCat-Flash-Thinking: A Technical Report (2025) [](https://arxiv.org/abs/2509.18883) |
294 | 296 | - **MiMo-V2-Flash**: MiMo-V2-Flash Technical Report (2026) [](https://arxiv.org/abs/2601.02780) |
295 | 297 | - **SWE-Master**: SWE-Master: Unleashing the Potential of Software Engineering Agents via Post-Training (2026) [](https://arxiv.org/abs/2602.03411) [](https://github.com/RUCAIBox/SWE-Master) |
| 298 | +- **SWE-Protégé**: SWE-Protégé: Learning to Selectively Collaborate With an Expert Unlocks Small Language Models as Software Engineering Agents (2026) [](https://arxiv.org/abs/2602.22124) |
296 | 299 |
|
297 | 300 | ### ⚡ Inference-Time Scaling |
298 | 301 |
|
|
0 commit comments