Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
Awesome-Code-LLM
👨💻 An awesome and curated list of best code-LLM for research.
https://github.com/huybery/Awesome-Code-LLM
Last synced: 3 days ago
JSON representation
-
🚀 Leaderboard
-
📚 Paper
-
▶️ Instruction Tuning
- [Paper
- [Paper - project/octopack)] *Niklas Muennighoff, Qian Liu, Armel Zebaze, Qinkai Zheng, Binyuan Hui, Terry Yue Zhuo, Swayam Singh, Xiangru Tang, Leandro von Werra, Shayne Longpre.* 2023.08
- [Paper
- [Repo
- [Paper - uiuc/magicoder)] *Yuxiang Wei, Zhe Wang, Jiawei Liu, Yifeng Ding, Lingming Zhang* 2023.12
- [Paper
- [Paper - project/octopack)] *Niklas Muennighoff, Qian Liu, Armel Zebaze, Qinkai Zheng, Binyuan Hui, Terry Yue Zhuo, Swayam Singh, Xiangru Tang, Leandro von Werra, Shayne Longpre.* 2023.08
-
▶️ Pre-Training
-
▶️ Alignment with Feedback
-
▶️ Prompting
- [Paper
- [Paper - Guang Lou, Weizhu Chen.* 2022.07
- [Paper - tau Yih, Daniel Fried, Sida I Wang.* 2022.11
- [Paper - tau Yih, Sida I. Wang, Xi Victoria Lin.* 2023.02
- [Paper - Guang Lou, Weizhu Chen.* 2022.07
- [Paper - tau Yih, Daniel Fried, Sida I Wang.* 2022.11
- [Paper - tau Yih, Sida I. Wang, Xi Victoria Lin.* 2023.02
- [Paper - Lezama.* 2023.06
- [Paper
- [Paper - Lezama.* 2023.06
- [Paper
- [Paper - Guang Lou, Weizhu Chen.* 2022.07
- [Paper - tau Yih, Daniel Fried, Sida I Wang.* 2022.11
- [Paper - tau Yih, Sida I. Wang, Xi Victoria Lin.* 2023.02
-
▶️ Evaluation & Benchmark
- [Paper
- [Paper - tau Yih, Daniel Fried, Sida Wang, Tao Yu.* 2022.11
- [Paper
- [Paper
- [Paper
- [Paper
- [Paper - Guang Lou, Weizhu Chen.* 2023.10
- [Paper
- [Paper
- [Paper - compass/DevBench)] *Bowen Li, Wenhan Wu, Ziwei Tang, Lin Shi, John Yang, Jinyang Li, Shunyu Yao, Chen Qian, Binyuan Hui, Qicheng Zhang, Zhiyin Yu, He Du, Ping Yang, Dahua Lin, Chao Peng, Kai Chen* 2024.3
- [Paper
- [Paper
- [Paper
- [Paper
- [Paper - tau Yih, Daniel Fried, Sida Wang, Tao Yu.* 2022.11
- [Paper
- [Paper
-
▶️ Using LLMs while coding
-
-
🚀 Awesome Code LLMs Leaderboard
- CanAiCode Leaderboard
- CRUXEval Leaderboard
- Program Synthesis Models Leaderboard - source model with an intuitive leadership quadrant graph. They evaluate the performance of open-source code models to rank them based on their capabilities and market adoption. |
- EvalPlus Leaderboard
- InfiBench Leaderboard - world questions in the code domain. |
- Evalperf Leaderboard
- Aider Code Editing Leaderboard
- LiveCodeBench Leaderboard
- Spider Leaderboard - scale complex and cross-domain semantic parsing and text-to-SQL dataset annotated by 11 Yale students. The goal of the Spider challenge is to develop natural language interfaces to cross-domain databases. |
- Big Code Models Leaderboard - E. |
-
Star History
-
▶️ Using LLMs while coding
- ![Star History Chart - history.com/#huybery/Awesome-Code-LLM&Date)
- ![Star History Chart - history.com/#huybery/Awesome-Code-LLM&Date)
-
-
🚀 Top Code LLMs
- paper
- github
- WizardCoder-15B
- paper
- Qwen2.5-Coder-32B-Instruct - Coder) |
- Qwen2.5-Coder-14B-Instruct - Coder) |
- Qwen2.5-Coder-7B-Instruct - Coder) |
- DS-Coder-V2-Instruct - ai/DeepSeek-Coder-V2) |
- Qwen2.5-Coder-3B-Instruct - Coder) |
- DS-Coder-V2-Lite-Instruct - ai/DeepSeek-Coder-V2) |
- DeepSeek-Coder-6.7B-Instruct - ai/DeepSeek-Coder) |
- CodeLlama-70B-Instruct
- Qwen2.5-Coder-1.5B-Instruct - Coder) |
- StarCoder2-15B-Instruct-v0.1
- Qwen2.5-Coder-0.5B-Instruct - Coder) |
- CodeLlama-34B-Instruct
- CodeQwen1.5-7B-Chat
- CodeQwen1.5-7B
-
💡 Evaluation Toolkit:
-
📚 Awesome Code LLMs Papers
-
🐙 Awesome Code Benchmark & Evaluation Papers
- Star - Range Pre-trained Language Model for Code Completion**](https://arxiv.org/abs/2306.14893) <br> | `ICML'23` | `2023.10` | [Github](https://github.com/microsoft/CodeBERT) | - |
- Star - Level Code Completion Through Iterative Retrieval and Generation**](https://arxiv.org/abs/2306.03091) <br> | `EMNLP'23` | `2023.10` | [Github](https://github.com/microsoft/CodeT/tree/main/RepoCoder) | - |
- Star - Switching Capabilities of Code Generation Models**](https://arxiv.org/abs/2411.05830) <br> | `Preprint` | `2024.11` | [Github](https://github.com/NizarIslah/GitChameleon) | - |
- Star
- Star
- Star - compass/DevBench) | - |
- Star - File Code Completion**](https://arxiv.org/abs/2306.03091) <br> | `NeurIPS'23` | `2023.11` | [Github](https://github.com/amazon-science/cceval) | - |
- Star - E: A Scalable and Extensible Approach to Benchmarking Neural Code Generation**](https://arxiv.org/abs/2208.08227) <br> | `Preprint` | `2022.08` | [Github](https://github.com/nuprl/MultiPL-E) | [HF](https://huggingface.co/datasets/xlangai/DS-1000) |
- Star
- Star - bench: Can Language Models Resolve Real-World GitHub Issues?**](https://arxiv.org/abs/2310.06770) <br> | `ICLR'24` | `2024.03` | [Github](https://github.com/princeton-nlp/SWE-bench) | [HF](https://huggingface.co/datasets/princeton-nlp/SWE-bench) |
- Star - |
- Star - Level Code Auto-Completion Systems**](https://arxiv.org/abs/2306.03091) <br> | `ICLR'24` | `2023.06` | [Github](https://github.com/Leolty/repobench) | [HF](https://huggingface.co/datasets/tianyang/repobench_python_v1.1) |
- Star - round Code Auto-editing**](https://arxiv.org/abs/2305.18584) <br> | `ICLR'24` | `2023.05` | [Github](https://github.com/MrVPlusOne/Coeditor) | - |
- Star - 1000: A Natural and Reliable Benchmark for Data Science Code Generation**](https://arxiv.org/abs/2211.11501) <br> | `ICML'23` | `2022.11` | [Github](https://github.com/xlang-ai/DS-1000) | [HF](https://huggingface.co/datasets/xlangai/DS-1000) |
- **Can ChatGPT replace StackOverflow? A Study on Robustness and Reliability of Large Language Model Code Generation** - | - |
-
🐳 Awesome Code Instruction-Tuning Papers
- Star - uiuc/magicoder) | [HF](https://huggingface.co/ise-uiuc/Magicoder-DS-6.7B) |
- Star - project/octopack) | [HF](https://huggingface.co/bigcode/octocoder) |
- Star - Instruct**](https://arxiv.org/abs/2306.08568) <br> | `Preprint` | `2023.07` | [Github](https://github.com/nlpxucan/WizardLM) | [HF](https://huggingface.co/WizardLMTeam/WizardCoder-15B-V1.0) |
- Star - following LLaMA Model trained on code generation instructions**](https://github.com/sahil280114/codealpaca) <br> | `Preprint` | `2023.xx` | [Github](https://github.com/sahil280114/codealpaca) | [HF](https://huggingface.co/datasets/sahil2801/CodeAlpaca-20k) |
-
🌊 Awesome Code Pre-Training Papers
- Star - eval) | - |
- Star - Tier Code Large Language Models**](https://arxiv.org/abs/2411.04905) <br> | `Preprint` | `2024.11` | [Github](https://github.com/OpenCoder-llm/OpenCoder-llm) | [HF](https://huggingface.co/infly/OpenCoder-8B-Instruct) |
- Star - Coder Technical Report**](https://arxiv.org/abs/2409.12186) <br> | `Preprint` | `2024.09` | [Github](https://github.com/QwenLM/Qwen2.5-Coder) | [HF](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct) |
- Star - Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence**](https://arxiv.org/abs/2406.11931) <br> | `Preprint` | `2024.06` | [Github](https://github.com/deepseek-ai/DeepSeek-Coder-V2) | [HF](https://huggingface.co/deepseek-ai/DeepSeek-Coder-V2-Instruct) |
- Star - project/starcoder2) | [HF](https://huggingface.co/bigcode) |
- Star - Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence**](https://arxiv.org/abs/2401.14196) <br> | `Preprint` | `2024.01` | [Github](https://github.com/deepseek-ai/DeepSeek-Coder) | [HF](https://huggingface.co/deepseek-ai/deepseek-coder-33b-instruct) |
- Star - llama/codellama) | [HF](https://huggingface.co/meta-llama/CodeLlama-7b-hf) |
- **Textbooks Are All You Need** - | [HF](https://huggingface.co/microsoft/phi-1) |
- Star - 16b) |
- Star - project/starcoder) | [HF](https://huggingface.co/bigcode/starcoder) |
- Star - Turn Program Synthesis**](https://arxiv.org/abs/2203.13474) <br> | `ICLR'23` | `2022.03` | [Github](https://github.com/salesforce/CodeGen) | [HF](https://huggingface.co/Salesforce/codegen25-7b-multi_P) |
- Star - Trained Model for Code Generation with Multilingual Evaluations on HumanEval-X**](https://arxiv.org/abs/2303.17568) <br> | `Preprint` | `2023.03` | [Github](https://github.com/THUDM/CodeGeeX) | [HF](https://huggingface.co/collections/THUDM/codegeex4-6694e777e98246f00632fcf1) |
- **SantaCoder: don't reach for the stars!** - | [HF](https://huggingface.co/bigcode/santacoder) |
-
🐋 Awesome Code Prompting Papers
- Star - Repair for Code Generation**](https://arxiv.org/abs/2306.09896) <br> | `ICLR'24` | `2023.06` | [Github](https://github.com/theoxo/self-repair) | - |
- Star - |
- Star - by-step**](https://arxiv.org/abs/2402.16906) <br> | `ACL'24` | `2024.02` | [Github](https://github.com/FloridSleeves/LLMDebugger) | - |
- Star - to-Code Generation with Execution**](https://arxiv.org/abs/2302.08468) <br> | `ICML'23` | `2023.02` | [Github](https://github.com/niansong1996/lever) | - |
- Star - |
- **Teaching Large Language Models to Self-Debug** - | - |
- **SelfEvolve: A Code Evolution Framework via Large Language Models** - | - |
-
🐬 Awesome Code Alignment Papers
- **PLUM: Preference Learning Plus Test Cases Yields Better Code Language Models** - | - |
- Star - scut/RLTF) | - |
- Star - based Code Generation using Deep Reinforcement Learning**](https://arxiv.org/abs/2301.13816) <br> | `TMLR'23` | `2023.01` | [Github](https://github.com/reddy-lab-code-research/PPOCoder) | - |
- Star - |
- **PanGu-Coder2: Boosting Large Language Models for Code with Ranking Feedback** - | - |
- **ProSec: Fortifying Code LLMs with Proactive Security Alignment** - | -
-
-
News
- 2024-11-12 - Coder-32B-Instruct now the most powerful open-source code model.
- 2024-11-08
-
Acknowledgement
-
🐙 Awesome Code Benchmark & Evaluation Papers
-
Programming Languages
Categories
Sub Categories
▶️ Evaluation & Benchmark
17
🐙 Awesome Code Benchmark & Evaluation Papers
16
▶️ Prompting
14
🌊 Awesome Code Pre-Training Papers
13
▶️ Pre-Training
13
▶️ Alignment with Feedback
8
▶️ Instruction Tuning
7
🐋 Awesome Code Prompting Papers
7
🐬 Awesome Code Alignment Papers
6
▶️ Using LLMs while coding
5
🐳 Awesome Code Instruction-Tuning Papers
4
Keywords