SkillAgentSearch skills...

CodeLLMSurvey

[TOSEM'25] The official GitHub page for the survey paper "A Survey on Large Language Models for Code Generation".

Install / Use

/learn @juyongjiang/CodeLLMSurvey
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

<div align="center"> <h1>:robot: A Survey on LLMs for Code Generation</h1> <a href="https://awesome.re"> <img src="https://awesome.re/badge.svg" alt="Awesome"> </a> <a href="https://img.shields.io/badge/PRs-Welcome-red"> <img src="https://img.shields.io/badge/PRs-Welcome-red" alt="PRs Welcome"> </a> <a href="https://img.shields.io/github/last-commit/juyongjiang/CodeLLMSurvey?color=green"> <img src="https://img.shields.io/github/last-commit/juyongjiang/CodeLLMSurvey?color=green" alt="Last Commit"> </a> </div> <!-- ![](assets/codellms.png) --> <p align="center" width="100%"> <img src="assets/codellms.png" alt="Code LLMs Workflow" style="width: 100%; min-width: 100px; display: block; margin: auto;"> <br> A diagram illustrating the general training, inference, and evaluation workflow for Code LLMs and their associated databases. The training workflow is mainly divided into four distinct stages: Stage ① and ② are the pre-training phase, whereas Stages ③ and ④ represent the post-training phases. It is important to note that Stage ② and ④ are optional. For instance, StarCoder incorporates only Stage ①. WizardCoder, fine-tuned upon StarCoder, includes only Stage ③, while Code Llama, continually pre-trained on Llama 2, encompasses Stages ② and ③. DeepSeek-Coder-V2, continually pre-trained on DeepSeek-V2, covers Stages ②, ③, and ④. Note that pre-trained model can be directly used for inference by prompt engineering. </p>

[!IMPORTANT] Good news! :tada: Our survey paper has been successfully accepted by TOSEM. :fire::fire::fire:

A curated collection of papers and resources on Large Language Models for Code Generation.

Please refer to our survey "A Survey on Large Language Models for Code Generation" for the detailed contents. Paper page

Please let us know if you discover any mistakes or have suggestions by emailing us: csjuyongjiang@gmail.com

If you find our survey beneficial for your research, please consider citing the following paper:

@article{jiang2024survey,
  title={A Survey on Large Language Models for Code Generation},
  author={Jiang, Juyong and Wang, Fan and Shen, Jiasi and Kim, Sungju and Kim, Sunghun},
  journal={arXiv preprint arXiv:2406.00515},
  year={2024}
}

Table of Contents

Paper

Pre-Training & Foundation Model

  • LLM-Based Test-Driven Interactive Code Generation: User Study and Empirical Evaluation, TSE, 2024
  • CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation, EMNLP, 2021
  • Bias Assessment and Mitigation in LLM-based Code Generation, arXiv, 2023
  • CodeScope: An Execution-based Multilingual Multitask Multidimensional Benchmark for Evaluating LLMs on Code Understanding and Generation, ACL, 2024
  • Synchromesh: Reliable code generation from pre-trained language models, ICLR, 2022
  • Automatic Code Generation using Pre-Trained Language Models, arXiv, 2021
  • CoderEval: {A} Benchmark of Pragmatic Code Generation with Generative Pre-trained Models, ICSE, 2024
  • ProphetNet-X: Large-Scale Pre-training Models for English, Chinese, Multi-lingual, Dialog, and Code Generation, ACL, 2021
  • ChatCoder: Chat-based Refine Requirement Improves LLMs' Code Generation, arXiv, 2023
  • Test-Driven Development and LLM-based Code Generation, ASE, 2024
  • Toward a New Era of Rapid Development: Assessing GPT-4-Vision's Capabilities in UML-Based Code Generation, LLM4CODE, 2024
  • When LLM-based Code Generation Meets the Software Development Process, arXiv, 2024
  • Learn to Code Sustainably: An Empirical Study on LLM-based Green Code Generation, arXiv, 2024
  • Bridging the Language Gap: Enhancing Multilingual Prompt-Based Code Generation in LLMs via Zero-Shot Cross-Lingual Transfer, arXiv, 2024
  • CodeGeeX: A Pre-Trained Model for Code Generation with Multilingual Benchmarking on HumanEval-X, KDD, 2023
  • Evaluating and Enhancing the Robustness of Code Pre-trained Models through Structure-Aware Adversarial Samples Generation, EMNLP, 2023
  • ClarifyGPT: Empowering LLM-based Code Generation with Intention Clarification, arXiv, 2023
  • Code Generation Tools (Almost) for Free? {A} Study of Few-Shot, Pre-Trained Language Models on Code, arXiv, 2022
  • Classification-Based Automatic {HDL} Code Generation Using LLMs, arXiv, 2024
  • Exploring the Effectiveness of {LLM} based Test-driven Interactive Code Generation: User Study and Empirical Evaluation, ICSE, 2024
  • Make Every Move Count: LLM-based High-Quality {RTL} Code Generation Using {MCTS}, arXiv, 2024
  • Towards more realistic evaluation of LLM-based code generation: an experimental study and beyond, arXiv, 2024
  • Game Agent Driven by Free-Form Text Command: Using LLM-based Code Generation and Behavior Branch, arXiv, 2024
  • Exploring the Effectiveness of LLM based Test-driven Interactive Code Generation: User Study and Empirical Evaluation, ICSE, 2024
  • LLM-based and Retrieval-Augmented Control Code Generation, LLM4Code, 2024
  • LLM-Based Code Generation Method for Golang Compiler Testing, FSE, 2023
  • CoderEval: A Benchmark of Pragmatic Code Generation with Generative Pre-trained Models, ICSE, 2024
  • LLM-based Control Code Generation using Image Recognition, LLM4Code, 2024
  • Benchmarking and Explaining Large Language Model-based Code Generation: {A} Causality-Centric Approach, arXiv, 2023
  • Code Llama: Open Foundation Models for Code, arXiv, 2023
  • Execution-based Code Generation using Deep Reinforcement Learning, TMLR, 2023
  • LongCoder: A Long-Range Pre-trained Language Model for Code Completion, ICML, 2023
  • Automating Code Review Activities by Large-Scale Pre-training, arXiv, 2022
  • Qwen2.5-Coder Technical Report, arXiv, 2024
  • DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence, arXiv, 2024
  • PanGu-Coder: Program Synthesis with Function-Level Language Modeling, arXiv, 2022
  • CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis, ICLR, 2023
  • Jigsaw: Large Language Models meet Program Synthesis, ICSE, 2022
  • ERNIE-Code: Beyond English-Centric Cross-lingual Pretraining for Programming Languages, ACL, 2022
  • SantaCoder: don't reach for the stars!, arXiv, 2023
  • CodeGen2: Lessons for Training LLMs on Programming and Natural Languages, ICLR, 2023
  • StarCoder: may the source be with you!, arXiv, 2023
  • Textbooks Are All You Need, arXiv, 2023
  • DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence, arXiv, 2024
  • StarCoder 2 and The Stack v2: The Next Generation, arXiv, 2024
  • CodeT5+: Open Code Large Language Models for Code Understanding and Generation, EMNLP, 2023

Instruction Tuning & Parameter-Efficient Fine-tuning

  • LLaMoCo: Instruction Tuning of Large Language Models for Optimization Code Generation, arXiv, 2024
  • Exploring Parameter-Efficient Fine-Tuning Techniques for Code Generation with Large Language Models, arXiv, 2023
  • An Exploratory Study on Fine-Tuning Large Language Models for Secure Code Generation, arXiv, 2024
  • {ITERTL:} An Iterative Framework for Fine-tuning LLMs for {RTL} Code Generation, arXiv, 2024
  • Fine Tuning Large Language Model for Secure Code Generation, Forge, 2024
  • Eliciting Instruction-tuned Code Language Models' Capabilities to Utilize Auxiliary Function for Code Generation, arXiv, 2024
  • Code Less, Align More: Efficient {LLM} Fine-tuning for Code Generation with Data Pruning, arXiv, 2024
  • Evaluating Instruction-Tuned Large Language Models on Code Comprehension and Generation, arXiv, 2023
  • DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions, arXiv, 2023
  • Code Alpaca: An Instruction-following LLaMA Model trained on code generation instructions, ``, 2023
  • OctoPack: Instruction Tuning Code Large Language Models, arXiv, 2023
  • WizardCoder: Empowering Code Large Language Models with Evol-Instruct, arXiv, 2023
  • Magicoder: Source Code Is All You Need, arXiv, 2023
  • Enhancing Code Generation Performance of Smaller Models by Distilling the Reasoning Ability of LLMs, COLING, 2024
  • Multi-Programming Language Ensemble for Code Generation in Large Language Model, arXiv, 2024
  • Personalized Distillation: Empowering Open-Sourced LLMs with Adaptive Learning for Code Generation, EMNLP, 2023
  • Astraios: Parameter-Efficient Instruction Tuning Code Large Language Models, arXiv, 2024

Reinforcement Learning with Feedback

  • Aligning Crowd-sourced Human Feedback for Code Generation wi
View on GitHub
GitHub Stars192
CategoryDevelopment
Updated1d ago
Forks9

Security Score

80/100

Audited on Mar 27, 2026

No findings