A Survey of Large Language Models for Code: Evolution, Benchmarking, and Future Trends
General large language models (LLMs), represented by ChatGPT, have demonstrated significant potential in tasks such as code generation in software engineering. This has led to the development of specialized LLMs for software engineering, known as Code LLMs. A considerable portion of Code LLMs is der...
Saved in:
Published in | arXiv.org |
---|---|
Main Authors | , , , , , , |
Format | Paper |
Language | English |
Published |
Ithaca
Cornell University Library, arXiv.org
17.11.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | General large language models (LLMs), represented by ChatGPT, have demonstrated significant potential in tasks such as code generation in software engineering. This has led to the development of specialized LLMs for software engineering, known as Code LLMs. A considerable portion of Code LLMs is derived from general LLMs through model fine-tuning. As a result, Code LLMs are often updated frequently and their performance can be influenced by the base LLMs. However, there is currently a lack of systematic investigation into Code LLMs and their performance. In this study, we conduct a comprehensive survey and analysis of the types of Code LLMs and their differences in performance compared to general LLMs. We aim to address three questions: (1) What LLMs are specifically designed for software engineering tasks, and what is the relationship between these Code LLMs? (2) Do Code LLMs really outperform general LLMs in software engineering tasks? (3) Which LLMs are more proficient in different software engineering tasks? To answer these questions, we first collect relevant literature and work from five major databases and open-source communities, resulting in 134 works for analysis. Next, we categorize the Code LLMs based on their publishers and examine their relationships with general LLMs and among themselves. Furthermore, we investigate the performance differences between general LLMs and Code LLMs in various software engineering tasks to demonstrate the impact of base models and Code LLMs. Finally, we comprehensively maintained the performance of LLMs across multiple mainstream benchmarks to identify the best-performing LLMs for each software engineering task. Our research not only assists developers of Code LLMs in choosing base models for the development of more advanced LLMs but also provides insights for practitioners to better understand key improvement directions for Code LLMs. |
---|---|
AbstractList | General large language models (LLMs), represented by ChatGPT, have demonstrated significant potential in tasks such as code generation in software engineering. This has led to the development of specialized LLMs for software engineering, known as Code LLMs. A considerable portion of Code LLMs is derived from general LLMs through model fine-tuning. As a result, Code LLMs are often updated frequently and their performance can be influenced by the base LLMs. However, there is currently a lack of systematic investigation into Code LLMs and their performance. In this study, we conduct a comprehensive survey and analysis of the types of Code LLMs and their differences in performance compared to general LLMs. We aim to address three questions: (1) What LLMs are specifically designed for software engineering tasks, and what is the relationship between these Code LLMs? (2) Do Code LLMs really outperform general LLMs in software engineering tasks? (3) Which LLMs are more proficient in different software engineering tasks? To answer these questions, we first collect relevant literature and work from five major databases and open-source communities, resulting in 134 works for analysis. Next, we categorize the Code LLMs based on their publishers and examine their relationships with general LLMs and among themselves. Furthermore, we investigate the performance differences between general LLMs and Code LLMs in various software engineering tasks to demonstrate the impact of base models and Code LLMs. Finally, we comprehensively maintained the performance of LLMs across multiple mainstream benchmarks to identify the best-performing LLMs for each software engineering task. Our research not only assists developers of Code LLMs in choosing base models for the development of more advanced LLMs but also provides insights for practitioners to better understand key improvement directions for Code LLMs. |
Author | Wang, Yanlin Ye, Mingxi Ning, Kaiwen Zheng, Dewu Chen, Jiachi Zheng, Zibin Zhang, Jingwen |
Author_xml | – sequence: 1 givenname: Zibin surname: Zheng fullname: Zheng, Zibin – sequence: 2 givenname: Kaiwen surname: Ning fullname: Ning, Kaiwen – sequence: 3 givenname: Yanlin surname: Wang fullname: Wang, Yanlin – sequence: 4 givenname: Jingwen surname: Zhang fullname: Zhang, Jingwen – sequence: 5 givenname: Dewu surname: Zheng fullname: Zheng, Dewu – sequence: 6 givenname: Mingxi surname: Ye fullname: Ye, Mingxi – sequence: 7 givenname: Jiachi surname: Chen fullname: Chen, Jiachi |
BookMark | eNqNis0KgkAURocoyMp3uNDWwMaftF2J0aJWSVsZ8mqa3akZR-jtc9EDtDnfge_M2Jgk4YhZ3PPWq8jnfMpsrRvXdXm44UHgWey6g4tRPX5AlnASqsKBVBkxyFkW2GoopYJk0C2kvWxNV0tyYI90uz-FetRUOSCogIPpjELIFFKhF2xSilaj_ds5Wx7SLDmuXkq-Deoub6RRNFw5j2LfDaI4jL3_qi-tuUGo |
ContentType | Paper |
Copyright | 2023. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
Copyright_xml | – notice: 2023. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
DBID | 8FE 8FG ABJCF ABUWG AFKRA AZQEC BENPR BGLVJ CCPQU DWQXO HCIFZ L6V M7S PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
DatabaseName | ProQuest SciTech Collection ProQuest Technology Collection Materials Science & Engineering Collection ProQuest Central (Alumni) ProQuest Central ProQuest Central Essentials AUTh Library subscriptions: ProQuest Central Technology Collection ProQuest One Community College ProQuest Central SciTech Premium Collection ProQuest Engineering Collection Engineering Database Publicly Available Content Database ProQuest One Academic Eastern Edition (DO NOT USE) ProQuest One Academic ProQuest One Academic UKI Edition ProQuest Central China Engineering Collection |
DatabaseTitle | Publicly Available Content Database Engineering Database Technology Collection ProQuest Central Essentials ProQuest One Academic Eastern Edition ProQuest Central (Alumni Edition) SciTech Premium Collection ProQuest One Community College ProQuest Technology Collection ProQuest SciTech Collection ProQuest Central China ProQuest Central ProQuest Engineering Collection ProQuest One Academic UKI Edition ProQuest Central Korea Materials Science & Engineering Collection ProQuest One Academic Engineering Collection |
DatabaseTitleList | Publicly Available Content Database |
Database_xml | – sequence: 1 dbid: 8FG name: ProQuest Technology Collection url: https://search.proquest.com/technologycollection1 sourceTypes: Aggregation Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Physics |
EISSN | 2331-8422 |
Genre | Working Paper/Pre-Print |
GroupedDBID | 8FE 8FG ABJCF ABUWG AFKRA ALMA_UNASSIGNED_HOLDINGS AZQEC BENPR BGLVJ CCPQU DWQXO FRJ HCIFZ L6V M7S M~E PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
ID | FETCH-proquest_journals_28940589693 |
IEDL.DBID | 8FG |
IngestDate | Tue Sep 24 20:16:21 EDT 2024 |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-proquest_journals_28940589693 |
OpenAccessLink | https://www.proquest.com/docview/2894058969/abstract/?pq-origsite=%requestingapplication% |
PQID | 2894058969 |
PQPubID | 2050157 |
ParticipantIDs | proquest_journals_2894058969 |
PublicationCentury | 2000 |
PublicationDate | 20231117 |
PublicationDateYYYYMMDD | 2023-11-17 |
PublicationDate_xml | – month: 11 year: 2023 text: 20231117 day: 17 |
PublicationDecade | 2020 |
PublicationPlace | Ithaca |
PublicationPlace_xml | – name: Ithaca |
PublicationTitle | arXiv.org |
PublicationYear | 2023 |
Publisher | Cornell University Library, arXiv.org |
Publisher_xml | – name: Cornell University Library, arXiv.org |
SSID | ssj0002672553 |
Score | 3.4943087 |
SecondaryResourceType | preprint |
Snippet | General large language models (LLMs), represented by ChatGPT, have demonstrated significant potential in tasks such as code generation in software engineering.... |
SourceID | proquest |
SourceType | Aggregation Database |
SubjectTerms | Large language models Questions Software engineering |
Title | A Survey of Large Language Models for Code: Evolution, Benchmarking, and Future Trends |
URI | https://www.proquest.com/docview/2894058969/abstract/ |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1LS8NAEB5qg-DNJz5qGdBjQ4h5exFbEoPYUnzRW9ndJHjQpCZpwYu_3Z1togehlyVLICRhmPny5Zv5AC65xKDC4VwPEtPSbR7IPCgCm2YRSrSf-J6v5haMJ278Yt_PnFkH4rYXhmSVbU5UiTopBHHkhvwwsMkCzw0MxokFELVxs_jUyT-K_rM2ZhpboJk0E496xqO7X7blyvUkdrb-JVxVRaJd0KZskZZ70EnzfdhW4ktRHcDrLT4ty1X6hUWGDyTNluuaRkTyKnuvUEJLHMnDawxXTbAMcCgj7O2DKbZ7gCxPMFITQnAtdD2Eiyh8HsV6ezPzJnCq-d9jWkfQzYs8PQbkLjU6W77j8cz2OfNNkXgZoSPhWMxJT6C36Uqnm0-fwQ55qFODnen1oFuXy_RcVtqa99VL7IM2DCfTR7kbf4c__q-IGg |
link.rule.ids | 786,790,12792,21416,33408,33779,43635,43840 |
linkProvider | ProQuest |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1bS8MwFD5oi-ibV7xMDejjitSkN1_EjZaqXRk6ZW-lSVP2oNtst4H_3pOs0wdhLyEQCEk4nHzny8n5AK45YlDhcG4FhU0txgP0gyJgqhYhov3C93xdt6CXuvEbexo6w4Zwq5u0ypVP1I66mAjFkd9gYMCUBJ4b3E-_LKUapV5XGwmNTTAZxVDFALMTpv2XX5bl1vUQM9N_jlbfHtEumP18Kqs92JDjfdjSSZeiPoD3B_I6rxbym0xKkqiUbGyX9CFRGmUfNUFISbrYvSPhojGSNumgZY0-c81yt0k-LkikK4OQZYLrIVxF4aAbW6vFZI3B1Nnf9ugRGBj5y2Mg3FUfnKnveLxkPs99WxReqVCRcGjuyBNorZvpdP3wJWzHg16SJY_p8xnsKB119cnO9lpgzKq5PMfbdsYvmiP9AbtKheM |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=A+Survey+of+Large+Language+Models+for+Code%3A+Evolution%2C+Benchmarking%2C+and+Future+Trends&rft.jtitle=arXiv.org&rft.au=Zheng%2C+Zibin&rft.au=Ning%2C+Kaiwen&rft.au=Wang%2C+Yanlin&rft.au=Zhang%2C+Jingwen&rft.date=2023-11-17&rft.pub=Cornell+University+Library%2C+arXiv.org&rft.eissn=2331-8422 |