The Impact of Knowledge Distillation on the Energy Consumption and Runtime Efficiency of NLP Models
Context. While models like BERT and GPT are powerful, they require substantial resources. Knowledge distillation can be employed as a technique to enhance their efficiency. Yet, we lack a clear understanding on their performance and energy consumption. This uncertainty is a major concern, especially...
Saved in:
Published in | 2024 IEEE/ACM 3rd International Conference on AI Engineering – Software Engineering for AI (CAIN) pp. 129 - 133 |
---|---|
Main Authors | , , , , , , |
Format | Conference Proceeding |
Language | English |
Published |
ACM
14.04.2024
|
Subjects | |
Online Access | Get full text |
DOI | 10.1145/3644815.3644966 |
Cover
Abstract | Context. While models like BERT and GPT are powerful, they require substantial resources. Knowledge distillation can be employed as a technique to enhance their efficiency. Yet, we lack a clear understanding on their performance and energy consumption. This uncertainty is a major concern, especially in practical applications, where these models could strain resources and limit accessibility for developers with limited means. Our drive also comes from the pressing need for environmentally-friendly and sustainable applications in light of growing environmental worries. To address this, it is crucial to accurately measure their energy consumption.Goal. This study aims to determine how Knowledge Distillation affects the energy consumption and performance of NLP models.Method. We benchmark BERT, Distilled-BERT, GPT-2, and Distilled-GPT-2 using three different tasks from 3 different categories selected from a third-party dataset. The energy consumption, CPU utilization, memory utilization, and inference time of the considered NLP models are measured and statistically analyzed.Results. We observed notable differences between the original and the distilled version of the measured NLP models. Distilled versions tend to consume less energy, while distilled GPT-2 uses less CPU. Conclusion. The results of this study highlight the critical impact of model choice on performance and energy consumption metrics. Future research should consider a wider range of distilled models, diverse benchmarks, and deployment environments, as well as explore the ecological footprint of these models, particularly in the context of environmental sustainability. |
---|---|
AbstractList | Context. While models like BERT and GPT are powerful, they require substantial resources. Knowledge distillation can be employed as a technique to enhance their efficiency. Yet, we lack a clear understanding on their performance and energy consumption. This uncertainty is a major concern, especially in practical applications, where these models could strain resources and limit accessibility for developers with limited means. Our drive also comes from the pressing need for environmentally-friendly and sustainable applications in light of growing environmental worries. To address this, it is crucial to accurately measure their energy consumption.Goal. This study aims to determine how Knowledge Distillation affects the energy consumption and performance of NLP models.Method. We benchmark BERT, Distilled-BERT, GPT-2, and Distilled-GPT-2 using three different tasks from 3 different categories selected from a third-party dataset. The energy consumption, CPU utilization, memory utilization, and inference time of the considered NLP models are measured and statistically analyzed.Results. We observed notable differences between the original and the distilled version of the measured NLP models. Distilled versions tend to consume less energy, while distilled GPT-2 uses less CPU. Conclusion. The results of this study highlight the critical impact of model choice on performance and energy consumption metrics. Future research should consider a wider range of distilled models, diverse benchmarks, and deployment environments, as well as explore the ecological footprint of these models, particularly in the context of environmental sustainability. |
Author | Yuan, Ye Malavolta, Ivano Stoico, Vincenzo Zhang, Jingzhi Chen, Kaiwei Zhang, Zongyao Shi, Jiacheng |
Author_xml | – sequence: 1 givenname: Ye surname: Yuan fullname: Yuan, Ye email: y.yuan3@student.vu.nl organization: Vrije Universiteit,Amsterdam,The Netherlands – sequence: 2 givenname: Jingzhi surname: Zhang fullname: Zhang, Jingzhi email: j.zhang6@student.vu.nl organization: Vrije Universiteit,Amsterdam,The Netherlands – sequence: 3 givenname: Zongyao surname: Zhang fullname: Zhang, Zongyao email: z.zhang14@student.vu.nl organization: Vrije Universiteit,Amsterdam,The Netherlands – sequence: 4 givenname: Kaiwei surname: Chen fullname: Chen, Kaiwei email: k.chen2@student.vu.nl organization: Vrije Universiteit,Amsterdam,The Netherlands – sequence: 5 givenname: Jiacheng surname: Shi fullname: Shi, Jiacheng email: j.shi2@student.vu.nl organization: Vrije Universiteit,Amsterdam,The Netherlands – sequence: 6 givenname: Vincenzo surname: Stoico fullname: Stoico, Vincenzo email: v.stoico@vu.nl organization: Vrije Universiteit,Amsterdam,The Netherlands – sequence: 7 givenname: Ivano surname: Malavolta fullname: Malavolta, Ivano email: i.malavolta@vu.nl organization: Vrije Universiteit,Amsterdam,The Netherlands |
BookMark | eNotT0tLxDAYjKCgrj178ZA_sGvSvI9SV12sD2Q9L2n6ZQ20aWm6SP-9XRWGmcMMw8wlOo1dBISuKVlRysUtk5xrKlZHNVKeoMwoozkhighDxTnKUgoVEUJzzZi5QG77BXjT9taNuPP4OXbfDdR7wPchjaFp7Bi6iGeMc24dYdhPuOhiOrT9r2NjjT8OcQztbHsfXIDopmPVa_mOX7oamnSFzrxtEmT_ukCfD-tt8bQs3x43xV25tLk241JSp7hTwhOdz1yxnMA83FRKOump8rVWVHPqrTfSe2ccrThTuSYUwHDOFujmrzcAwK4fQmuHaUfnt8Ioxn4AOD5VZA |
CODEN | IEEPAD |
ContentType | Conference Proceeding |
DBID | 6IE 6IL CBEJK ESBDL RIE RIL |
DOI | 10.1145/3644815.3644966 |
DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Xplore POP ALL IEEE Xplore All Conference Proceedings IEEE Xplore Open Access Journals (WRLC) IEEE Electronic Library (IEL) IEEE Proceedings Order Plans (POP All) 1998-Present |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: RIE name: IEEE Electronic Library (IEL) url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
EISBN | 9798400705915 |
EndPage | 133 |
ExternalDocumentID | 10555973 |
Genre | orig-research |
GroupedDBID | 6IE 6IL ACM ALMA_UNASSIGNED_HOLDINGS APO CBEJK ESBDL LHSKQ RIE RIL |
ID | FETCH-LOGICAL-a289t-61c74c75f08275fb320e0079b76c6f17fd871841faf96ffc9c1b4372801ee9443 |
IEDL.DBID | RIE |
IngestDate | Thu May 08 06:04:18 EDT 2025 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-a289t-61c74c75f08275fb320e0079b76c6f17fd871841faf96ffc9c1b4372801ee9443 |
OpenAccessLink | https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/document/10555973 |
PageCount | 5 |
ParticipantIDs | ieee_primary_10555973 |
PublicationCentury | 2000 |
PublicationDate | 2024-April-14 |
PublicationDateYYYYMMDD | 2024-04-14 |
PublicationDate_xml | – month: 04 year: 2024 text: 2024-April-14 day: 14 |
PublicationDecade | 2020 |
PublicationTitle | 2024 IEEE/ACM 3rd International Conference on AI Engineering – Software Engineering for AI (CAIN) |
PublicationTitleAbbrev | CAIN |
PublicationYear | 2024 |
Publisher | ACM |
Publisher_xml | – name: ACM |
SSID | ssib055848339 |
Score | 1.9259596 |
Snippet | Context. While models like BERT and GPT are powerful, they require substantial resources. Knowledge distillation can be employed as a technique to enhance... |
SourceID | ieee |
SourceType | Publisher |
StartPage | 129 |
SubjectTerms | Analytical models Biological system modeling Computational modeling Empirical Software Engineering Energy consumption Energy-Efficient Software Green products Memory management NLP models Performance evaluation |
Title | The Impact of Knowledge Distillation on the Energy Consumption and Runtime Efficiency of NLP Models |
URI | https://ieeexplore.ieee.org/document/10555973 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1LS8NAEF5sT55UjPhmD16Tut3dJHuuLfVBKWKht7LPizUVTS_-emc2iYogCCEJCWTDzCbfzM58M4RcGaVZaQNPnSvyVDidp2C1mlQwbaUBd8zpmOU7y6cLcbeUy5asHrkw3vuYfOYzPI2xfLexW1wqG2AzRzCAeY_0YJ41ZK1u8khA0pJz1ZbvYUIOOPoeTGZ4VFgG8Uf_lAgfkz0y6wZuskaes21tMvvxqybjv99snyTfTD06_8KgA7Ljq0NiQff0NtIf6SbQ-27ZjN7gB71ust8obGD80XEk_9FRpGLG_wfVlaOP2ELiBW7HEhPIz8RHzR7mFLunrd8TspiMn0bTtG2mkGrwqWpwEW0hbCEDYD7sDR9ee7APlClymwdWBAeuUylY0EHlIVhlmcGYHiCY90oIfkT61abyx4RyaRjT4GdwD2oN3HjmShAyD9IFL-QJSVBCq9emXsaqE87pH9fPyO4QTAWM0TBxTvr129ZfANTX5jKq-BOaAqle |
linkProvider | IEEE |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1LS8NAEB58HPSkYsW3e_Ca1O3u5nGuLa2toUgLvZV9XqypaHrx1zu7aVQEQQhJSCAJs0m--XbmmwG4VbmkmXYsMiZNIm5kEqHXqiJOpRYK6ZiRIcu3SAYz_jAX841YPWhhrLUh-czGfjfE8s1Kr_1UWds3c0QHmG3DLgI_F7Vcq3l9BGJpxli-KeBDuWgzzz6oiP0294UQf3RQCQDSP4CiuXWdN_IcrysV649fVRn__WyH0PrW6pHJFwodwZYtj0Hj6JNhEECSlSOjZuKM3PtPelnnvxFc0P0jvSD_I90gxgx_ECJLQ558E4kXPB2KTHiFpr9UMZ4Q3z9t-d6CWb837Q6iTTuFSCKrqpAk6pTrVDhEfVwr1rmz6CHkKk104mjqDJKnjFMnXZ44p3NNlY_qIYZZm3POTmCnXJX2FAgTilKJTINZHFjHlKUmQyMzJ4yzXJxBy1to8VpXzFg0xjn_4_gN7A2mj-PFeFiMLmC_g46Dj9hQfgk71dvaXiHwV-o6DPcnqGmsqw |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2024+IEEE%2FACM+3rd+International+Conference+on+AI+Engineering+%E2%80%93+Software+Engineering+for+AI+%28CAIN%29&rft.atitle=The+Impact+of+Knowledge+Distillation+on+the+Energy+Consumption+and+Runtime+Efficiency+of+NLP+Models&rft.au=Yuan%2C+Ye&rft.au=Zhang%2C+Jingzhi&rft.au=Zhang%2C+Zongyao&rft.au=Chen%2C+Kaiwei&rft.date=2024-04-14&rft.pub=ACM&rft.spage=129&rft.epage=133&rft_id=info:doi/10.1145%2F3644815.3644966&rft.externalDocID=10555973 |