The Impact of Knowledge Distillation on the Energy Consumption and Runtime Efficiency of NLP Models

Context. While models like BERT and GPT are powerful, they require substantial resources. Knowledge distillation can be employed as a technique to enhance their efficiency. Yet, we lack a clear understanding on their performance and energy consumption. This uncertainty is a major concern, especially...

Full description

Saved in:
Bibliographic Details
Published in2024 IEEE/ACM 3rd International Conference on AI Engineering – Software Engineering for AI (CAIN) pp. 129 - 133
Main Authors Yuan, Ye, Zhang, Jingzhi, Zhang, Zongyao, Chen, Kaiwei, Shi, Jiacheng, Stoico, Vincenzo, Malavolta, Ivano
Format Conference Proceeding
LanguageEnglish
Published ACM 14.04.2024
Subjects
Online AccessGet full text
DOI10.1145/3644815.3644966

Cover

Abstract Context. While models like BERT and GPT are powerful, they require substantial resources. Knowledge distillation can be employed as a technique to enhance their efficiency. Yet, we lack a clear understanding on their performance and energy consumption. This uncertainty is a major concern, especially in practical applications, where these models could strain resources and limit accessibility for developers with limited means. Our drive also comes from the pressing need for environmentally-friendly and sustainable applications in light of growing environmental worries. To address this, it is crucial to accurately measure their energy consumption.Goal. This study aims to determine how Knowledge Distillation affects the energy consumption and performance of NLP models.Method. We benchmark BERT, Distilled-BERT, GPT-2, and Distilled-GPT-2 using three different tasks from 3 different categories selected from a third-party dataset. The energy consumption, CPU utilization, memory utilization, and inference time of the considered NLP models are measured and statistically analyzed.Results. We observed notable differences between the original and the distilled version of the measured NLP models. Distilled versions tend to consume less energy, while distilled GPT-2 uses less CPU. Conclusion. The results of this study highlight the critical impact of model choice on performance and energy consumption metrics. Future research should consider a wider range of distilled models, diverse benchmarks, and deployment environments, as well as explore the ecological footprint of these models, particularly in the context of environmental sustainability.
AbstractList Context. While models like BERT and GPT are powerful, they require substantial resources. Knowledge distillation can be employed as a technique to enhance their efficiency. Yet, we lack a clear understanding on their performance and energy consumption. This uncertainty is a major concern, especially in practical applications, where these models could strain resources and limit accessibility for developers with limited means. Our drive also comes from the pressing need for environmentally-friendly and sustainable applications in light of growing environmental worries. To address this, it is crucial to accurately measure their energy consumption.Goal. This study aims to determine how Knowledge Distillation affects the energy consumption and performance of NLP models.Method. We benchmark BERT, Distilled-BERT, GPT-2, and Distilled-GPT-2 using three different tasks from 3 different categories selected from a third-party dataset. The energy consumption, CPU utilization, memory utilization, and inference time of the considered NLP models are measured and statistically analyzed.Results. We observed notable differences between the original and the distilled version of the measured NLP models. Distilled versions tend to consume less energy, while distilled GPT-2 uses less CPU. Conclusion. The results of this study highlight the critical impact of model choice on performance and energy consumption metrics. Future research should consider a wider range of distilled models, diverse benchmarks, and deployment environments, as well as explore the ecological footprint of these models, particularly in the context of environmental sustainability.
Author Yuan, Ye
Malavolta, Ivano
Stoico, Vincenzo
Zhang, Jingzhi
Chen, Kaiwei
Zhang, Zongyao
Shi, Jiacheng
Author_xml – sequence: 1
  givenname: Ye
  surname: Yuan
  fullname: Yuan, Ye
  email: y.yuan3@student.vu.nl
  organization: Vrije Universiteit,Amsterdam,The Netherlands
– sequence: 2
  givenname: Jingzhi
  surname: Zhang
  fullname: Zhang, Jingzhi
  email: j.zhang6@student.vu.nl
  organization: Vrije Universiteit,Amsterdam,The Netherlands
– sequence: 3
  givenname: Zongyao
  surname: Zhang
  fullname: Zhang, Zongyao
  email: z.zhang14@student.vu.nl
  organization: Vrije Universiteit,Amsterdam,The Netherlands
– sequence: 4
  givenname: Kaiwei
  surname: Chen
  fullname: Chen, Kaiwei
  email: k.chen2@student.vu.nl
  organization: Vrije Universiteit,Amsterdam,The Netherlands
– sequence: 5
  givenname: Jiacheng
  surname: Shi
  fullname: Shi, Jiacheng
  email: j.shi2@student.vu.nl
  organization: Vrije Universiteit,Amsterdam,The Netherlands
– sequence: 6
  givenname: Vincenzo
  surname: Stoico
  fullname: Stoico, Vincenzo
  email: v.stoico@vu.nl
  organization: Vrije Universiteit,Amsterdam,The Netherlands
– sequence: 7
  givenname: Ivano
  surname: Malavolta
  fullname: Malavolta, Ivano
  email: i.malavolta@vu.nl
  organization: Vrije Universiteit,Amsterdam,The Netherlands
BookMark eNotT0tLxDAYjKCgrj178ZA_sGvSvI9SV12sD2Q9L2n6ZQ20aWm6SP-9XRWGmcMMw8wlOo1dBISuKVlRysUtk5xrKlZHNVKeoMwoozkhighDxTnKUgoVEUJzzZi5QG77BXjT9taNuPP4OXbfDdR7wPchjaFp7Bi6iGeMc24dYdhPuOhiOrT9r2NjjT8OcQztbHsfXIDopmPVa_mOX7oamnSFzrxtEmT_ukCfD-tt8bQs3x43xV25tLk241JSp7hTwhOdz1yxnMA83FRKOump8rVWVHPqrTfSe2ccrThTuSYUwHDOFujmrzcAwK4fQmuHaUfnt8Ioxn4AOD5VZA
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
ESBDL
RIE
RIL
DOI 10.1145/3644815.3644966
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Xplore POP ALL
IEEE Xplore All Conference Proceedings
IEEE Xplore Open Access Journals (WRLC)
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 9798400705915
EndPage 133
ExternalDocumentID 10555973
Genre orig-research
GroupedDBID 6IE
6IL
ACM
ALMA_UNASSIGNED_HOLDINGS
APO
CBEJK
ESBDL
LHSKQ
RIE
RIL
ID FETCH-LOGICAL-a289t-61c74c75f08275fb320e0079b76c6f17fd871841faf96ffc9c1b4372801ee9443
IEDL.DBID RIE
IngestDate Thu May 08 06:04:18 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a289t-61c74c75f08275fb320e0079b76c6f17fd871841faf96ffc9c1b4372801ee9443
OpenAccessLink https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/document/10555973
PageCount 5
ParticipantIDs ieee_primary_10555973
PublicationCentury 2000
PublicationDate 2024-April-14
PublicationDateYYYYMMDD 2024-04-14
PublicationDate_xml – month: 04
  year: 2024
  text: 2024-April-14
  day: 14
PublicationDecade 2020
PublicationTitle 2024 IEEE/ACM 3rd International Conference on AI Engineering – Software Engineering for AI (CAIN)
PublicationTitleAbbrev CAIN
PublicationYear 2024
Publisher ACM
Publisher_xml – name: ACM
SSID ssib055848339
Score 1.9259596
Snippet Context. While models like BERT and GPT are powerful, they require substantial resources. Knowledge distillation can be employed as a technique to enhance...
SourceID ieee
SourceType Publisher
StartPage 129
SubjectTerms Analytical models
Biological system modeling
Computational modeling
Empirical Software Engineering
Energy consumption
Energy-Efficient Software
Green products
Memory management
NLP models
Performance evaluation
Title The Impact of Knowledge Distillation on the Energy Consumption and Runtime Efficiency of NLP Models
URI https://ieeexplore.ieee.org/document/10555973
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1LS8NAEF5sT55UjPhmD16Tut3dJHuuLfVBKWKht7LPizUVTS_-emc2iYogCCEJCWTDzCbfzM58M4RcGaVZaQNPnSvyVDidp2C1mlQwbaUBd8zpmOU7y6cLcbeUy5asHrkw3vuYfOYzPI2xfLexW1wqG2AzRzCAeY_0YJ41ZK1u8khA0pJz1ZbvYUIOOPoeTGZ4VFgG8Uf_lAgfkz0y6wZuskaes21tMvvxqybjv99snyTfTD06_8KgA7Ljq0NiQff0NtIf6SbQ-27ZjN7gB71ust8obGD80XEk_9FRpGLG_wfVlaOP2ELiBW7HEhPIz8RHzR7mFLunrd8TspiMn0bTtG2mkGrwqWpwEW0hbCEDYD7sDR9ee7APlClymwdWBAeuUylY0EHlIVhlmcGYHiCY90oIfkT61abyx4RyaRjT4GdwD2oN3HjmShAyD9IFL-QJSVBCq9emXsaqE87pH9fPyO4QTAWM0TBxTvr129ZfANTX5jKq-BOaAqle
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1LS8NAEB58HPSkYsW3e_Ca1O3u5nGuLa2toUgLvZV9XqypaHrx1zu7aVQEQQhJSCAJs0m--XbmmwG4VbmkmXYsMiZNIm5kEqHXqiJOpRYK6ZiRIcu3SAYz_jAX841YPWhhrLUh-czGfjfE8s1Kr_1UWds3c0QHmG3DLgI_F7Vcq3l9BGJpxli-KeBDuWgzzz6oiP0294UQf3RQCQDSP4CiuXWdN_IcrysV649fVRn__WyH0PrW6pHJFwodwZYtj0Hj6JNhEECSlSOjZuKM3PtPelnnvxFc0P0jvSD_I90gxgx_ECJLQ558E4kXPB2KTHiFpr9UMZ4Q3z9t-d6CWb837Q6iTTuFSCKrqpAk6pTrVDhEfVwr1rmz6CHkKk104mjqDJKnjFMnXZ44p3NNlY_qIYZZm3POTmCnXJX2FAgTilKJTINZHFjHlKUmQyMzJ4yzXJxBy1to8VpXzFg0xjn_4_gN7A2mj-PFeFiMLmC_g46Dj9hQfgk71dvaXiHwV-o6DPcnqGmsqw
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2024+IEEE%2FACM+3rd+International+Conference+on+AI+Engineering+%E2%80%93+Software+Engineering+for+AI+%28CAIN%29&rft.atitle=The+Impact+of+Knowledge+Distillation+on+the+Energy+Consumption+and+Runtime+Efficiency+of+NLP+Models&rft.au=Yuan%2C+Ye&rft.au=Zhang%2C+Jingzhi&rft.au=Zhang%2C+Zongyao&rft.au=Chen%2C+Kaiwei&rft.date=2024-04-14&rft.pub=ACM&rft.spage=129&rft.epage=133&rft_id=info:doi/10.1145%2F3644815.3644966&rft.externalDocID=10555973