Quantization Framework for Fast Spiking Neural Networks
Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit pr...
Saved in:
Published in | Frontiers in neuroscience Vol. 16; p. 918793 |
---|---|
Main Authors | , , |
Format | Journal Article |
Language | English |
Published |
Frontiers Media S.A
19.07.2022
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit precision of the artificial neurons and the time required by the spiking neurons to represent the same bit precision. This implicit link suggests that techniques used to reduce the activation bit precision of ANNs, such as quantization, can help shorten the inference latency of SNNs. However, carrying ANN quantization knowledge over to SNNs is not straightforward, as there are many fundamental differences between them. Here we propose a quantization framework for fast SNNs (QFFS) to overcome these difficulties, providing a method to build SNNs with enhanced latency and reduced loss of accuracy relative to the baseline ANN model. In this framework, we promote the compatibility of ANN information quantization techniques with SNNs, and suppress “occasional noise” to minimize accuracy loss. The resulting SNNs overcome the accuracy degeneration observed previously in SNNs with a limited number of time steps and achieve an accuracy of 70.18% on ImageNet within 8 time steps. This is the first demonstration that SNNs built by ANN-to-SNN conversion can achieve a similar latency to SNNs built by direct training. |
---|---|
AbstractList | Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit precision of the artificial neurons and the time required by the spiking neurons to represent the same bit precision. This implicit link suggests that techniques used to reduce the activation bit precision of ANNs, such as quantization, can help shorten the inference latency of SNNs. However, carrying ANN quantization knowledge over to SNNs is not straightforward, as there are many fundamental differences between them. Here we propose a quantization framework for fast SNNs (QFFS) to overcome these difficulties, providing a method to build SNNs with enhanced latency and reduced loss of accuracy relative to the baseline ANN model. In this framework, we promote the compatibility of ANN information quantization techniques with SNNs, and suppress “occasional noise” to minimize accuracy loss. The resulting SNNs overcome the accuracy degeneration observed previously in SNNs with a limited number of time steps and achieve an accuracy of 70.18% on ImageNet within 8 time steps. This is the first demonstration that SNNs built by ANN-to-SNN conversion can achieve a similar latency to SNNs built by direct training. Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit precision of the artificial neurons and the time required by the spiking neurons to represent the same bit precision. This implicit link suggests that techniques used to reduce the activation bit precision of ANNs, such as quantization, can help shorten the inference latency of SNNs. However, carrying ANN quantization knowledge over to SNNs is not straightforward, as there are many fundamental differences between them. Here we propose a quantization framework for fast SNNs (QFFS) to overcome these difficulties, providing a method to build SNNs with enhanced latency and reduced loss of accuracy relative to the baseline ANN model. In this framework, we promote the compatibility of ANN information quantization techniques with SNNs, and suppress "occasional noise" to minimize accuracy loss. The resulting SNNs overcome the accuracy degeneration observed previously in SNNs with a limited number of time steps and achieve an accuracy of 70.18% on ImageNet within 8 time steps. This is the first demonstration that SNNs built by ANN-to-SNN conversion can achieve a similar latency to SNNs built by direct training.Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit precision of the artificial neurons and the time required by the spiking neurons to represent the same bit precision. This implicit link suggests that techniques used to reduce the activation bit precision of ANNs, such as quantization, can help shorten the inference latency of SNNs. However, carrying ANN quantization knowledge over to SNNs is not straightforward, as there are many fundamental differences between them. Here we propose a quantization framework for fast SNNs (QFFS) to overcome these difficulties, providing a method to build SNNs with enhanced latency and reduced loss of accuracy relative to the baseline ANN model. In this framework, we promote the compatibility of ANN information quantization techniques with SNNs, and suppress "occasional noise" to minimize accuracy loss. The resulting SNNs overcome the accuracy degeneration observed previously in SNNs with a limited number of time steps and achieve an accuracy of 70.18% on ImageNet within 8 time steps. This is the first demonstration that SNNs built by ANN-to-SNN conversion can achieve a similar latency to SNNs built by direct training. |
Author | Li, Chen Ma, Lei Furber, Steve |
AuthorAffiliation | 2 Beijing Academy of Artificial Intelligence , Beijing , China 3 National Biomedical Imaging Center, Peking University , Beijing , China 1 Advanced Processor Technologies (APT) Group, Department of Computer Science, The University of Manchester , Manchester , United Kingdom |
AuthorAffiliation_xml | – name: 1 Advanced Processor Technologies (APT) Group, Department of Computer Science, The University of Manchester , Manchester , United Kingdom – name: 2 Beijing Academy of Artificial Intelligence , Beijing , China – name: 3 National Biomedical Imaging Center, Peking University , Beijing , China |
Author_xml | – sequence: 1 givenname: Chen surname: Li fullname: Li, Chen – sequence: 2 givenname: Lei surname: Ma fullname: Ma, Lei – sequence: 3 givenname: Steve surname: Furber fullname: Furber, Steve |
BookMark | eNp1kUtvEzEUhS1URB_wA9jNkk3C-G1vkFBFSqUKhFokdtYdP4LbiR1sTyv49UySIlEkVtfyuee71z6n6Cjl5BF6jfslpUq_DSmmuiQ9IUuNldT0GTrBQpAF4_Tb0V_nY3Ra623fC6IYeYGOKddE9RifIPllgtTiL2gxp25VYOMfcrnrQi7dCmrrrrfxLqZ198lPBca5tJ1eX6LnAcbqXz3WM_R19eHm_OPi6vPF5fn7q4VljLTFAOAH4YN1gnBCqLVhACVk4MI5i51jlGjvtAxODg4sxUpZhkEzyYUUjp6hywPXZbg12xI3UH6aDNHsL3JZGygt2tEbLjAB6jjGEFhwA3gt2YD5wK0M8-CZ9e7A2k7DxjvrU5uf9AT6VEnxu1nne6MpY0rpGfDmEVDyj8nXZjaxWj-OkHyeqiFCa9lTzuTcKg-ttuRaiw_Gxrb_5JkcR4N7s4vQ7CM0uwjNIcLZif9x_lnw_57ft3WjuA |
CitedBy_id | crossref_primary_10_3389_fnins_2023_1224457 crossref_primary_10_3390_electronics13091744 crossref_primary_10_1109_ACCESS_2024_3391889 crossref_primary_10_1109_JETCAS_2023_3328863 crossref_primary_10_1109_TCDS_2023_3329747 crossref_primary_10_3389_fnins_2023_1230002 crossref_primary_10_3389_fnins_2024_1449020 crossref_primary_10_1002_aisy_202300383 crossref_primary_10_1109_TETCI_2024_3359539 crossref_primary_10_3389_fnins_2023_1154241 crossref_primary_10_3389_fnins_2024_1440000 |
Cites_doi | 10.3389/fnins.2019.00095 10.48550/arXiv.2105.12917 10.1109/IJCNN.2015.7280696 10.1109/DAC18074.2021.9586266 10.3389/fnins.2021.629000 10.1016/j.patcog.2020.107281 10.3389/neuro.11.011.2008 10.24963/ijcai.2021/321 10.1109/TPAMI.2021.3114196 10.3389/fnins.2020.00535 10.1145/3477145.3477158 10.48550/arXiv.2102.04159 10.48550/arXiv.1806.08342 10.48550/arXiv.1902.08153 10.48550/arXiv.2008.03658 10.1109/IJCNN52387.2021.9534111 10.48550/arXiv.2106.06984 10.1145/3407197.3407203 10.3389/fnins.2017.00682 10.48550/arXiv.2103.00476 10.1109/IJCNN52387.2021.9533874 |
ContentType | Journal Article |
Copyright | Copyright © 2022 Li, Ma and Furber. Copyright © 2022 Li, Ma and Furber. 2022 Li, Ma and Furber |
Copyright_xml | – notice: Copyright © 2022 Li, Ma and Furber. – notice: Copyright © 2022 Li, Ma and Furber. 2022 Li, Ma and Furber |
DBID | AAYXX CITATION 7X8 5PM DOA |
DOI | 10.3389/fnins.2022.918793 |
DatabaseName | CrossRef MEDLINE - Academic PubMed Central (Full Participant titles) DOAJ Directory of Open Access Journals |
DatabaseTitle | CrossRef MEDLINE - Academic |
DatabaseTitleList | MEDLINE - Academic CrossRef |
Database_xml | – sequence: 1 dbid: DOA name: DOAJ Directory of Open Access Journals url: https://www.doaj.org/ sourceTypes: Open Website |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Anatomy & Physiology |
EISSN | 1662-453X |
ExternalDocumentID | oai_doaj_org_article_5612a3d511af4fdbae974b15b5c7fccf PMC9344889 10_3389_fnins_2022_918793 |
GroupedDBID | --- 29H 2WC 53G 5GY 5VS 8FE 8FH 9T4 AAFWJ AAYXX ABUWG ACGFO ACGFS ACXDI ADRAZ AEGXH AENEX AFKRA AFPKN AIAGR ALMA_UNASSIGNED_HOLDINGS AZQEC BBNVY BENPR BHPHI BPHCQ CITATION CS3 DIK DU5 E3Z EBS EJD EMOBN F5P FRP GROUPED_DOAJ GX1 HCIFZ HYE KQ8 LK8 M2P M48 M7P O5R O5S OK1 OVT P2P PGMZT PIMPY PQQKQ PROAC RNS RPM W2D 7X8 5PM |
ID | FETCH-LOGICAL-c442t-baaeb6efcd625223ccfba867f56ddc1dd4329ed97fd7bdac3188c41a9475676d3 |
IEDL.DBID | DOA |
ISSN | 1662-453X 1662-4548 |
IngestDate | Wed Aug 27 01:27:17 EDT 2025 Thu Aug 21 18:16:30 EDT 2025 Fri Jul 11 07:04:31 EDT 2025 Tue Jul 01 01:39:41 EDT 2025 Thu Apr 24 23:09:09 EDT 2025 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | true |
IsScholarly | true |
Language | English |
License | This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-c442t-baaeb6efcd625223ccfba867f56ddc1dd4329ed97fd7bdac3188c41a9475676d3 |
Notes | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 23 This article was submitted to Neuromorphic Engineering, a section of the journal Frontiers in Neuroscience Reviewed by: Dongsuk Jeon, Seoul National University, South Korea; Yukuan Yang, Tsinghua University, China Edited by: Guoqi Li, Tsinghua University, China |
OpenAccessLink | https://doaj.org/article/5612a3d511af4fdbae974b15b5c7fccf |
PMID | 35928011 |
PQID | 2699703547 |
PQPubID | 23479 |
ParticipantIDs | doaj_primary_oai_doaj_org_article_5612a3d511af4fdbae974b15b5c7fccf pubmedcentral_primary_oai_pubmedcentral_nih_gov_9344889 proquest_miscellaneous_2699703547 crossref_citationtrail_10_3389_fnins_2022_918793 crossref_primary_10_3389_fnins_2022_918793 |
ProviderPackageCode | CITATION AAYXX |
PublicationCentury | 2000 |
PublicationDate | 2022-07-19 |
PublicationDateYYYYMMDD | 2022-07-19 |
PublicationDate_xml | – month: 07 year: 2022 text: 2022-07-19 day: 19 |
PublicationDecade | 2020 |
PublicationTitle | Frontiers in neuroscience |
PublicationYear | 2022 |
Publisher | Frontiers Media S.A |
Publisher_xml | – name: Frontiers Media S.A |
References | Sengupta (B23) 2019; 13 B25 Hwang (B11) 2021; 15 Esser (B7) 2019 Ho (B10) 2020 Mueller (B18) 2021 Schaefer (B22) 2020 Qin (B19) 2020; 105 Diehl (B5) 2015 Fang (B8) 2021 Chowdhury (B2) 2021 Rathi (B20) 2020 Li (B15) Deng (B4) 2021 Bu (B1) 2021 Li (B13) Ding (B6) 2021 Han (B9) 2020 Lu (B16) 2020; 14 Davison (B3) 2009; 2 Warden (B24) 2019 Krishnamoorthi (B12) 2018 Rueckauer (B21) 2017; 11 Lui (B17) 2021 Li (B14) |
References_xml | – volume: 13 start-page: 95 year: 2019 ident: B23 article-title: Going deeper in spiking neural networks: VGG and residual architectures publication-title: Front. Neurosci doi: 10.3389/fnins.2019.00095 – ident: B15 article-title: Bsnn: towards faster and better conversion of artificial neural networks to spiking neural networks with bistable neurons publication-title: arXiv preprint arXiv:2105.12917 doi: 10.48550/arXiv.2105.12917 – start-page: 1 volume-title: 2015 International Joint Conference on Neural Networks (IJCNN) year: 2015 ident: B5 article-title: Fast-classifying, high-accuracy spiking deep networks through weight and threshold balancing doi: 10.1109/IJCNN.2015.7280696 – year: 2020 ident: B10 article-title: TCL: an ann-to-snn conversion with trainable clipping layers publication-title: arXiv preprint arXiv:2008.04509 doi: 10.1109/DAC18074.2021.9586266 – volume: 15 start-page: 135 year: 2021 ident: B11 article-title: Low-latency spiking neural networks using pre-charged membrane potential and delayed evaluation publication-title: Front. Neurosci doi: 10.3389/fnins.2021.629000 – volume: 105 start-page: 107281 year: 2020 ident: B19 article-title: Binary neural networks: a survey publication-title: Pattern Recognit doi: 10.1016/j.patcog.2020.107281 – volume: 2 start-page: 11 year: 2009 ident: B3 article-title: Pynn: a common interface for neuronal network simulators publication-title: Front Neuroinform doi: 10.3389/neuro.11.011.2008 – start-page: 13558 volume-title: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition year: 2020 ident: B9 article-title: RMP-SNN: residual membrane potential neuron for enabling deeper high-accuracy and low-latency spiking neural network – year: 2021 ident: B6 article-title: Optimal ann-snn conversion for fast and accurate inference in deep spiking neural networks publication-title: arXiv preprint arXiv:2105.11654 doi: 10.24963/ijcai.2021/321 – year: 2021 ident: B1 article-title: Optimal ann-snn conversion for high-accuracy and ultra-low-latency spiking neural networks publication-title: International Conference on Learning Representations – ident: B25 doi: 10.1109/TPAMI.2021.3114196 – volume: 14 start-page: 535 year: 2020 ident: B16 article-title: Exploring the connection between binary and spiking neural networks publication-title: Front. Neurosci doi: 10.3389/fnins.2020.00535 – year: 2021 ident: B17 article-title: Hessian aware quantization of spiking neural networks publication-title: arXiv preprint arXiv:2104.14117 doi: 10.1145/3477145.3477158 – year: 2021 ident: B8 article-title: Deep residual learning in spiking neural networks publication-title: arXiv preprint arXiv:2102.04159 doi: 10.48550/arXiv.2102.04159 – year: 2018 ident: B12 article-title: Quantizing deep convolutional networks for efficient inference: a whitepaper publication-title: arXiv preprint arXiv:1806.08342 doi: 10.48550/arXiv.1806.08342 – year: 2019 ident: B7 article-title: Learned step size quantization publication-title: arXiv preprint arXiv:1902.08153 doi: 10.48550/arXiv.1902.08153 – year: 2020 ident: B20 article-title: Diet-snn: direct input encoding with leakage and threshold optimization in deep spiking neural networks publication-title: arXiv preprint arXiv:2008.03658 doi: 10.48550/arXiv.2008.03658 – year: 2021 ident: B2 article-title: Spatio-temporal pruning and quantization for low-latency spiking neural networks publication-title: arXiv preprint arXiv:2104.12528 doi: 10.1109/IJCNN52387.2021.9534111 – ident: B13 article-title: A free lunch from ann: towards efficient, accurate spiking neural networks calibration publication-title: arXiv preprint arXiv:2106.06984 doi: 10.48550/arXiv.2106.06984 – start-page: 1 year: 2020 ident: B22 article-title: Quantizing spiking neural networks with integers publication-title: International Conference on Neuromorphic Systems 2020 doi: 10.1145/3407197.3407203 – volume: 11 start-page: 682 year: 2017 ident: B21 article-title: Conversion of continuous-valued deep networks to efficient event-driven networks for image classification publication-title: Front. Neurosci doi: 10.3389/fnins.2017.00682 – year: 2021 ident: B4 article-title: Optimal conversion of conventional artificial neural networks to spiking neural networks publication-title: arXiv preprint arXiv:2103.00476 doi: 10.48550/arXiv.2103.00476 – start-page: 1 volume-title: 2021 International Joint Conference on Neural Networks (IJCNN) year: 2021 ident: B18 article-title: Minimizing inference time: optimization methods for converted deep spiking neural networks doi: 10.1109/IJCNN52387.2021.9533874 – volume-title: Tinyml: Machine Learning With Tensorflow Lite on Arduino and Ultra-Low-Power Microcontrollers year: 2019 ident: B24 – start-page: 34 ident: B14 article-title: Differentiable spike: rethinking gradient-descent for training spiking neural networks publication-title: Advances in Neural Information Processing Systems |
SSID | ssj0062842 |
Score | 2.5502822 |
Snippet | Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information... |
SourceID | doaj pubmedcentral proquest crossref |
SourceType | Open Website Open Access Repository Aggregation Database Enrichment Source Index Database |
StartPage | 918793 |
SubjectTerms | ANN-to-SNN conversion fast spiking neural networks inference latency Neuroscience occasional noise quantization spiking neural networks |
SummonAdditionalLinks | – databaseName: Scholars Portal Journals: Open Access dbid: M48 link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1La9tAEB4S99JLaJuWuGnDFkoPBSW29qU9lOKUmlBIIKSG3MQ-E4MrJ35A8-87I8mmgpBTrpJWq_1GszPf7uwMwGf0KAbRFDHz6HtkIqBKOTHwJBCrOHdO1cfFzi_U2UT8upbXO7Apb9UCuHyU2lE9qclidvz3_uE7Kvw3Ypxob09SNa0o83aeHxsqns134QUaJk16ei62mwoKZ-K82dh8vFnHNNUZ_DtuZzdo8j8rNH4Fe637yEaNvF_DTqzewP6oQur854F9YXVAZ71Svg_6co2wtecs2XgThcXQTWVju1yxq7spLZQzStCBL71oIsKXb2Ey_vn7x1nW1knIvBD5KnPWRqdi8gHJDJp775OzhdJJqhD8MATBcxOD0SloF6xHNS68GFojtFRaBf4OetW8igfAch1ToEKgruDCGdRuzX1hpUwiDpKSfRhsICp9m0ScalnMSiQThGpZo1oSqmWDah--bpvcNRk0nnr4lHDfPkjJr-sL88VN2epSSQU9LQ_oKtokUnA2IilyQ-mk1wkH34dPG6mVqCy0A2KrOF9jT8oYnOKk0H3QHXF2euzeqaa3ddptw5HKFub9c3ziIbykUdMi8dB8gN5qsY4f0btZuaP6n_0HFAX9dQ priority: 102 providerName: Scholars Portal |
Title | Quantization Framework for Fast Spiking Neural Networks |
URI | https://www.proquest.com/docview/2699703547 https://pubmed.ncbi.nlm.nih.gov/PMC9344889 https://doaj.org/article/5612a3d511af4fdbae974b15b5c7fccf |
Volume | 16 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1LS8NAEF7EkxfxifXFCuJBiG2yr-xRxVoEBdFCb2GfWNC02Pbgv3cmSaW56MXLHpJNdvNN5rU7O0PIOVgUvaDzkDiwPRLugaUs7zkkiJGMWSur42KPT3Iw5A8jMVop9YUxYXV64Bq4LlZvNMyDXWAij96aABawTYUVTkXnIkpf0HlLZ6qWwRKEblbvYYILpruxHJeYmzvLrjSW12YtLVQl629ZmO34yBWF098im42lSK_rGW6TtVDukN3rErzkjy96QavYzWpRfJeo5wUg1ByppP1lwBUFi5T2zWxOX6ZjXBOnmIsDXvpUB3_P9siwf_d6O0iakgiJ4zybJ9aYYGWIzoPfApodALAmlyoK6b1Lvecs08FrFb2y3jjg2Nzx1GiuhFTSs32yXk7KcEBopkL0WPPT5oxbDYysmMuNEJGHXpSiQ3pLiArX5AvHshXvBfgNiGpRoVogqkWNaodc_jwyrZNl_Nb5BnH_6Yh5rqsLQP2ioX7xF_U75GxJtQL4Ajc7TBkmCxhJag3STHDVIapFztaI7Tvl-K3KsK0ZeK25PvyPKR6RDfxqXA9O9TFZn38uwgkYMnN7Wv2z0N6PUmgfef4N5uf3kw |
linkProvider | Directory of Open Access Journals |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Quantization+Framework+for+Fast+Spiking+Neural+Networks&rft.jtitle=Frontiers+in+neuroscience&rft.au=Chen+Li&rft.au=Lei+Ma&rft.au=Lei+Ma&rft.au=Steve+Furber&rft.date=2022-07-19&rft.pub=Frontiers+Media+S.A&rft.eissn=1662-453X&rft.volume=16&rft_id=info:doi/10.3389%2Ffnins.2022.918793&rft.externalDBID=DOA&rft.externalDocID=oai_doaj_org_article_5612a3d511af4fdbae974b15b5c7fccf |
thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1662-453X&client=summon |
thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1662-453X&client=summon |
thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1662-453X&client=summon |