Quantization Framework for Fast Spiking Neural Networks

Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit pr...

Full description

Saved in:
Bibliographic Details
Published inFrontiers in neuroscience Vol. 16; p. 918793
Main Authors Li, Chen, Ma, Lei, Furber, Steve
Format Journal Article
LanguageEnglish
Published Frontiers Media S.A 19.07.2022
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit precision of the artificial neurons and the time required by the spiking neurons to represent the same bit precision. This implicit link suggests that techniques used to reduce the activation bit precision of ANNs, such as quantization, can help shorten the inference latency of SNNs. However, carrying ANN quantization knowledge over to SNNs is not straightforward, as there are many fundamental differences between them. Here we propose a quantization framework for fast SNNs (QFFS) to overcome these difficulties, providing a method to build SNNs with enhanced latency and reduced loss of accuracy relative to the baseline ANN model. In this framework, we promote the compatibility of ANN information quantization techniques with SNNs, and suppress “occasional noise” to minimize accuracy loss. The resulting SNNs overcome the accuracy degeneration observed previously in SNNs with a limited number of time steps and achieve an accuracy of 70.18% on ImageNet within 8 time steps. This is the first demonstration that SNNs built by ANN-to-SNN conversion can achieve a similar latency to SNNs built by direct training.
AbstractList Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit precision of the artificial neurons and the time required by the spiking neurons to represent the same bit precision. This implicit link suggests that techniques used to reduce the activation bit precision of ANNs, such as quantization, can help shorten the inference latency of SNNs. However, carrying ANN quantization knowledge over to SNNs is not straightforward, as there are many fundamental differences between them. Here we propose a quantization framework for fast SNNs (QFFS) to overcome these difficulties, providing a method to build SNNs with enhanced latency and reduced loss of accuracy relative to the baseline ANN model. In this framework, we promote the compatibility of ANN information quantization techniques with SNNs, and suppress “occasional noise” to minimize accuracy loss. The resulting SNNs overcome the accuracy degeneration observed previously in SNNs with a limited number of time steps and achieve an accuracy of 70.18% on ImageNet within 8 time steps. This is the first demonstration that SNNs built by ANN-to-SNN conversion can achieve a similar latency to SNNs built by direct training.
Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit precision of the artificial neurons and the time required by the spiking neurons to represent the same bit precision. This implicit link suggests that techniques used to reduce the activation bit precision of ANNs, such as quantization, can help shorten the inference latency of SNNs. However, carrying ANN quantization knowledge over to SNNs is not straightforward, as there are many fundamental differences between them. Here we propose a quantization framework for fast SNNs (QFFS) to overcome these difficulties, providing a method to build SNNs with enhanced latency and reduced loss of accuracy relative to the baseline ANN model. In this framework, we promote the compatibility of ANN information quantization techniques with SNNs, and suppress "occasional noise" to minimize accuracy loss. The resulting SNNs overcome the accuracy degeneration observed previously in SNNs with a limited number of time steps and achieve an accuracy of 70.18% on ImageNet within 8 time steps. This is the first demonstration that SNNs built by ANN-to-SNN conversion can achieve a similar latency to SNNs built by direct training.Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information transmission rates through the use of spikes. When using an ANN-to-SNN conversion technique there is a direct link between the activation bit precision of the artificial neurons and the time required by the spiking neurons to represent the same bit precision. This implicit link suggests that techniques used to reduce the activation bit precision of ANNs, such as quantization, can help shorten the inference latency of SNNs. However, carrying ANN quantization knowledge over to SNNs is not straightforward, as there are many fundamental differences between them. Here we propose a quantization framework for fast SNNs (QFFS) to overcome these difficulties, providing a method to build SNNs with enhanced latency and reduced loss of accuracy relative to the baseline ANN model. In this framework, we promote the compatibility of ANN information quantization techniques with SNNs, and suppress "occasional noise" to minimize accuracy loss. The resulting SNNs overcome the accuracy degeneration observed previously in SNNs with a limited number of time steps and achieve an accuracy of 70.18% on ImageNet within 8 time steps. This is the first demonstration that SNNs built by ANN-to-SNN conversion can achieve a similar latency to SNNs built by direct training.
Author Li, Chen
Ma, Lei
Furber, Steve
AuthorAffiliation 2 Beijing Academy of Artificial Intelligence , Beijing , China
3 National Biomedical Imaging Center, Peking University , Beijing , China
1 Advanced Processor Technologies (APT) Group, Department of Computer Science, The University of Manchester , Manchester , United Kingdom
AuthorAffiliation_xml – name: 1 Advanced Processor Technologies (APT) Group, Department of Computer Science, The University of Manchester , Manchester , United Kingdom
– name: 2 Beijing Academy of Artificial Intelligence , Beijing , China
– name: 3 National Biomedical Imaging Center, Peking University , Beijing , China
Author_xml – sequence: 1
  givenname: Chen
  surname: Li
  fullname: Li, Chen
– sequence: 2
  givenname: Lei
  surname: Ma
  fullname: Ma, Lei
– sequence: 3
  givenname: Steve
  surname: Furber
  fullname: Furber, Steve
BookMark eNp1kUtvEzEUhS1URB_wA9jNkk3C-G1vkFBFSqUKhFokdtYdP4LbiR1sTyv49UySIlEkVtfyuee71z6n6Cjl5BF6jfslpUq_DSmmuiQ9IUuNldT0GTrBQpAF4_Tb0V_nY3Ra623fC6IYeYGOKddE9RifIPllgtTiL2gxp25VYOMfcrnrQi7dCmrrrrfxLqZ198lPBca5tJ1eX6LnAcbqXz3WM_R19eHm_OPi6vPF5fn7q4VljLTFAOAH4YN1gnBCqLVhACVk4MI5i51jlGjvtAxODg4sxUpZhkEzyYUUjp6hywPXZbg12xI3UH6aDNHsL3JZGygt2tEbLjAB6jjGEFhwA3gt2YD5wK0M8-CZ9e7A2k7DxjvrU5uf9AT6VEnxu1nne6MpY0rpGfDmEVDyj8nXZjaxWj-OkHyeqiFCa9lTzuTcKg-ttuRaiw_Gxrb_5JkcR4N7s4vQ7CM0uwjNIcLZif9x_lnw_57ft3WjuA
CitedBy_id crossref_primary_10_3389_fnins_2023_1224457
crossref_primary_10_3390_electronics13091744
crossref_primary_10_1109_ACCESS_2024_3391889
crossref_primary_10_1109_JETCAS_2023_3328863
crossref_primary_10_1109_TCDS_2023_3329747
crossref_primary_10_3389_fnins_2023_1230002
crossref_primary_10_3389_fnins_2024_1449020
crossref_primary_10_1002_aisy_202300383
crossref_primary_10_1109_TETCI_2024_3359539
crossref_primary_10_3389_fnins_2023_1154241
crossref_primary_10_3389_fnins_2024_1440000
Cites_doi 10.3389/fnins.2019.00095
10.48550/arXiv.2105.12917
10.1109/IJCNN.2015.7280696
10.1109/DAC18074.2021.9586266
10.3389/fnins.2021.629000
10.1016/j.patcog.2020.107281
10.3389/neuro.11.011.2008
10.24963/ijcai.2021/321
10.1109/TPAMI.2021.3114196
10.3389/fnins.2020.00535
10.1145/3477145.3477158
10.48550/arXiv.2102.04159
10.48550/arXiv.1806.08342
10.48550/arXiv.1902.08153
10.48550/arXiv.2008.03658
10.1109/IJCNN52387.2021.9534111
10.48550/arXiv.2106.06984
10.1145/3407197.3407203
10.3389/fnins.2017.00682
10.48550/arXiv.2103.00476
10.1109/IJCNN52387.2021.9533874
ContentType Journal Article
Copyright Copyright © 2022 Li, Ma and Furber.
Copyright © 2022 Li, Ma and Furber. 2022 Li, Ma and Furber
Copyright_xml – notice: Copyright © 2022 Li, Ma and Furber.
– notice: Copyright © 2022 Li, Ma and Furber. 2022 Li, Ma and Furber
DBID AAYXX
CITATION
7X8
5PM
DOA
DOI 10.3389/fnins.2022.918793
DatabaseName CrossRef
MEDLINE - Academic
PubMed Central (Full Participant titles)
DOAJ Directory of Open Access Journals
DatabaseTitle CrossRef
MEDLINE - Academic
DatabaseTitleList
MEDLINE - Academic
CrossRef

Database_xml – sequence: 1
  dbid: DOA
  name: DOAJ Directory of Open Access Journals
  url: https://www.doaj.org/
  sourceTypes: Open Website
DeliveryMethod fulltext_linktorsrc
Discipline Anatomy & Physiology
EISSN 1662-453X
ExternalDocumentID oai_doaj_org_article_5612a3d511af4fdbae974b15b5c7fccf
PMC9344889
10_3389_fnins_2022_918793
GroupedDBID ---
29H
2WC
53G
5GY
5VS
8FE
8FH
9T4
AAFWJ
AAYXX
ABUWG
ACGFO
ACGFS
ACXDI
ADRAZ
AEGXH
AENEX
AFKRA
AFPKN
AIAGR
ALMA_UNASSIGNED_HOLDINGS
AZQEC
BBNVY
BENPR
BHPHI
BPHCQ
CITATION
CS3
DIK
DU5
E3Z
EBS
EJD
EMOBN
F5P
FRP
GROUPED_DOAJ
GX1
HCIFZ
HYE
KQ8
LK8
M2P
M48
M7P
O5R
O5S
OK1
OVT
P2P
PGMZT
PIMPY
PQQKQ
PROAC
RNS
RPM
W2D
7X8
5PM
ID FETCH-LOGICAL-c442t-baaeb6efcd625223ccfba867f56ddc1dd4329ed97fd7bdac3188c41a9475676d3
IEDL.DBID DOA
ISSN 1662-453X
1662-4548
IngestDate Wed Aug 27 01:27:17 EDT 2025
Thu Aug 21 18:16:30 EDT 2025
Fri Jul 11 07:04:31 EDT 2025
Tue Jul 01 01:39:41 EDT 2025
Thu Apr 24 23:09:09 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed true
IsScholarly true
Language English
License This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c442t-baaeb6efcd625223ccfba867f56ddc1dd4329ed97fd7bdac3188c41a9475676d3
Notes ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
This article was submitted to Neuromorphic Engineering, a section of the journal Frontiers in Neuroscience
Reviewed by: Dongsuk Jeon, Seoul National University, South Korea; Yukuan Yang, Tsinghua University, China
Edited by: Guoqi Li, Tsinghua University, China
OpenAccessLink https://doaj.org/article/5612a3d511af4fdbae974b15b5c7fccf
PMID 35928011
PQID 2699703547
PQPubID 23479
ParticipantIDs doaj_primary_oai_doaj_org_article_5612a3d511af4fdbae974b15b5c7fccf
pubmedcentral_primary_oai_pubmedcentral_nih_gov_9344889
proquest_miscellaneous_2699703547
crossref_citationtrail_10_3389_fnins_2022_918793
crossref_primary_10_3389_fnins_2022_918793
ProviderPackageCode CITATION
AAYXX
PublicationCentury 2000
PublicationDate 2022-07-19
PublicationDateYYYYMMDD 2022-07-19
PublicationDate_xml – month: 07
  year: 2022
  text: 2022-07-19
  day: 19
PublicationDecade 2020
PublicationTitle Frontiers in neuroscience
PublicationYear 2022
Publisher Frontiers Media S.A
Publisher_xml – name: Frontiers Media S.A
References Sengupta (B23) 2019; 13
B25
Hwang (B11) 2021; 15
Esser (B7) 2019
Ho (B10) 2020
Mueller (B18) 2021
Schaefer (B22) 2020
Qin (B19) 2020; 105
Diehl (B5) 2015
Fang (B8) 2021
Chowdhury (B2) 2021
Rathi (B20) 2020
Li (B15)
Deng (B4) 2021
Bu (B1) 2021
Li (B13)
Ding (B6) 2021
Han (B9) 2020
Lu (B16) 2020; 14
Davison (B3) 2009; 2
Warden (B24) 2019
Krishnamoorthi (B12) 2018
Rueckauer (B21) 2017; 11
Lui (B17) 2021
Li (B14)
References_xml – volume: 13
  start-page: 95
  year: 2019
  ident: B23
  article-title: Going deeper in spiking neural networks: VGG and residual architectures
  publication-title: Front. Neurosci
  doi: 10.3389/fnins.2019.00095
– ident: B15
  article-title: Bsnn: towards faster and better conversion of artificial neural networks to spiking neural networks with bistable neurons
  publication-title: arXiv preprint arXiv:2105.12917
  doi: 10.48550/arXiv.2105.12917
– start-page: 1
  volume-title: 2015 International Joint Conference on Neural Networks (IJCNN)
  year: 2015
  ident: B5
  article-title: Fast-classifying, high-accuracy spiking deep networks through weight and threshold balancing
  doi: 10.1109/IJCNN.2015.7280696
– year: 2020
  ident: B10
  article-title: TCL: an ann-to-snn conversion with trainable clipping layers
  publication-title: arXiv preprint arXiv:2008.04509
  doi: 10.1109/DAC18074.2021.9586266
– volume: 15
  start-page: 135
  year: 2021
  ident: B11
  article-title: Low-latency spiking neural networks using pre-charged membrane potential and delayed evaluation
  publication-title: Front. Neurosci
  doi: 10.3389/fnins.2021.629000
– volume: 105
  start-page: 107281
  year: 2020
  ident: B19
  article-title: Binary neural networks: a survey
  publication-title: Pattern Recognit
  doi: 10.1016/j.patcog.2020.107281
– volume: 2
  start-page: 11
  year: 2009
  ident: B3
  article-title: Pynn: a common interface for neuronal network simulators
  publication-title: Front Neuroinform
  doi: 10.3389/neuro.11.011.2008
– start-page: 13558
  volume-title: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
  year: 2020
  ident: B9
  article-title: RMP-SNN: residual membrane potential neuron for enabling deeper high-accuracy and low-latency spiking neural network
– year: 2021
  ident: B6
  article-title: Optimal ann-snn conversion for fast and accurate inference in deep spiking neural networks
  publication-title: arXiv preprint arXiv:2105.11654
  doi: 10.24963/ijcai.2021/321
– year: 2021
  ident: B1
  article-title: Optimal ann-snn conversion for high-accuracy and ultra-low-latency spiking neural networks
  publication-title: International Conference on Learning Representations
– ident: B25
  doi: 10.1109/TPAMI.2021.3114196
– volume: 14
  start-page: 535
  year: 2020
  ident: B16
  article-title: Exploring the connection between binary and spiking neural networks
  publication-title: Front. Neurosci
  doi: 10.3389/fnins.2020.00535
– year: 2021
  ident: B17
  article-title: Hessian aware quantization of spiking neural networks
  publication-title: arXiv preprint arXiv:2104.14117
  doi: 10.1145/3477145.3477158
– year: 2021
  ident: B8
  article-title: Deep residual learning in spiking neural networks
  publication-title: arXiv preprint arXiv:2102.04159
  doi: 10.48550/arXiv.2102.04159
– year: 2018
  ident: B12
  article-title: Quantizing deep convolutional networks for efficient inference: a whitepaper
  publication-title: arXiv preprint arXiv:1806.08342
  doi: 10.48550/arXiv.1806.08342
– year: 2019
  ident: B7
  article-title: Learned step size quantization
  publication-title: arXiv preprint arXiv:1902.08153
  doi: 10.48550/arXiv.1902.08153
– year: 2020
  ident: B20
  article-title: Diet-snn: direct input encoding with leakage and threshold optimization in deep spiking neural networks
  publication-title: arXiv preprint arXiv:2008.03658
  doi: 10.48550/arXiv.2008.03658
– year: 2021
  ident: B2
  article-title: Spatio-temporal pruning and quantization for low-latency spiking neural networks
  publication-title: arXiv preprint arXiv:2104.12528
  doi: 10.1109/IJCNN52387.2021.9534111
– ident: B13
  article-title: A free lunch from ann: towards efficient, accurate spiking neural networks calibration
  publication-title: arXiv preprint arXiv:2106.06984
  doi: 10.48550/arXiv.2106.06984
– start-page: 1
  year: 2020
  ident: B22
  article-title: Quantizing spiking neural networks with integers
  publication-title: International Conference on Neuromorphic Systems 2020
  doi: 10.1145/3407197.3407203
– volume: 11
  start-page: 682
  year: 2017
  ident: B21
  article-title: Conversion of continuous-valued deep networks to efficient event-driven networks for image classification
  publication-title: Front. Neurosci
  doi: 10.3389/fnins.2017.00682
– year: 2021
  ident: B4
  article-title: Optimal conversion of conventional artificial neural networks to spiking neural networks
  publication-title: arXiv preprint arXiv:2103.00476
  doi: 10.48550/arXiv.2103.00476
– start-page: 1
  volume-title: 2021 International Joint Conference on Neural Networks (IJCNN)
  year: 2021
  ident: B18
  article-title: Minimizing inference time: optimization methods for converted deep spiking neural networks
  doi: 10.1109/IJCNN52387.2021.9533874
– volume-title: Tinyml: Machine Learning With Tensorflow Lite on Arduino and Ultra-Low-Power Microcontrollers
  year: 2019
  ident: B24
– start-page: 34
  ident: B14
  article-title: Differentiable spike: rethinking gradient-descent for training spiking neural networks
  publication-title: Advances in Neural Information Processing Systems
SSID ssj0062842
Score 2.5502822
Snippet Compared with artificial neural networks (ANNs), spiking neural networks (SNNs) offer additional temporal dynamics with the compromise of lower information...
SourceID doaj
pubmedcentral
proquest
crossref
SourceType Open Website
Open Access Repository
Aggregation Database
Enrichment Source
Index Database
StartPage 918793
SubjectTerms ANN-to-SNN conversion
fast spiking neural networks
inference latency
Neuroscience
occasional noise
quantization
spiking neural networks
SummonAdditionalLinks – databaseName: Scholars Portal Journals: Open Access
  dbid: M48
  link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1La9tAEB4S99JLaJuWuGnDFkoPBSW29qU9lOKUmlBIIKSG3MQ-E4MrJ35A8-87I8mmgpBTrpJWq_1GszPf7uwMwGf0KAbRFDHz6HtkIqBKOTHwJBCrOHdO1cfFzi_U2UT8upbXO7Apb9UCuHyU2lE9qclidvz3_uE7Kvw3Ypxob09SNa0o83aeHxsqns134QUaJk16ei62mwoKZ-K82dh8vFnHNNUZ_DtuZzdo8j8rNH4Fe637yEaNvF_DTqzewP6oQur854F9YXVAZ71Svg_6co2wtecs2XgThcXQTWVju1yxq7spLZQzStCBL71oIsKXb2Ey_vn7x1nW1knIvBD5KnPWRqdi8gHJDJp775OzhdJJqhD8MATBcxOD0SloF6xHNS68GFojtFRaBf4OetW8igfAch1ToEKgruDCGdRuzX1hpUwiDpKSfRhsICp9m0ScalnMSiQThGpZo1oSqmWDah--bpvcNRk0nnr4lHDfPkjJr-sL88VN2epSSQU9LQ_oKtokUnA2IilyQ-mk1wkH34dPG6mVqCy0A2KrOF9jT8oYnOKk0H3QHXF2euzeqaa3ddptw5HKFub9c3ziIbykUdMi8dB8gN5qsY4f0btZuaP6n_0HFAX9dQ
  priority: 102
  providerName: Scholars Portal
Title Quantization Framework for Fast Spiking Neural Networks
URI https://www.proquest.com/docview/2699703547
https://pubmed.ncbi.nlm.nih.gov/PMC9344889
https://doaj.org/article/5612a3d511af4fdbae974b15b5c7fccf
Volume 16
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1LS8NAEF7EkxfxifXFCuJBiG2yr-xRxVoEBdFCb2GfWNC02Pbgv3cmSaW56MXLHpJNdvNN5rU7O0PIOVgUvaDzkDiwPRLugaUs7zkkiJGMWSur42KPT3Iw5A8jMVop9YUxYXV64Bq4LlZvNMyDXWAij96aABawTYUVTkXnIkpf0HlLZ6qWwRKEblbvYYILpruxHJeYmzvLrjSW12YtLVQl629ZmO34yBWF098im42lSK_rGW6TtVDukN3rErzkjy96QavYzWpRfJeo5wUg1ByppP1lwBUFi5T2zWxOX6ZjXBOnmIsDXvpUB3_P9siwf_d6O0iakgiJ4zybJ9aYYGWIzoPfApodALAmlyoK6b1Lvecs08FrFb2y3jjg2Nzx1GiuhFTSs32yXk7KcEBopkL0WPPT5oxbDYysmMuNEJGHXpSiQ3pLiArX5AvHshXvBfgNiGpRoVogqkWNaodc_jwyrZNl_Nb5BnH_6Yh5rqsLQP2ioX7xF_U75GxJtQL4Ajc7TBkmCxhJag3STHDVIapFztaI7Tvl-K3KsK0ZeK25PvyPKR6RDfxqXA9O9TFZn38uwgkYMnN7Wv2z0N6PUmgfef4N5uf3kw
linkProvider Directory of Open Access Journals
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Quantization+Framework+for+Fast+Spiking+Neural+Networks&rft.jtitle=Frontiers+in+neuroscience&rft.au=Chen+Li&rft.au=Lei+Ma&rft.au=Lei+Ma&rft.au=Steve+Furber&rft.date=2022-07-19&rft.pub=Frontiers+Media+S.A&rft.eissn=1662-453X&rft.volume=16&rft_id=info:doi/10.3389%2Ffnins.2022.918793&rft.externalDBID=DOA&rft.externalDocID=oai_doaj_org_article_5612a3d511af4fdbae974b15b5c7fccf
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1662-453X&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1662-453X&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1662-453X&client=summon