SpinalFlow: An Architecture and Dataflow Tailored for Spiking Neural Networks

Spiking neural networks (SNNs) are expected to be part of the future AI portfolio, with heavy investment from industry and government, e.g., IBM TrueNorth, Intel Loihi. While Artificial Neural Network (ANN) architectures have taken large strides, few works have targeted SNN hardware efficiency. Our...

Full description

Saved in:
Bibliographic Details
Published in2020 ACM/IEEE 47th Annual International Symposium on Computer Architecture (ISCA) pp. 349 - 362
Main Authors Narayanan, Surya, Taht, Karl, Balasubramonian, Rajeev, Giacomin, Edouard, Gaillardon, Pierre-Emmanuel
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.05.2020
Subjects
Online AccessGet full text
DOI10.1109/ISCA45697.2020.00038

Cover

Loading…
Abstract Spiking neural networks (SNNs) are expected to be part of the future AI portfolio, with heavy investment from industry and government, e.g., IBM TrueNorth, Intel Loihi. While Artificial Neural Network (ANN) architectures have taken large strides, few works have targeted SNN hardware efficiency. Our analysis of SNN baselines shows that at modest spike rates, SNN implementations exhibit significantly lower efficiency than accelerators for ANNs. This is primarily because SNN dataflows must consider neuron potentials for several ticks, introducing a new data structure and a new dimension to the reuse pattern. We introduce a novel SNN architecture, SpinalFlow, that processes a compressed, time-stamped, sorted sequence of input spikes. It adopts an ordering of computations such that the outputs of a network layer are also compressed, time-stamped, and sorted. All relevant computations for a neuron are performed in consecutive steps to eliminate neuron potential storage overheads. Thus, with better data reuse, we advance the energy efficiency of SNN accelerators by an order of magnitude. Even though the temporal aspect in SNNs prevents the exploitation of some reuse patterns that are more easily exploited in ANNs, at 4-bit input resolution and 90% input sparsity, SpinalFlow reduces average energy by 1.8 \times, compared to a 4-bit Eyeriss baseline. These improvements are seen for a range of networks and sparsity/resolution levels; SpinalFlow consumes 5 \times less energy and 5.4 \times less time than an 8-bit version of Eyeriss. We thus show that, depending on the level of observed sparsity, SNN architectures can be competitive with ANN architectures in terms of latency and energy for inference, thus lowering the barrier for practical deployment in scenarios demanding real-time learning.
AbstractList Spiking neural networks (SNNs) are expected to be part of the future AI portfolio, with heavy investment from industry and government, e.g., IBM TrueNorth, Intel Loihi. While Artificial Neural Network (ANN) architectures have taken large strides, few works have targeted SNN hardware efficiency. Our analysis of SNN baselines shows that at modest spike rates, SNN implementations exhibit significantly lower efficiency than accelerators for ANNs. This is primarily because SNN dataflows must consider neuron potentials for several ticks, introducing a new data structure and a new dimension to the reuse pattern. We introduce a novel SNN architecture, SpinalFlow, that processes a compressed, time-stamped, sorted sequence of input spikes. It adopts an ordering of computations such that the outputs of a network layer are also compressed, time-stamped, and sorted. All relevant computations for a neuron are performed in consecutive steps to eliminate neuron potential storage overheads. Thus, with better data reuse, we advance the energy efficiency of SNN accelerators by an order of magnitude. Even though the temporal aspect in SNNs prevents the exploitation of some reuse patterns that are more easily exploited in ANNs, at 4-bit input resolution and 90% input sparsity, SpinalFlow reduces average energy by 1.8 \times, compared to a 4-bit Eyeriss baseline. These improvements are seen for a range of networks and sparsity/resolution levels; SpinalFlow consumes 5 \times less energy and 5.4 \times less time than an 8-bit version of Eyeriss. We thus show that, depending on the level of observed sparsity, SNN architectures can be competitive with ANN architectures in terms of latency and energy for inference, thus lowering the barrier for practical deployment in scenarios demanding real-time learning.
Author Taht, Karl
Narayanan, Surya
Giacomin, Edouard
Balasubramonian, Rajeev
Gaillardon, Pierre-Emmanuel
Author_xml – sequence: 1
  givenname: Surya
  surname: Narayanan
  fullname: Narayanan, Surya
  organization: University of Utah,School of Computing,Salt Lake City,USA
– sequence: 2
  givenname: Karl
  surname: Taht
  fullname: Taht, Karl
  organization: University of Utah,School of Computing,Salt Lake City,USA
– sequence: 3
  givenname: Rajeev
  surname: Balasubramonian
  fullname: Balasubramonian, Rajeev
  organization: University of Utah,School of Computing,Salt Lake City,USA
– sequence: 4
  givenname: Edouard
  surname: Giacomin
  fullname: Giacomin, Edouard
  organization: University of Utah,Electrical and Computer Engineering,Salt Lake City,USA
– sequence: 5
  givenname: Pierre-Emmanuel
  surname: Gaillardon
  fullname: Gaillardon, Pierre-Emmanuel
  organization: University of Utah,Electrical and Computer Engineering,Salt Lake City,USA
BookMark eNotjEtOwzAUAI0EErT0BLDwBRKe7fjHLgoUKhVYtKwrJ34GqyGpnFQVt28kWM1iNDMjl13fISH3DHLGwD6sNlVZSGV1zoFDDgDCXJAZ09ywQikmr8liGGINBZPCcMZuyNvmEDvXLtv-9EjLjpap-Y4jNuMxIXWdp09udGGydOti2yf0NPSJTtU-dl_0HY_JtRPGU5_2wy25Cq4dcPHPOflcPm-r12z98bKqynXmuJFjxtDZ4EGrxjcGpamDssARwGvwXCM3gikuOEqhhFEctfboG4cheIe1EXNy9_eNiLg7pPjj0u_OMmEsV-IMwi1PRw
ContentType Conference Proceeding
DBID 6IE
6IH
CBEJK
RIE
RIO
DOI 10.1109/ISCA45697.2020.00038
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan (POP) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Xplore Digital Library
IEEE Proceedings Order Plans (POP) 1998-present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Xplore Digital Library
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 1728146615
9781728146614
EndPage 362
ExternalDocumentID 9138926
Genre orig-research
GroupedDBID 6IE
6IH
ACM
ALMA_UNASSIGNED_HOLDINGS
APO
CBEJK
GUFHI
LHSKQ
RIE
RIO
ID FETCH-LOGICAL-a285t-1ea9fd076cdc8e58bf6902e00d70d27e28316232e5363862e77dedcaeffdaeb83
IEDL.DBID RIE
IngestDate Wed Aug 06 17:54:10 EDT 2025
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a285t-1ea9fd076cdc8e58bf6902e00d70d27e28316232e5363862e77dedcaeffdaeb83
PageCount 14
ParticipantIDs ieee_primary_9138926
PublicationCentury 2000
PublicationDate 2020-May
PublicationDateYYYYMMDD 2020-05-01
PublicationDate_xml – month: 05
  year: 2020
  text: 2020-May
PublicationDecade 2020
PublicationTitle 2020 ACM/IEEE 47th Annual International Symposium on Computer Architecture (ISCA)
PublicationTitleAbbrev ISCA
PublicationYear 2020
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssib041538211
Score 2.4838188
Snippet Spiking neural networks (SNNs) are expected to be part of the future AI portfolio, with heavy investment from industry and government, e.g., IBM TrueNorth,...
SourceID ieee
SourceType Publisher
StartPage 349
SubjectTerms Accelerators
CNNs
Computer architecture
Energy efficiency
Energy resolution
Network topology
Neurons
Portfolios
Real-time systems
SNNs
Spiking neural networks
Throughput
Training
Title SpinalFlow: An Architecture and Dataflow Tailored for Spiking Neural Networks
URI https://ieeexplore.ieee.org/document/9138926
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1La8JAEF7UU09t0dI3e-ix0TWvzfYmtmILloIK3mSzOwulkoiNCP31nUnUSumhp4RA2LCTZb55fPMxdmco4x4HgWdo_mWILsVTJjSeTJ0gCQ4py6E-o9d4OA1fZtGsxu73XBgAKJvPoE23ZS3f5mZNqbKOoqqaH9dZHQO3iqu1-3dCOrkYzGzZcV2hOs_jfg_hgZIYBfrUwCWIhHKgoVK6kMExG-0WrzpHPtrrIm2br19zGf_7dSes9UPW4297N3TKapA12Wi8JLmrwSLfPPBexnsH9QKuM8sfNen25hs-0e-0kuWIXjm-RalzTiM79AIvZY_4Z4tNB0-T_tDbKid42k-iwuuCVs4KGRtrEoiS1GEQ7IMQVgrrS0BM0UXc40MU4PmLfZDSgjUanLMa0iQ4Y40sz-CccR0EkXKI47QJQ2s1IojECbRiCi40Ul2wJm3FfFkNx5hvd-Hy78dX7IiMUXUMXrNGsVrDDXr1Ir0tzfkNnMeipQ
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1NS8NAEF1qPehJpRW_3YNH027ztYm3Ui2tNkVoC72Vze4ExJIUTSn4651J2lrEg6eEQNiwk2XezLyZx9idpoy77ziWpvmXLroUK9SutmScCJLgkLIY6hMN_d7EfZ560wq73_bCAEBBPoMG3Ra1fJPpJaXKmiFV1Wx_j-2j3_daZbfW5u9x6exiOLPuj2uJsNkfddoIEEKJcaBNFC5BbSg7KiqFE-kesWizfMkdeW8s87ihv35NZvzv9x2z-k-7Hn_dOqITVoG0xqLRggSvuvNs9cDbKW_vVAy4Sg1_VKTcm634WL3RSoYjfuX4FiXPOQ3tUHO8FCzxzzqbdJ_GnZ611k6wlB14udUCFSZGSF8bHYAXxAmGwTYIYaQwtgREFS1EPjZ4Dp5A3wYpDRitIEmMgjhwTlk1zVI4Y1w5jhcmiOSUdl1jFGKIIBFoxxgSV8vwnNVoK2aLcjzGbL0LF38_vmUHvXE0mA36w5dLdkiGKfmDV6yafyzhGn18Ht8Upv0GSPul7g
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2020+ACM%2FIEEE+47th+Annual+International+Symposium+on+Computer+Architecture+%28ISCA%29&rft.atitle=SpinalFlow%3A+An+Architecture+and+Dataflow+Tailored+for+Spiking+Neural+Networks&rft.au=Narayanan%2C+Surya&rft.au=Taht%2C+Karl&rft.au=Balasubramonian%2C+Rajeev&rft.au=Giacomin%2C+Edouard&rft.date=2020-05-01&rft.pub=IEEE&rft.spage=349&rft.epage=362&rft_id=info:doi/10.1109%2FISCA45697.2020.00038&rft.externalDocID=9138926