Revisiting Neuron Coverage Metrics and Quality of Deep Neural Networks

Deep neural networks (DNN) have been widely applied in modern life, including critical domains like autonomous driving, making it essential to ensure the reliability and robustness of DNN-powered systems. As an analogy to code coverage metrics for testing conventional software, researchers have prop...

Full description

Saved in:
Bibliographic Details
Published in2022 IEEE International Conference on Software Analysis, Evolution and Reengineering (SANER) pp. 408 - 419
Main Authors Yang, Zhou, Shi, Jieke, Asyrofi, Muhammad Hilmi, Lo, David
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.03.2022
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Deep neural networks (DNN) have been widely applied in modern life, including critical domains like autonomous driving, making it essential to ensure the reliability and robustness of DNN-powered systems. As an analogy to code coverage metrics for testing conventional software, researchers have proposed neuron coverage metrics and coverage-driven methods to generate DNN test cases. However, Yan et al. doubt the usefulness of existing coverage criteria in DNN testing. They show that a coverage-driven method is less effective than a gradient-based method in terms of both uncovering defects and improving model robustness. In this paper, we conduct a replication study of the work by Yan et al. and extend the experiments for deeper analysis. A larger model and a dataset of higher resolution images are included to examine the generalizability of the results. We also extend the experiments with more test case generation techniques and adjust the process of improving model robustness to be closer to the practical life cycle of DNN development. Our experiment results confirm the conclusion from Yan et al. that coverage-driven methods are less effective than gradient-based methods. Yan et al. find that using gradient-based methods to retrain cannot repair defects uncovered by coverage-driven methods. They attribute this to the fact that the two types of methods use different perturbation strategies: gradient-based methods perform differentiable transformations while coverage-driven methods can perform additional non-differentiable transformations. We test several hypotheses and further show that even coverage-driven methods are constrained only to perform differentiable transformations, the uncovered defects still cannot be repaired by adversarial training with gradient-based methods. Thus, defensive strategies for coverage-driven methods should be further studied.
AbstractList Deep neural networks (DNN) have been widely applied in modern life, including critical domains like autonomous driving, making it essential to ensure the reliability and robustness of DNN-powered systems. As an analogy to code coverage metrics for testing conventional software, researchers have proposed neuron coverage metrics and coverage-driven methods to generate DNN test cases. However, Yan et al. doubt the usefulness of existing coverage criteria in DNN testing. They show that a coverage-driven method is less effective than a gradient-based method in terms of both uncovering defects and improving model robustness. In this paper, we conduct a replication study of the work by Yan et al. and extend the experiments for deeper analysis. A larger model and a dataset of higher resolution images are included to examine the generalizability of the results. We also extend the experiments with more test case generation techniques and adjust the process of improving model robustness to be closer to the practical life cycle of DNN development. Our experiment results confirm the conclusion from Yan et al. that coverage-driven methods are less effective than gradient-based methods. Yan et al. find that using gradient-based methods to retrain cannot repair defects uncovered by coverage-driven methods. They attribute this to the fact that the two types of methods use different perturbation strategies: gradient-based methods perform differentiable transformations while coverage-driven methods can perform additional non-differentiable transformations. We test several hypotheses and further show that even coverage-driven methods are constrained only to perform differentiable transformations, the uncovered defects still cannot be repaired by adversarial training with gradient-based methods. Thus, defensive strategies for coverage-driven methods should be further studied.
Author Shi, Jieke
Lo, David
Yang, Zhou
Asyrofi, Muhammad Hilmi
Author_xml – sequence: 1
  givenname: Zhou
  surname: Yang
  fullname: Yang, Zhou
  email: zyang@smu.edu.sg
  organization: School of Computing and Information Systems, Singapore Management University,Singapore
– sequence: 2
  givenname: Jieke
  surname: Shi
  fullname: Shi, Jieke
  email: jiekeshi@smu.edu.sg
  organization: School of Computing and Information Systems, Singapore Management University,Singapore
– sequence: 3
  givenname: Muhammad Hilmi
  surname: Asyrofi
  fullname: Asyrofi, Muhammad Hilmi
  email: mhilmia@smu.edu.sg
  organization: School of Computing and Information Systems, Singapore Management University,Singapore
– sequence: 4
  givenname: David
  surname: Lo
  fullname: Lo, David
  email: davidlo@smu.edu.sg
  organization: School of Computing and Information Systems, Singapore Management University,Singapore
BookMark eNotzNFKwzAUgOEIeuGmTyBCXqA1OWlykstRNyfMiVOvR9qcjmBtR9pN9vaKevXd_PwTdt71HTF2K0UupXB3r7P1fKNVoSAHAZALIbQ5YxNpjC4UWqMu2WJDxzjEMXY7vqZD6jte9kdKfkf8icYU64H7LvCXg2_jeOJ9w--J9r-tb38Yv_r0MVyxi8a3A13_O2Xvi_lbucxWzw-P5WyVRRBqzOoKigbIe-NdXQQ0waNBKQmsrYIB2zhTY1U0AqsgLTqoPToKIZAGiaCm7ObvG4lou0_x06fT1lnQiFp9AzRNSW8
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1109/SANER53432.2022.00056
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Xplore POP ALL
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 1665437863
9781665437868
EndPage 419
ExternalDocumentID 9825775
Genre orig-research
GrantInformation_xml – fundername: Singapore Ministry of Education
  funderid: 10.13039/501100001459
GroupedDBID 6IE
6IL
CBEJK
RIE
RIL
ID FETCH-LOGICAL-i203t-cb24f2eaa6a9c4d76da76711e288bd628f96c7b4f07bd18792ca79eddde521723
IEDL.DBID RIE
IngestDate Thu Jun 29 18:36:50 EDT 2023
IsDoiOpenAccess false
IsOpenAccess true
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i203t-cb24f2eaa6a9c4d76da76711e288bd628f96c7b4f07bd18792ca79eddde521723
PageCount 12
ParticipantIDs ieee_primary_9825775
PublicationCentury 2000
PublicationDate 2022-March
PublicationDateYYYYMMDD 2022-03-01
PublicationDate_xml – month: 03
  year: 2022
  text: 2022-March
PublicationDecade 2020
PublicationTitle 2022 IEEE International Conference on Software Analysis, Evolution and Reengineering (SANER)
PublicationTitleAbbrev SANER
PublicationYear 2022
Publisher IEEE
Publisher_xml – name: IEEE
Score 2.3007984
Snippet Deep neural networks (DNN) have been widely applied in modern life, including critical domains like autonomous driving, making it essential to ensure the...
SourceID ieee
SourceType Publisher
StartPage 408
SubjectTerms Coverage-Driven Testing
Deep learning
Deep Learning Testing
Measurement
Neural networks
Neurons
Perturbation methods
Software
Software Quality
Training
Title Revisiting Neuron Coverage Metrics and Quality of Deep Neural Networks
URI https://ieeexplore.ieee.org/document/9825775
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3NS8MwFH_MnTypbOI3OXi0W5ulSXOUuTGEDZkOdhv5eAFRuuG6g_71JumcIh48NZRC2j7IL6_9fQBcKyGtBx6VGKnzhDnLE80kJrzwG36P6K5WpY0nfDRj9_N83oCbnRYGESP5DDthGP_l26XZhE9lXenbGSHyPdjzjVut1dqKcrJUdh9vJ4NpHoSSvu2j0YczxFL_CE2JmDE8gPHXbDVV5KWzqXTHfPwyYvzv7RxC-1udRx52uHMEDSxbMJxGmXggMZNouFGSfmBn-uWCjENqllkTVVpSe2a8k6Ujd4ireK169YfIBl-3YTYcPPVHyTYjIXmmaa9KjKbMUVSKK2mYFdwqwUWWIS0KbTktnORGaOZSoW1IFqfGlwetX9XykE3VO4ZmuSzxBIgzmgttuHWK-7YJC5Nm1u-enOxZywp9Cq3wDhar2gZjsX38s79Pn8N-qEJN17qAZvW2wUuP35W-ioX7BJr8nYk
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1NT8IwGH6DeNCTGjB-24NHB1vZ2vVoEILKiEFIuJF-JkazERkH_fW2HaIxHjytWZZ025v06bs9HwBXnDJlgYcHkokkiI0igYiZDkhqN_wW0U2lSstGZDCN72fJrAbXGy2M1tqTz3TLDf2_fFXIlftU1ma2naE02YJti_sJrtRaa1lOFLL2082oN06cVNI2ftg7cbpg6h-xKR41-nuQfc1XkUVeWqtStOTHLyvG_97QPjS_9XnocYM8B1DTeQP6Yy8UdzRm5C03ctR1_Ey7YKDM5WbJJeK5QpVrxjsqDLrVeuGv5a_24PngyyZM-71JdxCsUxKCZxx2ykAKHBusOSecyVhRojglNIo0TlOhCE4NI5KK2IRUKJctjqUtkFZ2XUtcOlXnEOp5kesjQEYKQoUkynBiGyedyjBSdv9kWEepOBXH0HDvYL6ojDDm68c_-fv0JewMJtlwPrwbPZzCrqtIRd46g3r5ttLnFs1LceGL-AkEeqDT
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2022+IEEE+International+Conference+on+Software+Analysis%2C+Evolution+and+Reengineering+%28SANER%29&rft.atitle=Revisiting+Neuron+Coverage+Metrics+and+Quality+of+Deep+Neural+Networks&rft.au=Yang%2C+Zhou&rft.au=Shi%2C+Jieke&rft.au=Asyrofi%2C+Muhammad+Hilmi&rft.au=Lo%2C+David&rft.date=2022-03-01&rft.pub=IEEE&rft.spage=408&rft.epage=419&rft_id=info:doi/10.1109%2FSANER53432.2022.00056&rft.externalDocID=9825775