Bridging the Gap between Real-world and Synthetic Images for Testing Autonomous Driving Systems

Deep Neural Networks (DNNs) for Autonomous Driving Systems (ADS) are typically trained on real-world images and tested using synthetic images from simulators. This approach results in training and test datasets with dissimilar distributions, which can potentially lead to erroneously decreased test a...

Full description

Saved in:
Bibliographic Details
Published inIEEE/ACM International Conference on Automated Software Engineering : [proceedings] pp. 732 - 744
Main Authors Amini, Mohammad Hossein, Nejati, Shiva
Format Conference Proceeding
LanguageEnglish
Published ACM 27.10.2024
Subjects
Online AccessGet full text
ISSN2643-1572
DOI10.1145/3691620.3695067

Cover

Abstract Deep Neural Networks (DNNs) for Autonomous Driving Systems (ADS) are typically trained on real-world images and tested using synthetic images from simulators. This approach results in training and test datasets with dissimilar distributions, which can potentially lead to erroneously decreased test accuracy. To address this issue, the literature suggests applying domain-to-domain translators to test datasets to bring them closer to the training datasets. However, translating images used for testing may unpredictably affect the reliability, effectiveness and efficiency of the testing process. Hence, this paper investigates the following questions in the context of ADS: Could translators reduce the effectiveness of images used for ADS-DNN testing and their ability to reveal faults in ADS-DNNs? Can translators result in excessive time overhead during simulation-based testing? To address these questions, we consider three domain-to-domain translators: CycleGAN and neural style transfer, from the literature, and SAEVAE, our proposed translator. Our results for two critical ADS tasks - lane keeping and object detection - indicate that translators significantly narrow the gap in ADS test accuracy caused by distribution dissimilarities between training and test data, with SAEVAE outperforming the other two translators. We show that, based on the recent diversity, coverage, and fault-revealing ability metrics for testing deep-learning systems, translators do not compromise the diversity and the coverage of test data nor do they lead to revealing fewer faults in ADS-DNNs. Further, among the translators considered, SAEVAE incurs a negligible overhead in simulation time and can be efficiently integrated into simulation-based testing. Finally, we show that translators increase the correlation between offline and simulation-based testing results, which can help reduce the cost of simulation-based testing. Our replication package is available online [1].CCS CONCEPTS* Software and its engineering → Empirical software validation; * Computing methodologies → Machine learning algorithms.
AbstractList Deep Neural Networks (DNNs) for Autonomous Driving Systems (ADS) are typically trained on real-world images and tested using synthetic images from simulators. This approach results in training and test datasets with dissimilar distributions, which can potentially lead to erroneously decreased test accuracy. To address this issue, the literature suggests applying domain-to-domain translators to test datasets to bring them closer to the training datasets. However, translating images used for testing may unpredictably affect the reliability, effectiveness and efficiency of the testing process. Hence, this paper investigates the following questions in the context of ADS: Could translators reduce the effectiveness of images used for ADS-DNN testing and their ability to reveal faults in ADS-DNNs? Can translators result in excessive time overhead during simulation-based testing? To address these questions, we consider three domain-to-domain translators: CycleGAN and neural style transfer, from the literature, and SAEVAE, our proposed translator. Our results for two critical ADS tasks - lane keeping and object detection - indicate that translators significantly narrow the gap in ADS test accuracy caused by distribution dissimilarities between training and test data, with SAEVAE outperforming the other two translators. We show that, based on the recent diversity, coverage, and fault-revealing ability metrics for testing deep-learning systems, translators do not compromise the diversity and the coverage of test data nor do they lead to revealing fewer faults in ADS-DNNs. Further, among the translators considered, SAEVAE incurs a negligible overhead in simulation time and can be efficiently integrated into simulation-based testing. Finally, we show that translators increase the correlation between offline and simulation-based testing results, which can help reduce the cost of simulation-based testing. Our replication package is available online [1].CCS CONCEPTS* Software and its engineering → Empirical software validation; * Computing methodologies → Machine learning algorithms.
Author Nejati, Shiva
Amini, Mohammad Hossein
Author_xml – sequence: 1
  givenname: Mohammad Hossein
  surname: Amini
  fullname: Amini, Mohammad Hossein
  email: mh.amini@uottawa.ca
  organization: University of Ottawa,Ottawa,Canada
– sequence: 2
  givenname: Shiva
  surname: Nejati
  fullname: Nejati, Shiva
  email: snejati@uottawa.ca
  organization: University of Ottawa,Ottawa,Canada
BookMark eNotjk1PAjEYhKvRRETOXjz0Dyy-7faDHhEVSUhMZO-k276LTdgu2RYJ_94lenomk5nJ3JOb2EUk5JHBlDEhn0tlmOIwHShB6SsyMdrMBIBmXMz0NRlxJcqCSc3vyCSlUMMgpWJMjcj2pQ9-F-KO5m-kS3ugNeYTYqRfaPfFqev3ntro6eYch0QOjq5au8NEm66nFaZ86c6PuYtd2x0Tfe3Dz8XanFPGNj2Q28buE07-OSbV-1u1-CjWn8vVYr4u7PAxF7zh4LWwXtZcKCOgVrqGUggQM3C2EVIiNKVx3HnnGNeD5xmTTnvvgZdj8vQ3GxBxe-hDa_vzloFWwhhW_gLz9lap
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1145/3691620.3695067
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Xplore POP ALL
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Computer Science
EISBN 9798400712487
EISSN 2643-1572
EndPage 744
ExternalDocumentID 10764991
Genre orig-research
GroupedDBID 6IE
6IF
6IH
6IK
6IL
6IM
6IN
6J9
AAJGR
AAWTH
ABLEC
ACREN
ADYOE
ADZIZ
AFYQB
ALMA_UNASSIGNED_HOLDINGS
AMTXH
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
CHZPO
IEGSK
IPLJI
M43
OCL
RIE
RIL
ID FETCH-LOGICAL-a248t-2f20d74ad5b246940b67b03440480caf455e0f39c2cdcc127cafd115c7ddd023
IEDL.DBID RIE
IngestDate Wed Jan 15 06:20:43 EST 2025
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a248t-2f20d74ad5b246940b67b03440480caf455e0f39c2cdcc127cafd115c7ddd023
PageCount 13
ParticipantIDs ieee_primary_10764991
PublicationCentury 2000
PublicationDate 2024-Oct.-27
PublicationDateYYYYMMDD 2024-10-27
PublicationDate_xml – month: 10
  year: 2024
  text: 2024-Oct.-27
  day: 27
PublicationDecade 2020
PublicationTitle IEEE/ACM International Conference on Automated Software Engineering : [proceedings]
PublicationTitleAbbrev ASE
PublicationYear 2024
Publisher ACM
Publisher_xml – name: ACM
SSID ssib057256116
ssj0051577
Score 2.285419
Snippet Deep Neural Networks (DNNs) for Autonomous Driving Systems (ADS) are typically trained on real-world images and tested using synthetic images from simulators....
SourceID ieee
SourceType Publisher
StartPage 732
SubjectTerms Accuracy
Autonomous driving systems (ADS)
Autonomous vehicles
Correlation
Deep learning
Generative adversarial networks
Image-to-image translation
Measurement
Object detection
Online testing
Software
Software engineering
Software reliability
Testing
Training
Title Bridging the Gap between Real-world and Synthetic Images for Testing Autonomous Driving Systems
URI https://ieeexplore.ieee.org/document/10764991
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1JSwMxFA7ak6e6VNzJwWtqJmvn6FarYBGt0FvJNiDitLQzB_31vsyiKAjehpAZQuYl73vJ-76H0KkLCVhKYogQlBORGUaMNTZmWXArubZOR6Lw_ViNnsXdVE4bsnrFhQkhVMlnoR8fq7t8P3dlPCqDFa4VIHQIdtbBzmqyVms8UoPzTiLWqbdh8NNaN1o-iZBnXAEQYhCjqlRS9bOYSuVLhl00bkdRp5C89svC9t3HL4HGfw9zE_W-aXv44cshbaG1kG-jblu3ATfLeAfNLiJLC7pgQH_4xixwk62FHwE2kkpEFZvc46f3HHqAaeHbN9h3VhgQLp5EXQ5497wsIiFiXq7w1fIlnkvgRv68hybD68nliDSFFohhYlAQljHqtTBeWgbhsqBWaRu1ACPh3JlMSBloxlPHnHcuYRraPEBJp7334PR3USef52EP4TSDT8iUQ7MSxkEswmmgNkuMdloZs496cb5mi1pKY9ZO1cEf7YdogwGKiM6C6SPUKZZlOAYUUNiT6u9_Ap-iscY
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3JTsMwFLRQOcCpLEXs-MA1JXG8kCNbaaGtEASpt8hbJIRIqjY5wNfznAUEEhK3yHIiy3r2zHPejBE61TaASAmkR6kfejSVxJNKKldlESoWCqWFEwpPpnz4TO9mbNaI1SstjLW2Kj6zffdY_cs3uS7dURmscMGBoUOyswrAT1kt12rDhwmA78CxnXojBqQWonHzCSg7CzlQIQJZKo-Yz39ep1KhyaCLpu046iKS135ZqL7--GXR-O-BbqDet3APP3xB0iZasdkW6rY3N-BmIW-j5NLptKALBv6Hb-UcN_Va-BGIo1fZqGKZGfz0nkEPCC48eoOdZ4mB4-LYOXPAuxdl4SQRebnE14sXdzKBGwP0HooHN_HV0GuuWvAkoeeFR1LiG0GlYYpAwkx9xYVyboBOcq5lShmzfhpGmmijdUAEtBkgk1oYYwD2d1AnyzO7i3CUwidYFEIzp1JDNhL61ldpIIUWXMo91HPzlcxrM42knar9P9pP0NownoyT8Wh6f4DWCXAKBx1EHKJOsSjtEXCCQh1XkfAJHNG1Ew
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=IEEE%2FACM+International+Conference+on+Automated+Software+Engineering+%3A+%5Bproceedings%5D&rft.atitle=Bridging+the+Gap+between+Real-world+and+Synthetic+Images+for+Testing+Autonomous+Driving+Systems&rft.au=Amini%2C+Mohammad+Hossein&rft.au=Nejati%2C+Shiva&rft.date=2024-10-27&rft.pub=ACM&rft.eissn=2643-1572&rft.spage=732&rft.epage=744&rft_id=info:doi/10.1145%2F3691620.3695067&rft.externalDocID=10764991