Optimized model architectures for deep learning on genomic data

The success of deep learning in various applications depends on task-specific architecture design choices, including the types, hyperparameters, and number of layers. In computational biology, there is no consensus on the optimal architecture design, and decisions are often made using insights from...

Full description

Saved in:
Bibliographic Details
Published inCommunications biology Vol. 7; no. 1; p. 516
Main Authors Gündüz, Hüseyin Anil, Mreches, René, Moosbauer, Julia, Robertson, Gary, To, Xiao-Yin, Franzosa, Eric A, Huttenhower, Curtis, Rezaei, Mina, McHardy, Alice C, Bischl, Bernd, Münch, Philipp C, Binder, Martin
Format Journal Article
LanguageEnglish
Published England Nature Publishing Group 30.04.2024
Nature Publishing Group UK
Nature Portfolio
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The success of deep learning in various applications depends on task-specific architecture design choices, including the types, hyperparameters, and number of layers. In computational biology, there is no consensus on the optimal architecture design, and decisions are often made using insights from more well-established fields such as computer vision. These may not consider the domain-specific characteristics of genome sequences, potentially limiting performance. Here, we present GenomeNet-Architect, a neural architecture design framework that automatically optimizes deep learning models for genome sequence data. It optimizes the overall layout of the architecture, with a search space specifically designed for genomics. Additionally, it optimizes hyperparameters of individual layers and the model training procedure. On a viral classification task, GenomeNet-Architect reduced the read-level misclassification rate by 19%, with 67% faster inference and 83% fewer parameters, and achieved similar contig-level accuracy with ~100 times fewer parameters compared to the best-performing deep learning baselines.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISSN:2399-3642
2399-3642
DOI:10.1038/s42003-024-06161-1