Mixed Precision Fermi-Operator Expansion on Tensor Cores from a Machine Learning Perspective
We present a second-order recursive Fermi-operator expansion scheme using mixed precision floating point operations to perform electronic structure calculations using tensor core units. A performance of over 100 teraFLOPs is achieved for half-precision floating point operations on Nvidia’s A100 tens...
Saved in:
Published in | Journal of chemical theory and computation Vol. 17; no. 4; pp. 2256 - 2265 |
---|---|
Main Authors | , , , , , , |
Format | Journal Article |
Language | English |
Published |
United States
American Chemical Society
13.04.2021
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | We present a second-order recursive Fermi-operator expansion scheme using mixed precision floating point operations to perform electronic structure calculations using tensor core units. A performance of over 100 teraFLOPs is achieved for half-precision floating point operations on Nvidia’s A100 tensor core units. The second-order recursive Fermi-operator scheme is formulated in terms of a generalized, differentiable deep neural network structure, which solves the quantum mechanical electronic structure problem. We demonstrate how this network can be accelerated by optimizing the weight and bias values to substantially reduce the number of layers required for convergence. We also show how this machine learning approach can be used to optimize the coefficients of the recursive Fermi-operator expansion to accurately represent the fractional occupation numbers of the electronic states at finite temperatures. |
---|---|
Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 23 89233218CNA000001 LA-UR-21-20350 USDOE Laboratory Directed Research and Development (LDRD) Program USDOE Office of Science (SC), Basic Energy Sciences (BES) |
ISSN: | 1549-9618 1549-9626 1549-9626 |
DOI: | 10.1021/acs.jctc.1c00057 |