Analysis of ADC Quantization Effect in Processing-In-Memory Macro in Various Low-Precision Deep Neural Networks

This paper introduces a method for adjusting flash ADC levels, focusing on ternary inputs and binary weights (1, −1) and (1, 0), to improve test accuracy in DNN and CNN models. It proposes two approaches for mapping bitline voltages with noise in PIM macro to MAC values to optimize ADC levels: rough...

Full description

Saved in:
Bibliographic Details
Published in2024 International Conference on Electronics, Information, and Communication (ICEIC) pp. 1 - 2
Main Authors Jin, Seung-Mo, Kang, Shin-Uk, Choo, Min-Seong
Format Conference Proceeding
LanguageEnglish
Published IEEE 28.01.2024
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:This paper introduces a method for adjusting flash ADC levels, focusing on ternary inputs and binary weights (1, −1) and (1, 0), to improve test accuracy in DNN and CNN models. It proposes two approaches for mapping bitline voltages with noise in PIM macro to MAC values to optimize ADC levels: rough tuning, which linearly maps MAC values, and fine-tuning, which maps the range of MAC values determined through rough tuning in a new way. Additionally, it demonstrates that this approach can enhance test accuracy for ternary inputs without requiring the highest possible flash ADC levels, providing insights into the direction of PIM macro design.
ISSN:2767-7699
DOI:10.1109/ICEIC61013.2024.10457226