Youngseok Kim, S. Kim, et al.
ACM JETC
We report a capacitor-based cross-point array that can be used to train analog-based Deep Neural Networks (DNNs), fabricated with trench capacitors in 14nm technology. The fundamental DNN functionalities of multiply-accumulate and weight-update are demonstrated. We also demonstrate the best symmetry and linearity ever reported for an analog cross-point array system. For DNNs, the capacitor leakage does not impact learning accuracy even without any refresh cycle, as the weights are continuously updated during training. This makes capacitor an ideal candidate for neural network training. We also discuss the scalability of this array using optimized low-leakage DRAM technology.
Youngseok Kim, S. Kim, et al.
ACM JETC
G. Tsutsui, C. Durfee, et al.
VLSI Technology 2018
Jianshi Tang, Douglas M. Bishop, et al.
IEDM 2018
S. Kim, John A. Ott, et al.
IEDM 2019