MRS Meetings and Events

 

EQ10.02.09 2022 MRS Fall Meeting

Optimized Weight Programming for Analog Memory-Based Deep Neural Networks

When and Where

Nov 28, 2022
3:45pm - 4:15pm

Sheraton, 2nd Floor, Independence West

Presenter

Co-Author(s)

Charles Mackin1,Malte Rasch1,An Chen1,Jonathan Timcheck2,Robert Bruce1,Ning Li1,Pritish Narayanan1,Stefano Ambrogio1,Manuel Le Gallo1,S.R. Nandakumar1,Andrea Fasoli1,Jose Luquin1,Alexander Friz1,Abu Sebastian1,Hsinyu Tsai1,Geoffrey Burr1

IBM1,Stanford University2

Abstract

Charles Mackin1,Malte Rasch1,An Chen1,Jonathan Timcheck2,Robert Bruce1,Ning Li1,Pritish Narayanan1,Stefano Ambrogio1,Manuel Le Gallo1,S.R. Nandakumar1,Andrea Fasoli1,Jose Luquin1,Alexander Friz1,Abu Sebastian1,Hsinyu Tsai1,Geoffrey Burr1

IBM1,Stanford University2
Analog memory-based deep neural networks provide energy-efficiency and per-area throughput gains relative to state-of-the-art digital counterparts such as graphics processing units. Recent advances focus largely on hardware-aware algorithmic training and improvements to circuits, architectures, and memory devices. Optimal translation of software-trained weights into analog hardware weights—given the plethora of complex memory non-idealities—represents an equally important task. We report a generalized computational framework that automates the crafting of complex weight programming strategies to minimize accuracy degradations during inference, particularly over time. The framework is agnostic to network structure and generalizes well across recurrent, convolutional, and transformer neural networks. As a highly flexible numerical heuristic, the approach accommodates arbitrary device-level complexity, making it potentially relevant for a variety of analog memories. Interestingly, this computational technique can optimize inference accuracy without the need to run inference simulations or evaluate large training, validation, or test datasets. By quantifying the limit of achievable inference accuracy, it also enables analog memory-based deep neural network accelerators to reach their full inference potential.

Symposium Organizers

Wei Zhang, Xi'an Jiaotong University
Valeria Bragaglia, IBM Research Europe - Zurich
Juejun Hu, Massachusetts Institute of Technology
Andriy Lotnyk, Leibniz Institute of Surface Engineering

Publishing Alliance

MRS publishes with Springer Nature