The authors report the implementation of a Transformer-based model on the same architecture used in Large Language Models in a 14nm analog AI accelerator with 35 million Phase Change Memory devices, which achieves near iso-accuracy despite hardware imperfections and noise.
- An Chen
- Stefano Ambrogio
- Geoffrey W. Burr