Spiking Transformer Hardware Accelerators in 3D Integration

By Boxun Xu1, Junyoung Hwang2, Pruek Vanna-iampikul2, 3, Sung Kyu Lim2, Peng Li1
1 Department of Electrical and Computer Engineering, University of California, Santa Barbara, CA, USA
2 Department of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA
3 Department of Electrical Engineering, Burapha University, Chonburi, Thailand

Spiking neural networks (SNNs) are powerful models of spatiotemporal computation and are well suited for deployment on resource-constrained edge devices and neuromorphic hardware due to their low power consumption. Leveraging attention mechanisms similar to those found in their artificial neural network counterparts, recently emerged spiking transformers have showcased promising performance and efficiency by capitalizing on the binary nature of spiking operations. Recognizing the current lack of dedicated hardware support for spiking transformers, this paper presents the first work on 3D spiking transformer hardware architecture and design methodology. We present an architecture and physical design co-optimization approach tailored specifically for spiking transformers. Through memory-on-logic and logic-on-logic stacking enabled by 3D integration, we demonstrate significant energy and delay improvements compared to conventional 2D CMOS integration.

To read the full article, click here