A Heterogeneous Chiplet Architecture for Accelerating End-to-End Transformer Models
By Harsh Sharma, Pratyush Dhingra, Janardhan Rao Doppa, Partha Pratim Pande (Washington State University) and Umit Ogras (University of Wisconsin Madison)
Transformers have revolutionized deep learning and generative modeling, enabling unprecedented advancements in natural language processing tasks. However, the size of transformer models is increasing continuously, driven by enhanced capabilities across various deep-learning tasks. This trend of ever-increasing model size has given rise to new challenges in terms of memory and computing requirements. Conventional computing platforms, including GPUs, suffer from suboptimal performance due to the memory demands imposed by models with millions/billions of parameters. The emerging chiplet-based platforms provide a new avenue for compute- and data-intensive machine learning (ML) applications enabled by a Network-on-Interposer (NoI). However, designing suitable hardware accelerators for executing Transformer inference workloads is challenging due to a wide variety of complex computing kernels in the Transformer architecture. In this paper, we leverage chiplet-based heterogeneous integration (HI) to design a high-performance and energy-efficient multi-chiplet platform to accelerate transformer workloads. We demonstrate that the proposed NoI architecture caters to the data access patterns inherent in a transformer model. The optimized placement of the chiplets and the associated NoI links and routers enable superior performance compared to the state-of-the-art hardware accelerators. The proposed NoI-based architecture demonstrates scalability across varying transformer models and improves latency and energy efficiency by up to 22.8x and 5.36x respectively.
To read the full article, click here
Related Chiplet
- 12nm EURYTION RFK1 - UCIe SP based Ka-Ku Band Chiplet Transceiver
- Interconnect Chiplet
- Bridglets
- Automotive AI Accelerator
- Direct Chiplet Interface
Related Technical Papers
- AIG-CIM: A Scalable Chiplet Module with Tri-Gear Heterogeneous Compute-in-Memory for Diffusion Acceleration
- RapidChiplet: A Toolchain for Rapid Design Space Exploration of Chiplet Architectures
- Chiplet Cloud: Building AI Supercomputers for Serving Large Generative Language Models
- DCRA: A Distributed Chiplet-based Reconfigurable Architecture for Irregular Applications
Latest Technical Papers
- Die-Level Transformation of 2D Shuttle Chips into 3D-IC for Advanced Rapid Prototyping using Meta Bonding
- STAMP-2.5D: Structural and Thermal Aware Methodology for Placement in 2.5D Integration
- MCMComm: Hardware-Software Co-Optimization for End-to-End Communication in Multi-Chip-Modules
- FoldedHexaTorus: An Inter-Chiplet Interconnect Topology for Chiplet-based Systems using Organic and Glass Substrates
- ChipletQuake: On-die Digital Impedance Sensing for Chiplet and Interposer Verification