Chiplets for generative AI
By Jawad Nasrullah, CEO - Palo Alto Electron
Generative AI models, known for their large size and substantial computational demands, are pushing the boundaries of traditional computing infrastructure. As the industry seeks solutions to mitigate costs, execution times, and the environmental impact of these models, the concept of scale-out computing traditionally seen at the data center level is being integrated into IC (Integrated Circuit) packaging using chiplet technology. This integration aims to address the challenges of power consumption and thermal design. The talk explores innovative strategies to enhance chip efficiency and reduce overheads. Key approaches include the development of AI-specific core chiplets, the implementation of efficient communication fabrics, the expansion of on-chip memory, the incorporation of more components within the IC package, the improvement of die-to-die interfaces, and the adoption of vertical chip stacking technologies. These techniques are vital for reducing power and mitigating hotspots.
Related Videos
- How Chiplets Accelerate Generative AI Applications
- Chiplets for the future of AI
- Connectivity for AI Everywhere: The Role of Chiplets
- Photonic Fabric Interface Chiplets for AI XPU Optical Connectivity
Latest Videos
- Revolutionizing SoC Design: The Shift to Chiplet-Based Architectures
- Embedded World, Nuremberg: Arm’s Suraj Gajendra on AI, Chiplets, and the Future of Automotive Compute
- Keysight Expands Chiplet Interconnect Support with UCIe 2.0 & BoW
- Revolutionizing AI & Chiplets: Baya Systems CEO on $36M Series B, UALink, Future of Data Movement
- Machine Learning Applications in EDA for Chiplet Reliability