Breaking the Memory Wall: How d-Matrix Is Redefining AI Inference with Chiplets
By Maurizio Di Paolo Emilio, embedded.com | April 23, 2025
As AI workloads push the limits of performance, power efficiency, and memory bandwidth, chiplets are rapidly emerging as the architectural solution of choice. In this interview, Sree Ganesan, Vice President of Product at d-Matrix, dives deep into how its pioneering chiplet-based platform is revolutionizing AI inference. From solving the memory wall with Digital In-Memory Computing (DIMC) to enabling seamless multi-chiplet communication via custom interconnects, d-Matrix reveals how its innovations are unlocking 10x faster token generation, 3x better energy efficiency, and a scalable roadmap for generative AI.
To read the full article, click here
Related Chiplet
- Interconnect Chiplet
- 12nm EURYTION RFK1 - UCIe SP based Ka-Ku Band Chiplet Transceiver
- Bridglets
- Automotive AI Accelerator
- Direct Chiplet Interface
Related News
- EdgeCortix Awarded New 3 Billion Yen NEDO Project to Develop Advanced Energy-Efficient AI Chiplet for Edge Inference and Learning
- MSquare Technology Showcases Leadership in IP and Chiplet Innovation at the AI Hardware & Edge AI Summit
- ADTechnology Partners with Arm, Samsung Foundry, and Rebellions on AI CPU Chiplet Platform
- Rebellions Partners with Arm, Samsung Foundry, and ADTechnology on Next-Gen AI Computing Chiplet Technology
Latest News
- Open Compute Project Expands Open Chiplet Economy Ecosystem
- Arm Sets the Standard for Open, Converged AI Data Centers
- InPsytech showcases 3nm UCIe 3.0 technology at OCP 2025, Accelerating Innovation in Chiplet Ecosystem
- Ayar Labs Appoints Vivek Gupta as Chief Strategy Officer to Lead Hyperscaler Collaborations and Accelerate CPO Adoption
- ASML appoints next Chief Technology Officer