A technological startup proposes a novel way of addressing massive LLM using the fastest memory available for humanity


  • The GPU type PCIE card offers 10 pflop fp4 calculation power and 2 GB of SRAM
  • The SRAM is usually used in small quantities as cache in the processors (L1 to L3)
  • It also uses LPDDR5 instead of HBM memory, much more expensive.

The Silicon Valley Startup D-Matrix, which has the support of Microsoft, has developed a chiplet-based solution designed for rapid inference and small lots of LLM in business environments. Its architecture adopts a computing approach in totally digital memory, using modified SAM cells for faster speed and energy efficiency.

Corsair, the current D-Matrix product, is described as the “first computer platform of its type” and has two ASIC D-Matrix on a PCIE card of height and full length, with four chiplets per ASIC. It reaches a total of 9.6 PFLOP of computer power FP4 with 2 GB of SRAM -based performance memory. Unlike traditional designs that depend on the expensive HBM, Corsair uses LPDDR5 capacity memory, with up to 256 GB per card to handle larger models or inference work loads by lots.

Leave a Comment

Your email address will not be published. Required fields are marked *