- The new mega double -core architecture. Mini increases performance while saving energy
- Dynamic nucleus assignment optimizes workloads
- Mega nuclei for complex tasks and mini nuclei for routine processing
At the International Conference on Solid States Circuits (ISSCC) of 2025, the researchers presented a new mega.mini architecture.
Inspired by the famous “Big.little” paradigm of ARM, this universal generative processor, discussed extensively in ‘Mega.mini: a universal generative ia processor with a new large/small central architecture for NPU’, an academic document presented at the conference, promised a revolutionary approach to the design of the Neural Processing Unit (NPU).
Big’s architecture. Little of ARM has long been a basic element of efficient mobile and integrated systems, balancing the high performance nuclei with energy efficiency to optimize energy use. The mega.mini project seeks to bring a double -core philosophy similar to NPUs, which are essential to execute AI models efficiently.
Mega.mini: an NPU design that changes the game
It is likely that this approach involves pairing “mega” centers of high capacity for demanding tasks with light nuclei “mini” for routine processing. The main objective of this design is to optimize energy consumption while maximizing processing capacities for several generative artificial intelligence tasks (AI), ranging from the generation of natural language to complex reasoning.
Generative tool workloads, such as those that feed large language models or image synthesis systems, are notoriously intensive in resources. Mega’s architecture.Mini aims to delegate complex tasks to mega nuclei while discharging simpler operations to mini nuclei, equilibrium speed and energy efficiency.
Mega.mini also functions as a universal processor for generative AI. Unlike traditional faster CPUs that require customization for specific AI tasks, mega is being developed.Mini in such a way that developers can take advantage of architecture for different cases of use, including natural language processing (NLP) and multimodal ia systems that integrate text, image and audio processing.
It also optimizes the workloads, either executing Mass -based models or compact edge applications, assisted by its support for multiple types and data formats, from traditional floating point operations to emerging scarcity calculations.
This universal approach could simplify the IA development pipes and improve the efficiency of the implementation on the platforms, from mobile devices to high performance data centers.
The introduction of a double -core architecture to NPUS is a significant deviation of conventional designs: traditional NPUs often depend on a monolithic structure, which can lead to inefficiencies when varied tasks are processed.
Mega’s design addresses this limitation by creating specialized nuclei for specific types of operations. The mega nuclei are designed for high -performance tasks, such as matrix multiplications and large -scale calculations, essential for the training and execution of sophisticated large language (LLM) models, while the mini nuclei are optimized for low power operations, such as data preprocessing and inference tasks.