Slim-Llama: An Energy-Efficient LLM ASIC Processor Supporting 3-Billion Parameters at Just 4.69mW
Introducing Slim-Llama, a highly efficient ASIC designed to optimize language model performance while minimizing energy consumption, capable of handling up to 3 billion parameters.
Large Language Models (LLMs) have become a cornerstone of artificial intelligence, driving advancements in natural language processing and decision-making tasks. However, their extensive power demands, resulting from high computational overhead and frequent external memory access, significantly hinder their scalability and deployment, especially in energy-constrained environments such as edge devices. This escalates the cost of operation while also limiting accessibility to these LLMs, which therefore calls for energy-efficient approaches designed to handle billion-parameter models.
To address these limitations, researchers at the Korea Advanced Institute of Science and Technology (KAIST) developed Slim-Llama, a highly efficient Application-Specific Integrated Circuit (ASIC) designed to optimize the deployment of LLMs. This novel processor uses binary/ternary quantization to reduce the precision of model weights from real to 1 or 2 bits, thus minimizing significant memory and computational demands, while maintaining performance. Utilizing a Sparsity-aware Look-up Table (SLT) for sparse data management, along with output reuses and vector indexing, Slim-Llama effectively addresses common bottlenecks in LLM deployment, providing an energy-friendly solution for executing billions of parameters.
Manufactured with Samsung’s 28nm CMOS technology, Slim-Llama has a compact die area of 20.25mm² and features 500KB of on-chip SRAM, eliminating dependencies on external memory and thus dramatically reducing energy consumption. It achieves impressive latency of 489 milliseconds and supports models with up to 3 billion parameters, making it highly adaptable for modern AI applications. The processor delivers 4.92 TOPS at an efficiency of 1.31 TOPS/W and outsmarts previous solutions with a 4.59x improvement in energy efficiency, addressing the pressing need for sustainable and efficient hardware in large-scale AI systems. Slim-Llama not only establishes a new benchmark in energy-efficient AI hardware but also opens avenues for developing more accessible and environmentally friendly AI technologies.