Slim-Llama: An Energy-Efficient LLM ASIC Processor Supporting 3-Billion Parameters at Just 4.69mW
Discover Slim-Llama, a groundbreaking ASIC processor designed to run billion-parameter LLMs with remarkable energy efficiency.
Large Language Models (LLMs) have become a cornerstone of artificial intelligence, driving advancements in natural language processing and decision-making tasks. However, their extensive power demands, resulting from high computational overhead and frequent external memory access, significantly hinder their scalability and deployment, especially in energy-constrained environments such as edge devices. This escalates the cost of operation while also limiting accessibility to these LLMs, which therefore calls for energy-efficient approaches designed to handle billion-parameter models.
To address these limitations, researchers at the Korea Advanced Institute of Science and Technology (KAIST) developed Slim-Llama, a highly efficient Application-Specific Integrated Circuit (ASIC) designed to optimize the deployment of LLMs. This novel processor uses binary/ternary quantization to reduce the precision of model weights from real to 1 or 2 bits, thus minimizing significant memory and computational demands, leaving performance intact. It employs output reuses and vector indexing with optimizations to enhance data flows, effectively providing a scalable support mechanism for handling execution tasks within billions of LLMs.
Manufactured using Samsung’s 28nm CMOS technology, Slim-Llama features a compact die area of 20.25mm² and 500KB of on-chip SRAM, removing dependencies on external memory and achieving bandwidth of up to 1.6GB/s. This processor remarkably reaches a latency of 489 milliseconds using the Llama 1-bit model while supporting up to 3 billion parameters. The architectural innovations, including binary and ternary quantization and sparsity-aware optimization, deliver significant energy efficiency improvements, boasting a 4.59x enhancement over previous solutions with consumption levels from 4.69mW at 25MHz to 82.07mW at 200MHz. As a result, Slim-Llama is poised to handle real-time applications effectively, setting new benchmarks in energy-efficient AI systems.