KAIST Develops Next-generation Ultra-low Power LLM Accelerator | Be Korea-savvy

KAIST Develops Next-generation Ultra-low Power LLM Accelerator


A photo describing an artificial intelligence chip which processes a large language model with neuromorphic computing technology provided by the Ministry of Science and ICT on March 6, 2024 (Image courtesy of Yonhap)

A photo describing an artificial intelligence chip which processes a large language model with neuromorphic computing technology provided by the Ministry of Science and ICT on March 6, 2024 (Image courtesy of Yonhap)

SEOUL, Mar. 6 (Korea Bizwire)A research team at the Korea Advanced Institute of Science and Technology (KAIST) has developed the world’s first artificial intelligence (AI) semiconductor capable of processing a large language model (LLM) with ultra-low power consumption, the science ministry said Wednesday.

The team, led by Professor Yoo Hoi-jun at the KAIST PIM Semiconductor Research Center, developed a “Complementary-Transformer” AI chip, which processes GPT-2 with an ultra-low power consumption of 400 milliwatts and a high speed of 0.4 seconds, according to the Ministry of Science and ICT.

The 4.5-mm-square chip, developed using Korean tech giant Samsung Electronics Co.’s 28 nanometer process, has 625 times less power consumption compared with global AI chip giant Nvidia’s A-100 GPU, which requires 250 watts of power to process LLMs, the ministry explained.

The chip is also 41 times smaller in area than the Nvidia model, enabling it to be used on devices like mobile phones.

The ministry said the utilization of neuromorphic computing technology, specifically spiking neural networks (SNNs), is essential to the achievement.

Previously, the technology was less accurate than deep neural networks (DNNs) and mainly capable of simple image classifications, but the research team succeeded in improving the accuracy of the technology to match that of DNNs to apply it to LLMs.

The team said its new AI chip optimizes computational energy consumption while maintaining accuracy by using unique neural network architecture that fuses DNNs and SNNs, and effectively compresses the large parameters of LLMs.

(Yonhap)

Leave a Reply

Your email address will not be published. Required fields are marked *

You may use these HTML tags and attributes: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <cite> <code> <del datetime=""> <em> <i> <q cite=""> <strike> <strong>