Written by 10:55 am AI Device, NVIDIA

### Revolutionary Neural AI Device Developed by Asian Researchers Shows Remarkable Power and Energy Efficiency Compared to Nvidia

Claim Samsung-fabbed chip is the first ultra-low power LLM processor.

A group of researchers from the Korea Advanced Institute of Science and Technology (KAIST) presented their innovative ‘Complementary-Transformer’ AI chip at the recent 2024 International Solid-State Circuits Conference (ISSCC). This cutting-edge C-Transformer chip is touted as the world’s premier ultra-low power AI accelerator chip designed for extensive language model (LLM) processing.

In their announcement, the scientists from KAIST compare the power efficiency of the C-Transformer to Nvidia, asserting that their chip consumes 625 times less power and is 41 times smaller than Nvidia’s A100 Tensor Core GPU. They attribute the success of the Samsung-manufactured chip to advanced neuromorphic computing technology.

Despite the claim that the KAIST C-Transformer chip can perform LLM processing tasks equivalent to Nvidia’s robust A100 GPUs, no direct comparative performance metrics were provided in the press release or conference materials. This absence of concrete data raises questions, hinting that a performance evaluation might not favor the C-Transformer.

The detailed specifications of the C-Transformer chip reveal that it is currently produced on Samsung’s 28nm process, boasting a die area of 20.25mm2. Operating at a maximum frequency of 200 MHz and consuming less than 500mW, the chip can deliver up to 3.41 TOPS. While this performance may seem significantly slower than Nvidia’s claimed 624 TOPS for the A100 PCIe card, the KAIST chip’s remarkable power efficiency is highlighted (625 times less power usage).

The architecture of the C-Transformer chip is notable for its three primary functional feature blocks. These include a Homogeneous DNN-Transformer / Spiking-transformer Core (HDSC) with a Hybrid Multiplication-Accumulation Unit (HMAU) for efficient energy distribution processing, an Output Spike Speculation Unit (OSSU) to minimize latency and computations in spike domain processing, and an Implicit Weight Generation Unit (IWGU) with Extended Sign Compression (ESC) to reduce energy consumption from External Memory Access (EMA).

The unique aspect of the C-Transformer chip lies in its incorporation of tailored neuromorphic processing techniques to compress the vast parameters of LLMs. While neuromorphic computing was previously deemed inadequate for LLM applications, the KAIST team claims to have enhanced its accuracy to rival traditional deep neural networks (DNNs).

Despite the lack of direct comparisons with established AI accelerators, the potential of the first C-Transformer chip in mobile computing is undeniable. The progress made with a Samsung test chip and extensive GPT-2 testing signifies a promising development in the field of AI chip technology.

Visited 2 times, 1 visit(s) today
Tags: , Last modified: March 9, 2024
Close Search Window
Close