Hacker Read top | best | new | newcomments | leaders | about | bookmarklet login
Kaist develops next-generation ultra-low power LLM accelerator (en.yna.co.kr) similar stories update story
3 points by readline_prompt | karma 84 | avg karma 2.62 2024-03-06 06:21:47 | hide | past | favorite | 41 comments



view as:

> The 4.5-mm-square chip, developed using Korean tech giant Samsung Electronics Co.'s 28 nanometer process, has 625 times less power consumption compared with global AI chip giant Nvidia's A-100 GPU, which requires 250 watts of power to process LLMs, the ministry explained.

>processes GPT-2 with an ultra-low power consumption of 400 milliwatts and a high speed of 0.4 seconds

Not sure what's the point on comparing the two, an A100 will get you a lot more speed than 2.5 tokens/sec. GPT 2 is just a 1.5B param model, a Pi 4 would get you more tokens per second with just CPU inference.

Still, I'm sure there's improvements to be made and the direction is fantastic to see, especially after Coral TPUs have proven completely useless for LLM and whisper acceleration. Hopefully it ends up as something vaguely affordable.


Which of the model requirements of Coral TPUs [1] are the most problematic for LLMs?

[1] https://coral.ai/docs/edgetpu/models-intro/#model-requiremen...


Guessing as to what the GP meant--coral TPUs max out around 8M parameters, IIRC. That's a few orders of magnitude less than the smallest LLM model.

The part where they have like 3 bytes of memory so you switch from extremely high latency of RAM to laughably sluggish latency of USB serial. I think there's also no support below 8 bit quants, which you'd really need.

> New structure mimics the layout of neurons and synapses

What does that mean, practically?

How can you mimic that layout in silicon?


This means they use Spiking Neural Networks. It’s a software algorithm that most likely doesn’t work as well as regular NNs.

Well, our brains are closer to spiking neural networks than 'regular' neural networks. And they work pretty well. For the most part.

I feel like SNNs are like Brazil - they are the future, and shall remain so. I think more basic research is needed for them to mature. AFAIK the current SOTA is to train them with 'surrogate gradients', which shoe-horn them into the current NN training paradigm, and that sort of discards some of their worth. Have biologically-inspired learning rules, like STDP, _really_ been exhausted?


But this group claims to have demonstrated a way to use SNNs to run LLMs effectively and with vastly less energy usage.

If OpenAI or DeepMind makes such claim I'd pay attention. Otherwise it's always some (usually hw) guys trying to get a grant, or even just publish a paper.

p.s. People interested in biologically inspired data processing algorithms should look at Numenta's papers (earlier ones, because recently they switched to regular deep learning), and especially learn their justification for not using spikes.


They switched? Why? There goes their whole raison d'etre!

Spiking neural networks are not software, they are usually built directly into silicon chips because they are using pulse timing to encode information instead of multiple bits. The problem is that training them is difficult because they operate over time, not that they don't work. As of now, scaling training infrastructure is more important than theoretical power efficiency.

Neuromorphic computing basically uses individual "neurons", represented with either analog or digital circuits, which communicate using asynchronous pulses called "spikes". Unlike the human brain, neuromorphic chips are 2D, but we can replicate a good amount of neural dynamics in silicon.

It's unclear how they managed to use this to run LLMs, though. Getting GPT-2 running with SNNs is a legitimate achievement, because SNNs have traditionally lagged significantly behind conventional deep learning architectures.

https://web.stanford.edu/group/brainsinsilicon/documents/ANe... https://web.stanford.edu/group/brainsinsilicon/documents/Ben...


Want to reference Groq.com. They are developing their own inference hardware called an LPU https://wow.groq.com/lpu-inference-engine/

They also released their API a week or 2 ago. Its significantly faster than anything from OpenAI right now. Mixtral 8x7b operates at around 500 tokens per second. https://groq.com/


It's not so much an accelerator as it is addressing the main inference bottleneck (i.e. memory latency) with sheer brute force by throwing money at the problem. They've made accelerators out of pure L3 cache with a whopping 230 MB per card. They cited something like 500 cards to load one single Mixtral instance, which probably cost over $10M to build. It's a supercomputer essentially.

Or to put it another way: they’ve made a compute substrate with the correct ratios of processing power to memory capacity.

NVIDIA GPUs were optimised for different workloads, such as 3D rendering, that have different optimal ratios.

This “supercomputer” isn’t brute force or wasteful because it allows more requests per second. By having each response get processed faster it can pipeline more of them through per unit time and unit silicon area.


The correct ratio for one workload (production inference).

A recent presentation on the architecture

https://youtu.be/WQDMKTEgQnY?si=W0E9Kq6P280l3Wcl

IMO we still need an MLPerf submission or similar to really understand if this is more efficient or more efficient only if you also want to minimize latency.

Nvidia has pulled enough rabbits out of the hat when it comes to MLPerf I’m still not convinced they can’t work some CUDA magic and undercut them on efficiency.


They need 568 LPUs to load both Mixtral 8x7B and LLaMA 70B, because they need both those models available for the demo.

I imagine Mixtral by itself would only take something like 200-300 LPUs


Only $5M then.

$5M once, upfront. But given the significantly increased throughput, how fast does that pay for itself?

Depends on power usage. I’m curious how power hungry those are compared to server/workstation cards.

You need computers for all of them and megawatts of power, power supplies, cooling, and power distribution.

Naturally, but you need that for GPUs as well, no? What is the actual difference when running, when measured per token generated?

I'm pretty sure $20,000 per LPU isn't actually the cost of these LPUs. I saw someone else on HN asking if $20,000 could get them something and an employee said to reach out. Which makes me think $20,000 is enough to get some sort of model running at least, even if it's not necessarily an LLM.

Nothing wrong with that, though.

> pure L3 cache with a whopping 230 MB per card

Just to put these numbers in perspective a desktop 8 core 7800x3d has 96MB of L3 cache, and the top-end 96-core Epyc 9684X has 1.15GB of L3.


What's the cost per inference relative to H100? Isn't that the number to care about?

If you believe the marketing material it’s lower. Their API is the cheapest around, so either it’s true or they’re subsidizing.

Another consideration: Even if it's slightly more expensive, that can be OK if you care about inference speed. I'd pay 50% more for GPT-4 if it could deliver results that quick.

Based on some rough ballpark conservative estimates (one server with 2 A100 at $50000; 50 tokens/s one one of those servers; so 10 of those servers), upfront cost with consumer hardware seems to be 1/10 to 1/20 of what the Groq hardware costs. I would guess that realistically cloud providers can probably achieve half to 1/3 of that price

So unless you need the fast latency of Groq, consumer hardware seems to be a lot cheaper for the same thoughput.


Grayskull has 96 MB SRAM and people call it overpriced at $600 to $800. It is far more plausible that their chip costs are somewhere around $500.

Neuromorphic computing is cool, but not new tech. However, using a neuromorphic spiking architecture to run LLMs seems new. Unfortunately, there doesn't seem to be a paper associated with this work, so there's no deeper information on what exactly they're doing.

I heavily doubt that they are running LLMs on this.

The article says they ran GPT-2! Which isn't particularly large, but replicating a large language model with a spiking neural network seems like novel work at least.

Is their work similar to Rain.ai then?

The article says 400 milliwatt power draw.

Wolfram Alpha says thats roughly equivalent to cell phone power draw when sleeping.


We build software acceleration for LLM, effectively running smaller llama2 models at the same performance on several L4's as on 1xA100.

Quick shoutout to https://youtube.com/@TechTechPotato for those interested in keeping tabs on the AI hardware space. There is much more going on in this area than you would think if you only follow general media.

Legal | privacy