No video

How Groq’s LPUs Overtake GPUs For Fastest LLM AI!

  Рет қаралды 21,671

ipXchange

ipXchange

Күн бұрын

Пікірлер: 6
@Maisonier
@Maisonier Ай бұрын
I'd love to have a small black box at home with several Groq LPUs acting as LLMs for my local network. It would serve a typical family of five, each accessing it from their phones via WiFi while at home working, especially since internet connectivity can be an issue. I wonder if they'll ever sell such a device to the general public instead of just focusing on businesses?
@ipXchange
@ipXchange Ай бұрын
I couldn't say. They do make racks, but I wonder how many you would need to make something viable at home, and whether they'd let you buy not in bulk. That would be cool though. To be fair, you can use Groq cloud, but I guess you want to own your own infrastructure. Groq has deployed their LPU in super small use cases, so there might be a possibility you could get you hands on some private units...
@alertbri
@alertbri Ай бұрын
How does an LPU differ from an ASIC please?
@ipXchange
@ipXchange Ай бұрын
I suppose it could be considered a type of ASIC as it is a processor designed specifically for large language model processing. The way that an LPU differs from a GPU is that it does not do any parallel processing - it's very good at doing things in sequence. For applications like LLMs or audio, going forward in time is all that's required because the next word depends on the words that came before it. It's pretty much a 1D problem. This is in contrast to GPUs because a 2D or 3D picture needs to understand the whole context of a scene, hence why it requires parallel processing of all the pixels in order to understand what's going on. While parallel processing in GPUs can be used to enable faster LLM AI, at a certain point, the recombination of data slows the whole process down. The LPU, however, is able to just keep chugging along at the same pace because any parallelism is done in separate chips. At a certain number of devices, it seems that this wins out in terms of performance as the GPUs stop providing a net gain for more units added to the system. This is an oversimplification, but you get the idea. Thank you for the comment and question.
@Davorge
@Davorge Ай бұрын
@@ipXchange interesting, so why are billionaries dropping hundreds of millions in H100 clusters? wouldnt it be better for them to invest in LPU's moving forward?
@kahvac
@kahvac 20 күн бұрын
@@Davorge You have to start somewhere..if you keep waiting for the next best thing you will be left behind.
Do we really need NPUs now?
15:30
TechAltar
Рет қаралды 477 М.
艾莎撒娇得到王子的原谅#艾莎
00:24
在逃的公主
Рет қаралды 52 МЛН
Before VS during the CONCERT 🔥 "Aliby" | Andra Gogan
00:13
Andra Gogan
Рет қаралды 9 МЛН
الذرة أنقذت حياتي🌽😱
00:27
Cool Tool SHORTS Arabic
Рет қаралды 24 МЛН
ПОМОГЛА НАЗЫВАЕТСЯ😂
00:20
Chapitosiki
Рет қаралды 30 МЛН
Nature's Incredible ROTATING MOTOR (It’s Electric!) - Smarter Every Day 300
29:37
The moment we stopped understanding AI [AlexNet]
17:38
Welch Labs
Рет қаралды 953 М.
Linus Torvalds: Speaks on Hype and the Future of AI
9:02
SavvyNik
Рет қаралды 173 М.
ROCKET that LITERALLY BURNS WATER as FUEL
19:00
Integza
Рет қаралды 2,4 МЛН
Google CEO Sundar Pichai and the Future of AI | The Circuit
24:02
Bloomberg Originals
Рет қаралды 3,4 МЛН
Generative AI in a Nutshell - how to survive and thrive in the age of AI
17:57
AI Deception: How Tech Companies Are Fooling Us
18:59
ColdFusion
Рет қаралды 1,8 МЛН
艾莎撒娇得到王子的原谅#艾莎
00:24
在逃的公主
Рет қаралды 52 МЛН