[ 3 / biz / cgl / ck / diy / fa / ic / jp / lit / sci / vr / vt ] [ index / top / reports ] [ become a patron ] [ status ]
2023-11: Warosu is now out of extended maintenance.

/biz/ - Business & Finance

Search:


View post   

>> No.57744416 [View]
File: 130 KB, 517x590, llm-training-128-cpu.jpg [View same] [iqdb] [saucenao] [google]
57744416

>>57736404
Where will new senior devs come from after the current crop dies?

>> No.57722841 [View]
File: 130 KB, 517x590, llm-training-128-cpu.jpg [View same] [iqdb] [saucenao] [google]
57722841

>>57722226
Memory is the major bottleneck with LLM training. NVIDIA is now producing the H100 chips which are not strictly faster but have much more memory, and an even larger price tag to go with it.
On the other hand, the unified memory model overcomes this limitation much more elegantly where you can get an M-chip macbook with 97GB RAM shared between CPU and GPU. AMD also touts this unified memory model which allows much more efficient use of memory (was pushed into mainstream with PS4).

FWIW here is an LLM running on 128-cpu machine.
https://youtu.be/argfZlPZKdY?t=527

Navigation
View posts[+24][+48][+96]