Hacker Read top | best | new | newcomments | leaders | about | bookmarklet login

It's meant to compete with nvidia's DGX systems with 8 GPUs per node.


sort by: page size:

It's 24 Nvidia DGX-1 servers, which contain 8 GPUs each. It's worth noting that Nvidia already have their own 124-node DGX-1 installation, which would have 992 GPUs.

It's only single node, multiple GPUs, though, which already exists in other frameworks.

This will give us servers for deep learning that can have 8 GPUs and a couple of NVMe disks on PCI 4.0 (32 GB/s). With very good inter-GPU I/O and access to NVMe, it will enable commodity servers that are competitive with Nvidia's DGX-1 or DGX2, that include SXM2 (Nvlink with 80GB/s between GPUs).

I'm glad they're open-sourcing this, but I have to say that making 8 GPUs work together is not that big of a deal. Companies like Cirrascale are making up to 16 GPUs scale linearly with a blade.

How many gpus do they have?

Actually the PCI express topology is configurable, which is one of the innovations. You can put all 8 gpus on a single CPU bus or have 4 on each CPU (but have to use QPI between them)

> Nvidia uses a new NVLink Switch System with 36 NVLink switches to tie together 256 GH200 Grace Hopper chips and 144 TB of shared memory into one cohesive unit that looks and acts like one massive GPU

Is there more information about this?


Linux can do up to 16 GPUs.

Yeah, but it'll be slower than the equivalent Nvidia GPU cluster.

But still only single GPU for now. I also heard great things about it, but wanted to make the maximum use of my multi-GPU local setup.

One thing that isn't clear from the website is that it requires two GPUs, one for the host and one for the VM.

I would like to point out that their research seems lacking.

The industry is not "moving to one big GPU".

The industry has already moved to 4, 6, or 8 "big" GPUs per station/node.


Hmm, multiple GPU chips on the same board? Could this be where did Nvidia came up with the idea?

We're pretty excited near-term for getting to sub-second / sub-100ms interactive time on real GB workloads. That's pretty normal in GPU land. More so, where this is pretty clearly going, is using multiGPU boxes like DGX2s that already have 2 TB/s memory bandwidth. Unlike multinode cpu systems, I'd expect better scaling b/c no need to leave the node.

With GPUs, the software progression is single gpu -> multi-gpu -> multinode multigpu. By far, the hardest step is single gpu. They're showing that.


This actually seems like just a very clever market segmentation solution since the GPU was already limited to 8x PCIe lanes (its a laptop GPU see https://www.notebookcheck.net/NVIDIA-GeForce-RTX-4060-Laptop...). The 'addition' of the M.2 SSD makes it a unique offering. Limiting it to only one drive is another way to keep the thermal envelope down. Kudos to the Asus design and product development folks.

It'll be interesting to see if this goes the Larrabee route (a flat array of CPUs which share everything) or the traditional GPU route (multiple levels of shared resources).

The article is pretty light on details. What's revolutionary about 8 GPUs in a PC on a shelf?

Dual GPUs should be considered normal/consumer grade setup, hopefully they'll add it soon, on 4bits it's enough with plenty of space for context.

This whole thing is a fork of llamacpp, also hoping it'll all go upstream sooner or later.


A DGX-1 box has 8 Pascal GPUs. The reason it costs a lot more than 8 GTX 1080s is the remarkable interconnect and memory bandwidth.
next

Legal | privacy