←back to thread

548 points nsagent | 2 comments | | HN request time: 0s | source
Show context
albertzeyer ◴[] No.44566290[source]
This is exciting. So this is using unified memory of CUDA? I wonder how well that works. Is the behavior of the unified memory in CUDA actually the same as for Apple silicon? For Apple silicon, as I understand, the memory is anyway shared between GPU and CPU. But for CUDA, this is not the case. So when you have some tensor on CPU, how will it end up on GPU then? This needs a copy somehow. Or is this all hidden by CUDA?
replies(3): >>44566325 #>>44566412 #>>44571076 #
zcbenz ◴[] No.44566412[source]
In the absence of hardware unified memory, CUDA will automatically copy data between CPU/GPU when there are page faults.
replies(4): >>44566633 #>>44566987 #>>44567184 #>>44567252 #
1. fenced_load ◴[] No.44566633[source]
There is also NVLink c2c support between Nvidia's CPUs and GPUs that doesn't require any copy, CPUs and GPUs directly access each other's memory over a coherent bus. IIRC, they have 4 CPU + 4 GPU servers already available.
replies(1): >>44566751 #
2. benreesman ◴[] No.44566751[source]
Yeah NCCL is a whole world and it's not even the only thing involved, but IIRC that's the difference between 8xH100 PCI and 8xH100 SXM2.