Looks like ballpark a million dollars of GPU time if you want to train up one for yourself (4000 gpus/24 days).
Very nice write up that’s generous in sharing their learnings.
This is a solid and positive contribution.
Looks like ballpark a million dollars of GPU time if you want to train up one for yourself (4000 gpus/24 days).
Very nice write up that’s generous in sharing their learnings.
This is a solid and positive contribution.
It was 24 days (576 hours) not 24 hours. $663,552 @ $3/hr.
Found this a few days ago which might be neat for finding cheaper https://www.primeintellect.ai/
No affiliation with either
WARNING: This is highly speculative and napkin math
H200 (141 GB HBM3 - $3.99/h - 1.4x perf) 216 x 24 x 17 = 88128h = 351.895,104 (17 days and 216 cards)
B200 (192 GB HBM3e - $5.99/h - 2.8x perf) 158 x 24 x 9 = 34128h = $204.426,72
Probably wrong math, should be more efficient and cheaper. Doubt that they have 100/200 cards available for that long.
Source: I've only trained using RTX4090 and stuff like that with 8 cards.
Not affiliated in any way with Runpod.