←back to thread

766 points huseyinkeles | 1 comments | | HN request time: 0.208s | source
Show context
mhitza ◴[] No.45571218[source]
Should be "that you can train for $100"

Curios to try it someday on a set of specialized documents. Though as I understand the cost of running this is whatever GPU you can rent with 80GB of VRAM. Which kind of leaves hobbyists and students out. Unless some cloud is donating gpu compute capacity.

replies(2): >>45571268 #>>45571369 #
Onavo ◴[] No.45571369[source]
A GPU with 80GB VRAM costs around $1-3 USD an hour on commodity clouds (i.e. the non-Big 3 bare metal providers e.g. https://getdeploying.com/reference/cloud-gpu/nvidia-h100). I think it's accessible to most middle class users in first world countries.
replies(1): >>45571954 #
antinomicus ◴[] No.45571954[source]
Isn’t the whole point to run your model locally?
replies(4): >>45572029 #>>45572031 #>>45572477 #>>45572856 #
1. theptip ◴[] No.45572029[source]
No, that’s clearly not a goal of this project.

This is a learning tool. If you want a local model you are almost certainly better using something trained on far more compute. (Deepseek, Qwen, etc)