←back to thread

310 points skarat | 1 comments | | HN request time: 0.213s | source

Things are changing so fast with these vscode forks I m barely able to keep up. Which one are you guys using currently? How does the autocomplete etc, compare between the two?
Show context
monster_truck ◴[] No.43960400[source]
Void. I'd rather run my own models locally https://voideditor.com/
replies(1): >>43960805 #
traktorn ◴[] No.43960805[source]
Which model are you running locally? Is it faster than waiting for Claudes generation? What gear do you use?
replies(2): >>43962562 #>>43964972 #
1. jhonof ◴[] No.43962562[source]
Not OP but for autocomplete I am running Qwen2.5-Coder-7B and I quantized it using Q2_K. I followed this guide:

https://blog.steelph0enix.dev/posts/llama-cpp-guide/#quantiz...

And I get fast enough autcomplete results for it to be useful. I have and NVIDIA 4060 RTX in a laptop with 8 gigs of dedicated memory that I use for it. I still use claude for chat (pair programming) though, and I don't really use agents.