←back to thread

506 points Terretta | 1 comments | | HN request time: 0.208s | source
Show context
NitpickLawyer ◴[] No.45066063[source]
Tested this yesterday with Cline. It's fast, works well with agentic flows, and produces decent code. No idea why this thread is so negative (also got flagged while I was typing this?) but it's a decent model. I'd say it's at or above gpt5-mini level, which is awesome in my book (I've been maining gpt5-mini for a few weeks now, does the job on a budget).

Things I noted:

- It's fast. I tested it in EU tz, so ymmv

- It does agentic in an interesting way. Instead of editing a file whole or in many places, it does many small passes.

- Had a feature take ~110k tokens (parsing html w/ bs4). Still finished the task. Didn't notice any problems at high context.

- When things didn't work first try, it created a new file to test, did all the mocking / testing there, and then once it worked edited the main module file. Nice. GPT5-mini would often times edit working files, and then get confused and fail the task.

All in all, not bad. At the price point it's at, I could see it as a daily driver. Even agentic stuff w/ opus + gpt5 high as planners and this thing as an implementer. It's fast enough that it might be worth setting it up in parallel and basically replicate pass@x from research.

IMO it's good to have options at every level. Having many providers fight for the market is good, it keeps them on their toes, and brings prices down. GPT5-mini is at 2$/MTok, this is at 1.5$/MTok. This is basically "free", in the great scheme of things. I ndon't get the negativity.

replies(10): >>45066728 #>>45067116 #>>45067311 #>>45067436 #>>45067602 #>>45067936 #>>45068543 #>>45068653 #>>45068788 #>>45074597 #
coder543 ◴[] No.45067311[source]
Qwen3-Coder-480B hosted by Cerebras is $2/Mtok (both input and output) through OpenRouter.

OpenRouter claims Cerebras is providing at least 2000 tokens per second, which would be around 10x as fast, and the feedback I'm seeing from independent benchmarks indicates that Qwen3-Coder-480B is a better model.

replies(2): >>45067631 #>>45067760 #
stocksinsmocks ◴[] No.45067760[source]
There is a national superset of “NIH” bias that I think will impede adoption of Chinese-origin models for the foreseeable future. That’s a shame because by many objective metrics they’re a better value.
replies(1): >>45068189 #
dlachausse ◴[] No.45068189[source]
In my case it's not NIH, but rather that I don't trust or wish to support my nation's largest geopolitical adversary.
replies(4): >>45070723 #>>45070873 #>>45071387 #>>45075162 #
ulfw ◴[] No.45071387[source]
"largest geopolitical adversary"

I can't believe Americans all are falling for propaganda like this. So Russia is all fine now huh. You know the country you literally had nuclear warheads pointed at for decades and decades and decades on end.

replies(3): >>45071516 #>>45071933 #>>45072313 #
anticodon ◴[] No.45072313[source]
The fact is that China is one of the largest foreign USA debt holders makes it actually scarier than nuclear warheads.

If China would decide to sell US treasuries, it will be more devastating to the US economy than effect of 10 nuclear strikes.

replies(5): >>45072326 #>>45074085 #>>45074186 #>>45075159 #>>45081823 #
1. wqaatwt ◴[] No.45081823[source]
> more devastating to the US economy

It wouldn’t be that great for China either..