←back to thread

504 points Terretta | 2 comments | | HN request time: 0.001s | source
Show context
NitpickLawyer ◴[] No.45066063[source]
Tested this yesterday with Cline. It's fast, works well with agentic flows, and produces decent code. No idea why this thread is so negative (also got flagged while I was typing this?) but it's a decent model. I'd say it's at or above gpt5-mini level, which is awesome in my book (I've been maining gpt5-mini for a few weeks now, does the job on a budget).

Things I noted:

- It's fast. I tested it in EU tz, so ymmv

- It does agentic in an interesting way. Instead of editing a file whole or in many places, it does many small passes.

- Had a feature take ~110k tokens (parsing html w/ bs4). Still finished the task. Didn't notice any problems at high context.

- When things didn't work first try, it created a new file to test, did all the mocking / testing there, and then once it worked edited the main module file. Nice. GPT5-mini would often times edit working files, and then get confused and fail the task.

All in all, not bad. At the price point it's at, I could see it as a daily driver. Even agentic stuff w/ opus + gpt5 high as planners and this thing as an implementer. It's fast enough that it might be worth setting it up in parallel and basically replicate pass@x from research.

IMO it's good to have options at every level. Having many providers fight for the market is good, it keeps them on their toes, and brings prices down. GPT5-mini is at 2$/MTok, this is at 1.5$/MTok. This is basically "free", in the great scheme of things. I ndon't get the negativity.

replies(10): >>45066728 #>>45067116 #>>45067311 #>>45067436 #>>45067602 #>>45067936 #>>45068543 #>>45068653 #>>45068788 #>>45074597 #
jameshart ◴[] No.45067602[source]
If the Grok brand wasn’t terminally tarnished for you by the ‘mechahitler’ incident, I’m not sure what more it would take.

This is an offering being produced by a company whose idea of responsible AI use involves prompting a chatbot that “You spend a lot of time on 4chan, watching InfoWars videos” - https://www.404media.co/grok-exposes-underlying-prompts-for-...

A lot of people rightly don’t want any such thing anywhere near their code.

replies(16): >>45067741 #>>45067793 #>>45067834 #>>45067845 #>>45067876 #>>45067950 #>>45068178 #>>45068224 #>>45068385 #>>45068645 #>>45068805 #>>45068858 #>>45069087 #>>45069800 #>>45070448 #>>45071147 #
orochimaaru ◴[] No.45068224[source]
How exactly is a code assistant “partisan”? I don’t use X but I’m open to buying a Tesla and grok for code purposes.

Kinda weird to mix political sentiment with a coding technology.

replies(3): >>45068431 #>>45068451 #>>45068587 #
jameshart ◴[] No.45068431[source]
Well, you’d also be forgiven for thinking ‘how on earth can a social website chatbot be a white supremacist?’ And yet xAI managed to prove that is a legitimate concern.

xAI has a shocking track record of poor decisions when it comes to training and prompting their AIs. If anyone can make a partisan coding assistant, they can. Indeed, given their leadership and past performance, we might expect them to explicitly try.

replies(2): >>45068655 #>>45068892 #
1. dudeinjapan ◴[] No.45068892[source]
Perhaps you’ve never heard of Tay?

Microsoft did pioneering work in the Nazi chatbot space.

replies(1): >>45069529 #
2. arjvik ◴[] No.45069529[source]
Fwiw Tay was unintentional and was shut down immediately upon realization… very good case study for safety folks!