Most active commenters
  • lordnacho(3)

←back to thread

Nobody knows how to build with AI yet

(worksonmymachine.substack.com)
526 points Stwerner | 25 comments | | HN request time: 0.639s | source | bottom
Show context
karel-3d ◴[] No.44616917[source]
Reading articles like this feels like being in a different reality.

I don't work like this, I don't want to work like this and maybe most importantly I don't want to work with somebody who works like this.

Also I am scared that any library that I am using through the myriad of dependencies is written like this.

On the other hand... if I look at this as some alternate universe where I don't need to directly or indirectly touch any of this... I am happy that it works for these people? I guess? Just keep it away from me

replies(20): >>44617013 #>>44617014 #>>44617030 #>>44617053 #>>44617173 #>>44617207 #>>44617235 #>>44617244 #>>44617297 #>>44617336 #>>44617355 #>>44617366 #>>44617387 #>>44617482 #>>44617686 #>>44617879 #>>44617958 #>>44617997 #>>44618547 #>>44618568 #
lordnacho ◴[] No.44617013[source]
But you also can't not swim with the tide. If you drove a horse-buggy 100 years ago, it was probably worth your while to keep your eye on whether motor-cars went anywhere.

I was super skeptical about a year ago. Copilot was making nice predictions, that was it. This agent stuff is truly impressive.

replies(7): >>44617059 #>>44617096 #>>44617165 #>>44617303 #>>44617421 #>>44617514 #>>44618157 #
1. bloppe ◴[] No.44617165[source]
An I the only one who has to constantly tell Claude and Gemini to stop making edits to my codebase because they keep messing things up and breaking the build like ten times in a row, duplicating logic everywhere, etc? I keep hearing about how impressive agents are. I wish they could automate me out of my job faster
replies(9): >>44617236 #>>44617257 #>>44617322 #>>44617596 #>>44617644 #>>44618327 #>>44618377 #>>44619630 #>>44620251 #
2. mbrumlow ◴[] No.44617236[source]
Did you tell them to not duplicate code?
3. Benjammer ◴[] No.44617257[source]
Are you paying for the higher end models? Do you have proper system prompts and guidance in place for proper prompt engineering? Have you started to practice any auxiliary forms of context engineering?

This isn't a magic code genie, it's a very complicated and very powerful new tool that you need to practice using over time in order to get good results from.

replies(4): >>44617299 #>>44617324 #>>44618056 #>>44618063 #
4. goalieca ◴[] No.44617299[source]
It ain’t a magic code genie. And developers don’t spend most of their day typing lines of code. Lots of it is designing, figuring out what to build, understanding the code, maintenance considerations, and adhering to the style of whatever file you’re in. All these agents needing local context and still spit junk.
5. ◴[] No.44617322[source]
6. dingnuts ◴[] No.44617324[source]
guy 1: I put money in the slot machine everyone says wins all the time and I lose

you: HAVE YOU PUT MORE TOKENS IN???? ARE YOU PUTTING THEM IN THE EXPENSIVE MACHINES???

super compelling argument /s

if you want to provide working examples of "prompt engineering" or "context engineering" please do but "just keep paying until the behavior is impressive" isn't winning me as a customer

it's like putting out a demo program that absolutely sucks and promising that if I pay, it'll get good. why put out the shit demo and give me this impression, then, if it sucks?

replies(1): >>44617410 #
7. lordnacho ◴[] No.44617410{3}[source]
The way I ended up paying for Claude max was that I started on the cheap plan, it went well, then it wanted more money, and I paid because things were going well.

Then it ran out of money again, and I gave it even more money.

I'm in the low 4 figures a year now, and it's worth it. For a day's pay each year, I've got a junior dev who is super fast, makes good suggestions, and makes working code.

replies(1): >>44617689 #
8. esafak ◴[] No.44617596[source]
Create and point them to an agent.md file http://agent.md/
replies(1): >>44618361 #
9. csomar ◴[] No.44617644[source]
They need "context engineering" which what I'll describe best as "railing" them in. If you give them a bit of a loose space, they'll massacre your code base. You can use their freedom for exploration but not for implementation.

In essence, you have to do the "engineering" part of the app and they can write the code pretty fast for you. They can help you in the engineering part, but you still need to be able to weigh in whatever crap they recommend and adjust accordingly.

10. Avicebron ◴[] No.44617689{4}[source]
> For a day's pay each year

For anyone trying to back of the napkin at $1000 as 4-figures per year, averaged as a day salary, the baseline salary where this makes sense is about ~$260,000/yr? Is that about right lordnacho?

replies(2): >>44618010 #>>44618201 #
11. lordnacho ◴[] No.44618010{5}[source]
Yeah I thought that was a reasonable number in the ballpark. I mean, it probably makes sense to pay a lot more for it. A grand is probably well within the range where you shouldn't care about it, even if you only get a basic salary and it's a terrible year with no bonus.

And that's not saying AI tools are the real deal, either. It can be a lot less than a fully self driving dev and still be worth a significant fraction of an entry level dev.

12. tempodox ◴[] No.44618056[source]
That's the beauty of the hype: Anyone who cannot replicate it, is “holding it wrong”.
replies(2): >>44618116 #>>44622478 #
13. QuantumGood ◴[] No.44618063[source]

   > it's a very complicated and very powerful new tool that you need to practice using over time in order to get good results from.
Of course this is and would be expected to be true. Yet adoption of this mindset has been orders of magnitude slower than the increase in AI features and capabilities.
14. ◴[] No.44618116{3}[source]
15. antihipocrat ◴[] No.44618201{5}[source]
I assume it's after tax too..
16. exographicskip ◴[] No.44618327[source]
Duplicate logic is definitely a thing. That and littering comments all over the place.

Worth it to me as I can fix all the above after the fact.

Just annoying haha

replies(1): >>44619891 #
17. vishvananda ◴[] No.44618377[source]
I'm really baffled why the coding interfaces have not implemented a locking feature for some code. It seems like an obvious feature to be able to select a section of your code and tell the agent not to modify it. This could remove a whole class of problems where the agent tries to change tests to match the code or removes key functionality.

One could even imagine going a step further and having a confidence level associated with different parts of the code, that would help the LLM concentrate changes on the areas that you're less sure about.

replies(1): >>44619462 #
18. Benjammer ◴[] No.44619462[source]
Why are engineers so obstinate about this stuff? You really need a GUI built for you in order to do this? You can't take the time to just type up this instruction to the LLM? Do you realize that's possible? You can just write instructions "Don't modify XYZ.ts file under any circumstances". Not to mention all the tools have simple hotkeys to dismiss changes for an entire file with the press of a button if you really want to ignore changes to a file or whatever. In Cursor you can literally select a block of text and press a hotkey to "highlight" that code to the LLM in the chat, and you could absolutely tell it "READ BUT DON'T TOUCH THIS CODE" or something, directly tied to specific lines of code, literally the feature you are describing. BUT, you have to work with the LLM and tooling, it's not just going to be a button for you or something.

You can also literally do exactly what you said with "going a step further".

Open Claude Code, run `/init`. Download Superwhisper, open a new file at project root called BRAIN_DUMP.md, put your cursor in the file, activate Superwhisper, talk in stream of consciousness-style about all the parts of the code and your own confidence level, with any details you want to include. Go to your LLM chat, tell it to "Read file @BRAIN_DUMP.md" and organize all the contents into your own new file CODE_CONFIDENCE.md. Tell it to list the parts of the code base and give it's best assessment of the developer's confidence in that part of the code, given the details and tone in the brain dump for each part. Delete the brain dump file if you want. Now you literally have what you asked for, an "index" of sorts for your LLM that tells it the parts of the codebase and developer confidence/stability/etc. Now you can just refer to that file in your project prompting.

Please, everyone, for the love of god, just start prompting. Instead of posting on hacker news or reddit about your skepticism, literally talk to the LLM about it and ask it questions, it can help you work through almost any of this stuff people rant about.

replies(3): >>44620215 #>>44620666 #>>44622866 #
19. dvfjsdhgfv ◴[] No.44619630[source]
It happens to me, yes. Sometimes they get stuck in the process. I learned how to go around certain issues but it's very annoying.
20. viraptor ◴[] No.44619891[source]
Ask for no comments. Or extremely infrequent ones for complex sections only. Agreeing with https://news.ycombinator.com/item?id=44619462 here.
21. lightbulbish ◴[] No.44620215{3}[source]
_all_ models I’ve tried continuously, and still, have problems ignoring rules. I’m actually quite shocked someone would write this if you have experience in the area, as it so clearly contrasts with my own experience.

Despite explicit instructions in all sorts of rules and .md’s, the models still make changes where they should not. When caught they innocently say ”you’re right I shouldn’t have done that as it directly goes against your rule of <x>”.

Just to be clear, are you suggesting that currently, with your existing setup, the AI’s always follow your instructions in your rules and prompts? If so, I want your rules please. If not, I don’t understand why you would diss a solution which aims to hardcode away some of the llm prompt interpretation problems that exist

22. bradly ◴[] No.44620251[source]
This is way even as a paid user I stick to the browser tab llms. I am a context control freak and constantly just grabbing a new sessions and starting over. I don't try and fix a session and the incentives of a subscription vs api token payment model has inverse incentives.
23. vishvananda ◴[] No.44620666{3}[source]
I am by no means an AI skeptic. It is possible to encode all sorts of things into instructions, but I don’t think the future of programming is every individual constructing and managing artisan prompts. There are surely some new paradigms to be discovered here. A code locking interface seems like an interesting one to explore. I’m sure there are others.
24. orangecat ◴[] No.44622478{3}[source]
Or maybe it works well in some cases and not others?
25. gjadi ◴[] No.44622866{3}[source]
Or, you know, chmod -w XYZ.ts