←back to thread

Gemini CLI

(blog.google)
1342 points sync | 1 comments | | HN request time: 0.413s | source
Show context
wohoef ◴[] No.44378022[source]
A few days ago I tested Claude Code by completely vibe coding a simple stock tracker web app in streamlit python. It worked incredibly well, until it didn't. Seems like there is a critical project size where it just can't fix bugs anymore. Just tried this with Gemini CLI and the critical project size it works well for seems to be quite a bit bigger. Where claude code started to get lost, I simply told Gemini CLI to "Analyze the codebase and fix all bugs". And after telling it to fix a few more bugs, the application simply works.

We really are living in the future

replies(8): >>44378198 #>>44378469 #>>44378677 #>>44378994 #>>44379068 #>>44379186 #>>44379685 #>>44384682 #
1. crazylogger ◴[] No.44379186[source]
Ask the AI to document each module in a 100-line markdown. These should be very high level, don't contain any detail, but just include pointers to relevant files for AI to find out by itself. With a doc as the starting point, AI will have context to work on any module.

If the module just can't be documented in this way in under 100 lines, it's a good time to refactor. Chances are if Claude's context window is not enough to work with a particular module, a human dev can't either. It's all about pointing your LLM precisely at the context that matters.