←back to thread

Gemini CLI

(blog.google)
1336 points sync | 2 comments | | HN request time: 0.524s | source
Show context
cperry ◴[] No.44377336[source]
Hi - I work on this. Uptake is a steep curve right now, spare a thought for the TPUs today.

Appreciate all the takes so far, the team is reading this thread for feedback. Feel free to pile on with bugs or feature requests we'll all be reading.

replies(38): >>44377379 #>>44377463 #>>44377522 #>>44377570 #>>44377664 #>>44377874 #>>44378010 #>>44378095 #>>44378282 #>>44378477 #>>44378483 #>>44378624 #>>44378661 #>>44378918 #>>44378935 #>>44379294 #>>44379599 #>>44379809 #>>44379831 #>>44380039 #>>44380415 #>>44380918 #>>44380943 #>>44381462 #>>44381702 #>>44382998 #>>44383306 #>>44383505 #>>44384705 #>>44385575 #>>44385992 #>>44386024 #>>44386107 #>>44386388 #>>44387079 #>>44387136 #>>44387432 #>>44388145 #
danavar ◴[] No.44378477[source]
Is there a way to instantly, quickly prompt it in the terminal, without loading the full UI? Just to get a short response without filling the terminal page.

like to just get a short response - for simple things like "what's a nm and grep command to find this symbol in these 3 folders". I use gemini alot for this type of thing already

Or would that have to be a custom prompt I write?

replies(5): >>44378594 #>>44378617 #>>44378653 #>>44378886 #>>44380297 #
1. peterldowns ◴[] No.44378594[source]
I use `mods` for this https://github.com/charmbracelet/mods

other people use simon willison's `llm` tool https://github.com/simonw/llm

Both allow you to switch between models, send short prompts from a CLI, optionally attach some context. I prefer mods because it's an easier install and I never need to worry about Python envs and other insanity.

replies(1): >>44378637 #
2. indigodaddy ◴[] No.44378637[source]
Didn't know about mods, looks awesome.