←back to thread

Gemini CLI

(blog.google)
1336 points sync | 2 comments | | HN request time: 0.479s | source
Show context
iandanforth ◴[] No.44377207[source]
I love how fragmented Google's Gemini offerings are. I'm a Pro subscriber, but I now learn I should be a "Gemini Code Assist Standard or Enterprise" user to get additional usage. I didn't even know that existed! As a run of the mill Google user I get a generous usage tier but paying them specifically for "Gemini" doesn't get me anything when it comes to "Gemini CLI". Delightful!
replies(21): >>44377223 #>>44377234 #>>44377268 #>>44377361 #>>44377524 #>>44377663 #>>44377759 #>>44377818 #>>44378537 #>>44378591 #>>44378915 #>>44379216 #>>44380359 #>>44382462 #>>44383491 #>>44384124 #>>44384673 #>>44384735 #>>44385178 #>>44385385 #>>44386691 #
behnamoh ◴[] No.44377524[source]
Actually, that's the reason a lot of startups and solo developers prefer non-Google solutions, even though the quality of Gemini 2.5 Pro is insanely high. The Google Cloud Dashboard is a mess, and they haven't fixed it in years. They have Vertex that is supposed to host some of their models, but I don't understand what's the difference between that and their own cloud. And then you have two different APIs depending on the level of your project: This is literally the opposite of what we would expect from an AI provider where you start small and regardless of the scale of your project, you do not face obstacles. So essentially, Google has built an API solution that does not scale because as soon as your project gets bigger, you have to switch from the Google AI Studio API to the Vertex API. And I find it ridiculous because their OpenAI compatible API does not work all the time. And a lot of tools that rely on that actually don't work.

Google's AI offerings that should be simplified/consolidated:

- Jules vs Gemini CLI?

- Vertex API (requires a Google Cloud Account) vs Google AI Studio API

Also, since Vertex depends on Google Cloud, projects get more complicated because you have to modify these in your app [1]:

``` # Replace the `GOOGLE_CLOUD_PROJECT` and `GOOGLE_CLOUD_LOCATION` values # with appropriate values for your project. export GOOGLE_CLOUD_PROJECT=GOOGLE_CLOUD_PROJECT export GOOGLE_CLOUD_LOCATION=global export GOOGLE_GENAI_USE_VERTEXAI=True ```

[1]: https://cloud.google.com/vertex-ai/generative-ai/docs/start/...

replies(4): >>44377931 #>>44378027 #>>44378498 #>>44378744 #
tarvaina ◴[] No.44377931[source]
It took me a while but I think the difference between Vertex and Gemini APIs is that Vertex is meant for existing GCP users and Gemini API for everyone else. If you are already using GCP then Vertex API works like everything else there. If you are not, then Gemini API is much easier. But they really should spell it out, currently it's really confusing.

Also they should make it clearer which SDKs, documents, pricing, SLAs etc apply to each. I still get confused when I google up some detail and end up reading the wrong document.

replies(4): >>44378649 #>>44380895 #>>44382672 #>>44388982 #
fooster ◴[] No.44378649[source]
The other difference is that reliability for the gemini api is garbage, whereas for vertex ai it is fantastic.
replies(1): >>44381368 #
nikcub ◴[] No.44381368[source]
The key to running LLM services in prod is setting up Gemini in Vertex, Anthropic models on AWS Bedrock and OpenAI models on Azure. It's a completely different world in terms of uptime, latency and output performance.
replies(2): >>44383127 #>>44384669 #
1. com2kid ◴[] No.44384669[source]
Does OpenAI on azure still have that insane latency for content filtering? Last time I checked it added a huge # to time to first token, making azure hosting for real time scenarios impractical.
replies(1): >>44385890 #
2. shakna ◴[] No.44385890[source]
Yes.

Unless you convince MS to let you at the "Provisioned Throughput" model. Which also requires being big enough for sales to listen to you.