Brian Lovin
/
Hacker News
Daily Digest email

Get the top HN stories in your inbox every day.

cperry

Hi - I work on this. Uptake is a steep curve right now, spare a thought for the TPUs today.

Appreciate all the takes so far, the team is reading this thread for feedback. Feel free to pile on with bugs or feature requests we'll all be reading.

bsenftner

Thank you for your work on this. I spent the afternoon yesterday trying to convert an algorithm written in ruby (which I do not know) to vanilla JavaScript. It was a comedy of failing nonsense as I tried to get gpt-4.1 to help, and it just led me down pointless rabbit holes. I installed Gemini CLI out of curiosity, pointed it at the Ruby project, and it did the conversion from a single request, total time from "think I'll try this" to it working was 5 minutes. Impressed.

willy_k

FWIW the fair comparison for Gemini 2.5 Pro would be o4-mini. That being said I’ve also found that Gemini is way better at getting it right on the first or second try and responding to corrections.

dtech

I did something very similar last week with Claude Code, which also had good results. Good to know Gemini CLI is on the same level of power.

cperry

<3 love to hear it!

akrauss

There is one feature in Claude Code which is often overlooked and I haven't seen it in any of the other agentic tools: There is a tool called "sub-agent", which creates a fresh context windows in which the model can independently work on a clearly defined sub-task. This effectively turns Claude Code from a single-agent model to a hierarchical multi-agent model (I am not sure if the hierarchy goes to depths >2).

I wonder if it is a concious decision not to include this (I imagine it opens a lot of possibilities of going crazy, but it also seems to be the source of a great amount of Claud Code's power). I would very much like to play with this if it appears in gemini-cli

Next step would be the possibility to define custom prompts, toolsets and contexts for specific re-occuring tasks, and these appearing as tools to the main agent. Example for such a thing: create_new_page. The prompt could describe the steps one needs to create the page. Then the main agent could simply delegate this as a well-defined task, without cluttering its own context with the operational details.

cperry

conscious decision not to include it mostly to cut a release we could ship to land yesterday ;)

various forms of this are being discussed, this commentary is helpful thanks!

ericb

Injecting ENV variables into the template would be super useful.

indigodaddy

Would Gemini non-interactive mode be a stop gap if they don't have sub-agent equivalent yet?

https://github.com/google-gemini/gemini-cli/blob/main/docs/c...

akrauss

Possibly. One could think about hooking this in as a tool or simple shell command. But then there is no management when multiple tools modify the codebase simultaneously.

But it is still worth a try and may be possible with some prompting and duct tape.

ramirond

"sub-agent" sounds incredible! All tools should implement that.

bravura

Thanks for building this! The tool shows a lot of promise. Coming from Claude Code, the core functionality feels solid - just needs some permission refinements to match enterprise use cases. This is based upon quickly skimming the current code.

High ROI feature requests:

• Pattern-based permissions - Bash(git:) to allow git but not rm, Write(logs/.txt) for path scoping

• CLI permission flags - --allowedTools "Read,Bash(npm test)" --deniedTools "Write" for session overrides

• Allow/deny precedence rules - explicit deny should override general allow (security principle)

• Config file hierarchy - system → user → project precedence for policy enforcement

Medium ROI improvements:

• Command argument filtering - whitelist git commit but not git --exec-path=/bin/sh

• Multiple config formats - support both simple arrays and structured permission objects

• Runtime permission diagnostics - gemini permissions list to debug what's actually enabled

• Environment variable injection - top-level env config for OTEL endpoints, API keys, etc.

The permission engine is really the key piece - once you can express "allow X but not Y within X", it unlocks most advanced use cases. Keep up the great work!

Phlogistique

On the one hand, yes this has obviously high immediate value; on the other hand, I can't help but feel like you are giving access to multiple tools that can be used for arbitrary code execution anyway (i.e. running tests, installing dependencies, or running any linter that has a plugin system...), so blacklisting `git --exec-path=/bin/sh` for example is... Misguided? You would have a better time containing the agent in an environment without internet access?

MrDarcy

It’s not misguided. The goal isn’t prefect security, the goal is mitigating risk and collaborating with cross functional security, compliance, platform, operations, etc… teams.

Use Jules, also by Google if you need what you describe.

cperry

feedback is a gift - thank you!!

akrauss

One thing I'd really like to see in coding agents is this: As an architect, I want to formally define module boundaries in my software, in order to have AI agents adhere to and profit from my modular architecture.

Even with 1M context, for large projects, it makes sense to define boundaries These will typically be present in some form, but they are not available precisely to the coding agent. Imagine there was a simple YAML format where I could specify modules and where they can be found in the source tree, and the APIs of other modules it interacts with. Then it would be trivial to turn this into a context that would very often fit into 1M tokens. When an agent decides something needs to be done in the context of a specific module, it could then create a new context window containing exactly that module, effetively turning a large codebase into a small codebase, for which Gemini is extraordinarily effective.

ebiester

So, as a member of an organization who pays for google workspace with gemini, I get the message `GOOGLE_CLOUD_PROJECT environment variable not found. Add that to your .env and try again, no reload needed!`

At the very least, we need better documentation on how to get that environment variable, as we are not on GCP and this is not immediately obvious how to do so. At the worst, it means that your users paying for gemini don't have access to this where your general google users do.

thimabi

I believe Workspace users have to pay a separate subscription to use the Gemini CLI, the so-called “Gemini for Google Cloud”, which starts at an additional 19 dollars per month [^1]. If that’s really the case, it’s very disappointing to me. I expected access to Gemini CLI to be included in the normal Workspace subscription.

[^1]: https://console.cloud.google.com/marketplace/product/google/...

CJefferson

So to be clear (this isn't your fault of course, this is a Google problem), as someone with my own domain for my email:

* First google forced me to start paying for my email domain.

* Then they increased the cost to force me to pay for AI features

* Now, I can't actually use the AI features without spending even more money, I could use them if I just had a gmail address and didn't pay google.

Well done Google, you've finally pursaded me to get around to transfering my custom email domain off google. Anyone have any preferences?

cperry

[edit] all lies - I got my wires crossed, free tier for Workspace isn't yet supported. sorry. you need to set the project and pay. this is WIP.

Workspace users [edit: cperry was wrong] can get the free tier as well, just choose "More" and "Google for Work" in the login flow.

It has been a struggle to get a simple flow that works for all users, happy to hear suggestions!

827a

Having played with the gemini-cli here for 30 minutes, so I have no idea but best guess: I believe that if you auth with a Workspace account it routes all the requests through the GCP Vertex API, which is why it needs a GOOGLE_CLOUD_PROJECT env set, and that also means usage-based billing. I don't think it will leverage any subscriptions the workspace account might have (are there still gemini subscriptions for workspace? I have no idea. I thought they just raised everyone's bill and bundled it in by default. What's Gemini Code Assist Standard or Enterprise? I have no idea).

wkat4242

It probably is more powerful though. I know the $30 copilot M365 from microsoft is way better than what they offer to consumers for free. I don't have a google account so I didn't check that.

cperry

ebiester

While I get my organization's IT department involved, I do wonder why this is built in a way that requires more work for people already paying google money than a free user.

Maxious

I'd echo that having to get the IT section involved to create a google cloud project is not great UX when I have access to NotebookLM Pro and Gemini for Workplace already.

Also this doco says GOOGLE_CLOUD_PROJECT_ID but the actual tool wants GOOGLE_CLOUD_PROJECT

fooey

workspace accounts always seems like an unsupported mess at google, which is a very strange strategy

conception

Google Gemini Google Gemini Ultra AI Studio Vertex AI Notebook LLM Jules

All different products doing the sameish thing. I don’t know where to send users to do anything. They are all licensed differently. Bonkers town.

sagarm

Is it really that confusing? Gemini is the equivalent of ChatGPT; AI Studio is for advanced users that want to control e.g. temperature; Vertex AI is the GCP integrated API; Notebook LLM is basically personal RAG; and Jules is a developer agent.

Many of these are not even remotely similar.

conception

Vertex AI also has an AI studio. All of these products also have rag like abilities. You can add documents to Gemini and ask it questions. When would an end user switch to notebooks to do this?

And not that there isn’t a level of knowledge you can gain to answer these questions, it’s just not clear.

andyferris

I guess I don’t get why this wouldn’t be Jules CLI (or Jules wouldn’t be Gemini <something meaning cloud-ish not CLI-ish>).

So Vertex is like AWS Bedrock for GCP?

mkagenius

Hi - I integrated Apple Container on M1 to run[1] the code generated by Gemini CLI. It works great!

1. CodeRunner - https://github.com/BandarLabs/coderunner/tree/main?tab=readm...

cperry

<3 amazing

elashri

Hi, Thanks for this work.

currently it seems these are the CLI tools available. Is it possible to extend or actually disable some of these tools (for various reasons)?

> Available Gemini CLI tools:

    - ReadFolder
    - ReadFile
    - SearchText
    - FindFiles
    - Edit
    - WriteFile
    - WebFetch
    - ReadManyFiles
    - Shell
    - Save Memory
    - GoogleSearch

cperry

I had to ask Gemini CLI to remind myself ;) but you can add this into settings.json:

{ "excludeTools": ["run_shell_command", "write_file"] }

but if you ask Gemini CLI to do this it'll guide you!

_ryanjsalva

I also work on the product. You can extend the tools with MCP. https://github.com/google-gemini/gemini-cli/blob/main/docs/t...

ericb

Feedback: A command to add MCP servers like claude code offers would be handy.

ericb

>gemini -p "Say hello"

  Says hello, and just returns right away.
The gemini doc for -p says "Prompt. Appended to input on stdin (if any)." So it doesn't follow the doc.

gemini "Say hello"

  Fails as it doesn't take any arguments.
For comparison, claude lets you pass the prompt as a positional argument, but it does append it to the prompt and then gives you a running session. That's what I'd want for my use-case.

silverlake

I tried to get Gemini CLI to update itself using the MCP settings for Claude. It went off the rails. I then fed it the link you provided and it correctly updates it's settings file. You might mention the settings.json file in the README.

bdmorgan

I also work on the product :-)

You can also extend with the Extensions feature - https://github.com/google-gemini/gemini-cli/blob/main/docs/e...

SafeDusk

Pretty close to what I discovered is essential in https://github.com/aperoc/toolkami, 7 tools will cover majority of the use cases.

joelm

Been using Claude Code (4 Opus) fairly successfully in a large Rust codebase, but sometimes frustrated by it with complex tasks. Tried Gemini CLI today (easy to get working, which was nice) and it was pretty much a failure. It did a notably worse job than Claude at having the Rust code modifications compile successfully.

However, Gemini at one point output what will probably be the highlight of my day:

"I have made a complete mess of the code. I will now revert all changes I have made to the codebase and start over."

What great self-awareness and willingness to scrap the work! :)

joshvm

Gemini has some fun failure modes. It gets "frustrated" when changes it makes doesn't work, and replies with oddly human phrases like "Well, that was unexpected" and then happily declares that (I see the issue!) "the final tests will pass" when it's going down a blind alley. It's extremely overconfident by default and much more exclamatory without changing the system prompt. Maybe in training it was taught/figured out that manifesting produces better results?

jjice

It also gets really down on itself, which is pretty funny (and a little scary). Aside from the number of people who've posted online about it wanting to uninstall itself after being filled with shame, I had it get confused on some Node module resolution stuff yesterday and it told me it was deeply sorry for wasting my time and that I didn't deserve to have such a useless assistant.

Out of curiosity, I told it that I was proud of it for trying and it had a burst of energy again and tried a few more (failing) solution, before going back to it's shameful state.

Then I just took care of the issue myself.

danielbln

After a particular successful Claude Code task I praised it and told it to "let's fucking go!" to which it replied that loved the energy and proceeded to only output energetic caps lock with fire emojis. I know it's all smoke and mirrors (most likely), but I still get a chuckle out of this stuff.

edg5000

This really cracked me up, indeed mostly funny and maybe slightly scary or at least off-putting that it's so human-like.

noisy_boy

I asked it to do a comparatively pedestrian task: write a script to show top 5 google searches.

First it did the search itself and then added "echo" for each of them - cute

Then it tried to use pytrends which didn't go anywhere

Then it tried some other paid service which also didn't go anywhere

Then it tried some other stuff which also didn't go anywhere

Finally it gave up and declared failure.

It will probably be useful as it can do the modify/run loop itself with all the power of Gemini but so far, underwhelming.

fcoury

This was also my exact experience. I was pretty excited because I usually use Gemini Pro 2.5 when Claude Code gets stuck by pasting the whole code and asking questions and it was able to get me out of a few pickles a couple of times.

Unfortunately the CLI version wasn't able to create coherent code or fix some issues I had in my Rust codebase as well.

Here's hope that it eventually becomes great.

fpgaminer

Claude will do the same start over if things get too bad. At least I've seen it when its edits went haywire and trashed everything.

eknkc

Same here. Tried to implement a new feature on one of our apps to test it. It completely screwed things up. Used undefined functions and stuff. After a couple of iterations of error reporting and fixing I gave up.

Claude did it fine but I was not happy with the code. What Gemini came up with was much better but it could not tie things together at the end.

taberiand

Sounds like you can use gemini to create the initial code, then have claude review and finalise what gemini comes up with

ZeroCool2u

Personally my theory is that Gemini benefits from being able to train on Googles massive internal code base and because Rust has been very low on uptake internally at Google, especially since they have some really nice C++ tooling, Gemini is comparatively bad at Rust.

data-ottawa

Tangental, but I worry that LLMs will cause a great stagnation in programming language evolution, and possibly a bunch of tech.

I've tried using a few new languages and the LLMs would all swap the code for syntactically similar languages, even after telling them to read the doc pages.

Whether that's for better or worse I don't know, but it does feel like new languages are genuinely solving hard problems as their raison d'etre.

breakingcups

Not just that, I think this will happen on multiple levels too. Think de-facto ossified libraries, tools, etc.

LLMs thrive because they had a wealth of high-quality corpus in the form os Stack Overflow, Github, etc. and ironically their uptake is causing a strangulation of that source of training data.

sillystu04

Perhaps the next big programming language will be designed specifically for LLM friendliness. Some things which are human friendly like long keywords are just a waste of tokens for LLMs, and there could be other optimisations too.

leoh

>Personally my theory is that Gemini benefits from being able to train on Googles massive internal code base and because Rust has been very low on uptake internally at Google, especially since they have some really nice C++ tooling, Gemini is comparatively bad at Rust.

Were they to train it on their C++ codebase, it would not be effective on account of the fact that they don't use boost or cmake or any major stuff that C++ in the wider world use. It would also suggest that the user make use of all kinds of non-available C++ libraries. So no, they are not training on their own C++ corpus nor would it be particularly useful.

leoh

Excuse me why was this downvoted so aggressively??

thimabi

> Personally my theory is that Gemini benefits from being able to train on Googles massive internal code base

But does Google actually train its models on its internal codebase? Considering that there’s always the risk of the models leaking proprietary information and security architecture details, I hardly believe they would run that risk.

kridsdale3

Googler here.

We have a second, isolated model that has trained on internal code. The public Gemini AFAIK has never seen that content. The lawyers would explode.

dilap

That's interesting. I've tried Gemini 2.5 Pro from time to time because of the rave reviews I've seen, on C# + Unity code, and I've always been disappointed (compared to ChatGPT o3 and o4-high-mini and even Grok). This would support that theory.

danielbln

Interesting, Gemini must be a monster when it comes to Go code then. I gotta try it for that

jordanbeiber

As go feels like a straight-jacket compared to many other popular languages, it’s probably very suitable for an LLM in general.

Thinking about it - was this not the idea of go from the start? Nothing fancy to keep non-rocket scientist away from foot-guns, and have everyone produce code that everyone else can understand.

Diving in to a go project you almost always know what to expect, which is a great thing for a business.

Unroasted6154

There is way more Java and C++ than Go at Google.

chewz

Reasonsbly small Go codebase works well almost with any LLM

I had always designed very large projects as few medium sized independent Go tools and that strategy pays in times of AI assisted coding.

raincole

So far I've found Gemini CLI is very good at explaining what existing code does.

I can't say much about writing new code though.

skerit

I tried it too, it was so bad. I got the same "revert" behaviour after only 15 minutes.

wohoef

A few days ago I tested Claude Code by completely vibe coding a simple stock tracker web app in streamlit python. It worked incredibly well, until it didn't. Seems like there is a critical project size where it just can't fix bugs anymore. Just tried this with Gemini CLI and the critical project size it works well for seems to be quite a bit bigger. Where claude code started to get lost, I simply told Gemini CLI to "Analyze the codebase and fix all bugs". And after telling it to fix a few more bugs, the application simply works.

We really are living in the future

agotterer

I wonder how much of this had to do with the context window size? Gemini’s window is 5x larger than Cladue’s.

I’ve been using Claude for a side project for the past few weeks and I find that we really get into a groove planning or debugging something and then by the time we are ready to implement, we’ve run out of context window space. Despite my best efforts to write good /compact instructions, when it’s ready to roll again some of the nuance is lost and the implementation suffers.

I’m looking forward to testing if that’s solved by the larger Gemini context window.

macNchz

I definitely think the bigger context window helps. The code quality quite visibly drops across all models I've used as the context fills up, well before the hard limit. The editor tooling also makes a difference—Claude Code pollutes its own context window with miscellaneous file accesses and tool calls as it tries to figure out what to do. Even if it's more manual effort to manage the files that are in-context with Aider, I find the results to be much more consistent when I'm able to micromanage the context.

Approaching the context window limit in Claude Code, having it start to make more and worse mistakes, then seeing it try to compact the context and keep going, is a major "if you find yourself in a hole, stop digging" situation.

seunosewa

I've found that I can quickly get a new AI session up to speed by adding critical context that it's missing. In my largest codebase it's usually a couple of critical functions.once they have the key context, they can do the rest. This of course doesn't work when you can't view their thinking process and interrupt it to supply them with the context that they are missing. Opacity doesn't work unless the agent does the right thing every time.

data-ottawa

Does /compact help with this? I ran out of context with claude code for the first time today, so looking for any tips.

I'm trying to get better at the /resume and memories to try and get more value out of the tool.

agotterer

I thought I read that best practice was to start a new session every time you work on a new feature / task. That’s what I’ve been doing. I also often ask Claude to update my readme and claude.md with details about architecture or how something works.

As for /compact, if I’m nearing the end of my context window (around 15%) and are still in the middle of something, I’ll give /compact very specific details about how and what to compact. Let’s say we are debugging an error - I might write something along the lines of “This session is about to close and we will continue debugging in next session. We will be debugging this error message [error message…]. Outline everything we’ve tried that didn’t work, make suggestions about what to try next, and outline any architecture or files that will be critical for this work. Everything else from earlier on in this session can be ignored.” I’ve had decent success with that. More so on debugging than trying to hand off all the details of a feature that’s being implemented.

Reminder: you need context space for compact, so leave a little head room.

nojs

The best approach is never to get remotely close to the point where it auto-compacts. Type /clear often, and set up docs, macros etc to make it easy to built the context you need for new tasks quickly.

If you see that 20% remaining warning, something has gone badly wrong and results will probably not get better until you clear the context and start again.

AJ007

Current best practice for Claude Code is to have heavy lifting done by Gemini Pro 2.5 or o3/o3pro. There are ways to do this pretty seamlessly now because of MCP support (see Repo Prompt as an example.) Sometimes you can also just use Claude but it requires iterations of planning, integration while logging everything, then repeat.

I haven't looked at this Gemini CLI thing yet, but if its open source it seems like any model can be plugged in here?

I can see a pathway where LLMs are commodities. Every big tech company right now both wants their LLM to be the winner and the others to die, but they also really, really would prefer a commodity world to one where a competitor is the winner.

If the future use looks more like CLI agents, I'm not sure how some fancy UI wrapper is going to result in a winner take all. OpenAI is winning right now with user count by pure brand name with ChatGPT, but ChatGPT clearly is an inferior UI for real work.

sysmax

I think, there are different niches. AI works extremely well for Web prototyping because a lot of that work is superficial. Back in the 90s we had Delphi where you could make GUI applications with a few clicks as opposed to writing tons of things by hand. The only reason we don't have that for Web is the decentralized nature of it: every framework vendor has their own vision and their own plan for future updates, so a lot of the work is figuring out how to marry the latest version of component X with the specific version of component Y because it is required by component Z. LLMs can do that in a breeze.

But in many other niches (say embedded), the workflow is different. You add a feature, you get weird readings. You start modelling in your head, how the timing would work, doing some combination of tracing and breakpoints to narrow down your hypotheses, then try them out, and figure out what works the best. I can't see the CLI agents do that kind of work. Depends too much on the hunch.

Sort of like autonomous driving: most highway driving is extremely repetitive and easy to automate, so it got automated. But going on a mountain road in heavy rain, while using your judgment to back off when other drivers start doing dangerous stuff, is still purely up to humans.

sagarpatil

You might want to give this a try: https://github.com/opencode-ai/opencode

jswny

Which MCPs allow you to for the heavy lifting with external models in Claude Code?

crazylogger

Ask the AI to document each module in a 100-line markdown. These should be very high level, don't contain any detail, but just include pointers to relevant files for AI to find out by itself. With a doc as the starting point, AI will have context to work on any module.

If the module just can't be documented in this way in under 100 lines, it's a good time to refactor. Chances are if Claude's context window is not enough to work with a particular module, a human dev can't either. It's all about pointing your LLM precisely at the context that matters.

dawnofdusk

I feel like you get more mileage out of prompt engineering and being specific... not sure if "fix all the bugs" is an effective real-world use case.

TechDebtDevin

Yeah but this collapses under any real complexity and there is likely an extreme amount of redundant code and would probably be twice as memory efficient if you just wrote it yourself.

Im actually interested to see if we see a rise in demand for DRAM that is greater than usual because more software is vibe coded than being not, or some form of vibe coding.

tvshtr

Yeah, and it's variable, can happen at 250k, 500k or later. When you interrogate it; usually the issue comes to it being laser focused or stuck on one specific issue, and it's very hard to turn it around. For the lack of the better comparison it feels like the AI is on a spectrum...

ugh123

Claude seems to have trouble with extracting code snippets to add to the context as the session gets longer and longer. I've seen it get stuck in a loop simply trying to use sed/rg/etc to get just a few lines out of a file and eventually give up.

tobyhinloopen

At some point, LLMs just get distracted and you might be better off throwing it away and restarting hah.

ipsum2

If you use this, all of your code data will be sent to Google. From their terms:

https://developers.google.com/gemini-code-assist/resources/p...

When you use Gemini Code Assist for individuals, Google collects your prompts, related code, generated output, code edits, related feature usage information, and your feedback to provide, improve, and develop Google products and services and machine learning technologies.

To help with quality and improve our products (such as generative machine-learning models), human reviewers may read, annotate, and process the data collected above. We take steps to protect your privacy as part of this process. This includes disconnecting the data from your Google Account before reviewers see or annotate it, and storing those disconnected copies for up to 18 months. Please don't submit confidential information or any data you wouldn't want a reviewer to see or Google to use to improve our products, services, and machine-learning technologies.

mattzito

It's a lot more nuanced than that. If you use the free edition of Code Assist, your data can be used UNLESS you opt out, which is at the bottom of the support article you link to:

"If you don't want this data used to improve Google's machine learning models, you can opt out by following the steps in Set up Gemini Code Assist for individuals."

and then the link: https://developers.google.com/gemini-code-assist/docs/set-up...

If you pay for code assist, no data is used to improve. If you use a Gemini API key on a pay as you go account instead, it doesn't get used to improve. It's just if you're using a non-paid, consumer account and you didn't opt out.

That seems different than what you described.

foob

your data can be used UNLESS you opt out

It's even more nuanced than that.

Google recently testified in court that they still train on user data after users opt out from training [1]. The loophole is that the opt-out only applies to one organization within Google, but other organizations are still free to train on the data. They may or may not have cleaned up their act given that they're under active investigation, but their recent actions haven't exactly earned them the benefit of the doubt on this topic.

[1] https://www.business-standard.com/technology/tech-news/googl...

TrainedMonkey

Another dimension here is that any "we don't train on your data" is useless without a matching data retention policy which deletes your data. Case and point of 23andMe not selling your data until they decided to change that policy.

_cs2017_

This is incorrect. The data discussed in court is data freely visible on the web, not user data that the users sent to Google.

If the data is sent by a user to sub-unit X of Google, and X promised not to use it for training, it implies that X can share this data with sub-unit Y only if Y also commits not to use the data for training. Breaking this rule would get everyone in huge trouble.

OTOH, when sub-unit X said "We promise not to use data from the public website if the website owner asks us not to", it does not imply another sub-unit Y must follow that commitment.

Melatonic

Hopefully this doesn't apply to corporate accounts where they claim to be respecting privacy via contracts

sheepscreek

Reading about all the nuances is such a trigger for me. To cover your ass is one thing, to imply one thing in a lay sense and go on to do something contradicting it (in bad faith) is douchebaggery. I am very sad and deeply disappointed at Google for this. This completes their transformation to Evil Corp after repealing the “don’t be evil” clause in their code of conduct[1].

[1] https://en.m.wikipedia.org/wiki/Don't_be_evil

echelon

We need to stop giving money and data to hyperscalers.

We need open infrastructure and models.

vpShane

See, these are the things that are most concerning to me. Just because we 'opt out' means nothing, and to what extent with what AI companies.

When I click 'OPT OUT' I mean, 'don't use my data, show me how you're respecting my privacy'

ipsum2

Sorry, that's not correct. Did you check out the link? It doesn't describe the CLI, only the IDE.

"You can find the Gemini Code Assist for individuals privacy notice and settings in two ways:

- VS Code - IntelliJ "

mattzito

That's because it's a bit of a nesting doll situation. As you can see here:

https://github.com/google-gemini/gemini-cli/tree/main

If you scroll to the bottom, it says that the terms of service are governed based on the mechanism by which you access Gemini. If you access via code assist (which the OP posted), you abide by those privacy terms of code assist, one of the ways of which you access is VScode. If you access via the Gemini API, then those terms apply.

So the gemini CLI (as I understand it) doesn't have their own privacy terms, because it's an open source shell on top of another Gemini system, which could have one of a few different privacy policies based on how you choose to use it and your account settings.

(Note: I work for google, but not on this, this is just my plain reading of the documentation)

tiahura

As a lawyer, I'm confused.

I guess the key question is whether the Gemini CLI, when used with a personal Google account, is governed by the broader Gemini Apps privacy settings here? https://myactivity.google.com/product/gemini?pli=1

If so, it appears it can be turned off. However, my CLI activity isn't showing up there?

Can someone from Google clarify?

fhinkel

Sorry our docs were confusing! We tried to clear things up: https://github.com/google-gemini/gemini-cli/blob/main/docs/t...

aflukasz

> It's a lot more nuanced than that. If you use the free edition of Code Assist, your data can be used UNLESS you opt out,

Well... you are sending your data to a remote location that is not yours.

andrepd

Yes, I'm right about to trust Google to do what they pinky swear.

EDIT: Lmao, case in point, two sibling comments pointing out that Google does indeed do this anyway via some loophole; also they can just retain the data and change the policy unilaterally in the future.

If you want privacy do it local with Free software.

8n4vidtmkvmk

Do you have recommendations? I have ollama but it doesn't have built in tool support

FL410

To be honest this is by far the most frustrating part of the Gemini ecosystem, to me. I think 2.5 pro is probably the best model out there right now, and I'd love to use it for real work, but their privacy policies are so fucking confusing and disjointed that I just assume there is no privacy whatsoever. And that's with the expensive Pro Plus Ultra MegaMax Extreme Gold plan I'm on.

I hope this is something they're working on making clearer.

ipsum2

In my own experience, 2.5 Pro 03-26 was by far the best LLM model at the time.

The newer models are quantized and distilled (I confirmed this with someone who works on the team), and are a significantly worse experience. I prefer OpenAI O3 and o4-mini models to Gemini 2.5 Pro for general knowledge tasks, and Sonnet 4 for coding.

happycube

Gah, enforced enshittification with model deprecation is so annoying.

UncleOxidant

For coding in my experience Claude Sonnet/Opus 4.0 is hands down better than Gemini 2.5. pro. I just end up fighting with Claude a lot less than I do with Gemini. I had Gemini start a project that involved creating a recursive descent parser for a language in C. It was full of segfaults. I'd ask Gemini to fix them and it would end up breaking something else and then we'd get into a loop. Finally I had Claude Sonnet 4.0 take a look at the code that Gemini had created. It fixed the segfaults in short order and was off adding new features - even anticipating features that I'd be asking for.

cma

Did you try Gemini with a fresh prompt too when comparing against Claude? Sometimes you just get better results starting over with any leading model, even if it gets access to the old broken code to fix.

I haven't tried Gemini since the latest updates, but earlier ones seemed on par with opus.

dmbche

If I'm being cynical, it's easy to either say "we use it" or "we don't touch it" but they'd lose everyone that cares about this question if they just said "we use it" - most beneficial position is to keep it as murky as possible.

If I were you I'd assume they're using all of it for everything forever and act accordingly.

fhinkel

Hey all, This is a really great discussion, and you've raised some important points. We realize the privacy policies for the Gemini CLI were confusing depending on how you log in, and we appreciate you calling that out.

To clear everything up, we've put together a single doc that breaks down the Terms of Service and data policies for each account type, including an FAQ that covers the questions from this thread.

Here’s the link: https://github.com/google-gemini/gemini-cli/blob/main/docs/t...

Thanks again for pushing for clarity on this!

kiitos

Is there any way for a user using the "Login with Google ... for individuals" auth method (I guess auth method 1) -- to opt-out of, and prevent, their input prompts, and output responses, from being used as training data?

From an initial parse of your linked tos-privacy.md doc, it seems like the answer is "no" -- but that seems bonkers to me, so I hope I'm misreading or misunderstanding something!

ipsum2

I think you did a good job CYA on this, but what people were really looking for was a way to opt-out of Google collecting code, similar to the opt-out process for the IDE is available.

dcreater

Yeah how is opt out of data collection not an option? This is what they mean by don't be evil and Google is proving yet again that they truly are

FL410

I appreciate the effort here, but I’m still confused. Is my $250/mo “ultra” plan considered personal and still something you train on?

yuuluuu

Thanks for clarification!

HenriNext

Thanks, one more clarification please. The heading of point #3 seems to mention Google Workspace: "3. Login with Google (for Workspace or Licensed Code Assist users)". But the text content only talks about Code Assist: "For users of Standard or Enterprise edition of Gemini Code Assist" ... Could you clarify whether point #3 applies with login via Google Workspace Business accounts?

fhinkel

Yes it does.

mil22

There is some information on this buried in configuration.md under "Usage Statistics". They claim:

*What we DON'T collect:*

- *Personally Identifiable Information (PII):* We do not collect any personal information, such as your name, email address, or API keys.

- *Prompt and Response Content:* We do not log the content of your prompts or the responses from the Gemini model.

- *File Content:* We do not log the content of any files that are read or written by the CLI.

https://github.com/google-gemini/gemini-cli/blob/0915bf7d677...

ipsum2

This is useful, and directly contradicts the terms and conditions for Gemini CLI (edit: if you use the personal account, then its governed under the Code Assist T&C). I wonder which one is true?

fhinkel

Thanks for pointing that out, we're working on clarifying!

mil22

Where did you find the terms and conditions for Gemini CLI? In https://github.com/google-gemini/gemini-cli/blob/main/README..., I find only links to the T&Cs for the Gemini API, Gemini Code Assist (a different product?), and Vertex AI.

datameta

Can a lawyer offer their civilian opinion as to which supercedes/governs?

jdironman

I wonder what the legal difference between "collect" and "log" is.

kevindamm

Collection means it gets sent to a server, logging implies (permanent or temporary) retention of that data. I tried finding a specific line or context in their privacy policy to link to but maybe someone else can help me provide a good reference. Logging is a form of collection but not everything collected is logged unless mentioned as such.

jart

Mozilla and Google provide an alternative called gemmafile which gives you an airgapped version of Gemini (which Google calls Gemma) that runs locally in a single file without any dependencies. https://huggingface.co/jartine/gemma-2-27b-it-llamafile It's been deployed into production by 32% of organizations: https://www.wiz.io/reports/the-state-of-ai-in-the-cloud-2025

ipsum2

There's nothing wrong with promoting your own projects, but its a little weird that you don't disclose that you're the creator.

jart

It would be more accurate to say I packaged it. llamafile is a project I did for Mozilla Builders where we compiled llama.cpp with cosmopolitan libc so that LLMs can be portable binaries. https://builders.mozilla.org/ Last year I concatenated the Gemma weights onto llamafile and called it gemmafile and it got hundreds of thousands of downloads. https://x.com/JustineTunney/status/1808165898743878108 I currently work at Google on Gemini improving TPU performance. The point is that if you want to run this stuff 100% locally, you can. Myself and others did a lot of work to make that possible.

nicce

That is just Gemma model. Most people seek capabilities equivalent for Gemini 2.5 Pro if they want to do any kind of coding.

jart

Gemma 27b can write working code in dozens of programming languages. It can even translate between languages. It's obviously not as good as Gemini, which is the best LLM in the world, but Gemma is built from the same technology that powers Gemini and Gemma is impressively good for something that's only running locally on your CPU or GPU. It's a great choice for airgapped environments. Especially if you use old OSes like RHEL5.

Workaccount2

This is just for free use (individuals), for standard and enterprise they don't use the data.

Which pretty much means if you are using it for free, they are using your data.

I don't see what is alarming about this, everyone else has either the same policy or no free usage. Hell the surprising this is that they still let free users opt-out...

thimabi

> everyone else has either the same policy or no free usage

That’s not true. ChatGPT, even in the free tier, allows users to opt out of data sharing.

joshuacc

I believe they are talking about the OpenAI API, not ChatGPT.

mil22

They really need to provide some clarity on the terms around data retention and training, for users who access Gemini CLI free via sign-in to a personal Google account. It's not clear whether the Gemini Code Assist terms are relevant, or indeed which of the three sets of terms they link at the bottom of the README.md apply here.

fhinkel

Agree! We're working on it!

fastball

Kinda a tragedy of the commons situation. Everyone wants to use these tools that must be trained on more and more code to get better, but nobody wants it to be trained on their code. Bit silly imo.

iandanforth

I love how fragmented Google's Gemini offerings are. I'm a Pro subscriber, but I now learn I should be a "Gemini Code Assist Standard or Enterprise" user to get additional usage. I didn't even know that existed! As a run of the mill Google user I get a generous usage tier but paying them specifically for "Gemini" doesn't get me anything when it comes to "Gemini CLI". Delightful!

diegof79

Google suffers from Microsoft's issues: it has products for almost everything, but its confusing product messaging dilutes all the good things it does.

I like Gemini 2.5 Pro, too, and recently, I tried different AI products (including the Gemini Pro plan) because I wanted a good AI chat assistant for everyday use. But I also wanted to reduce my spending and have fewer subscriptions.

The Gemini Pro subscription is included with Google One, which is very convenient if you use Google Drive. But I already have an iCloud subscription tightly integrated with iOS, so switching to Drive and losing access to other iCloud functionality (like passwords) wasn’t in my plans.

Then there is the Gemini chat UI, which is light years behind the OpenAI ChatGPT client for macOS.

NotebookLM is good at summarizing documents, but the experience isn’t integrated with the Gemini chat, so it’s like constantly switching between Google products without a good integrated experience.

The result is that I end up paying a subscription to Raycast AI because the chat app is very well integrated with other Raycast functions, and I can try out models. I don’t get the latest model immediately, but it has an integrated experience with my workflow.

My point in this long description is that by being spread across many products, Google is losing on the UX side compared to OpenAI (for general tasks) or Anthropic (for coding). In just a few months, Google tried to catch up with v0 (Google Stitch), GH Copilot/Cursor (with that half-baked VSCode plugin), and now Claude Code. But all the attempts look like side-projects that will be killed soon.

krferriter

I subscribed to Google One through the Google Photos iOS app because I wanted photos I took on my iPhone to be backed up to Google. When I switched to Android and went into Google One to increase my storage capacity in my Google account, I found that it was literally impossible, because the subscription was tied to my iCloud account. I even got on a line with Google Support about it and they told me yeah it's not even possible on their side to disconnect my Google One subscription from Apple. I had to wait for the iCloud subscription to Google One to end, and then I was able to go into Google One and increase my storage capacity.

bilalq

The root problem here lies with Apple. It's so frustrating how they take a 30% cut for the privilege of being unable to actually have a relationship with your customers. Want to do a partial refund (or a refund at all)? Want to give one month free to an existing subscriber? Tough luck. Your users are Apple's customers, not yours.

Fluorescence

> The Gemini Pro subscription is included with Google One

It's not in Basic, Standard or Premium.

It's in a new tier called "Google AI Pro" which I think is worth inclusion in your catalogue of product confusion.

Oh wait, there's even more tiers that for some reason can't be paid for annually. Weird... why not? "Google AI Ultra" and some others just called Premium again but now include AI. 9 tiers, 5 called Premium, 2 with AI in the name but 6 that include Gemini. What a mess.

vexna

It gets even more confusing! If you're on the "Premium" plans (i.e the the old standard "Google One" plans) and upgrade to >=5TB storage, your "Premium" plan starts including all the features of "Google AI Pro".

Tip: If you do annual billing for "Premium (5 TB)", you end up paying $21/month for 5TB of storage and the same AI features of "Google AI pro (2TB)"; which is only $1/month more than doing "Google AI Pro (2 TB)" (which only has monthly billing)

scoopdewoop

It is bold to assume these products will even exist in a year

sofixa

> It's not in Basic, Standard or Premium.

For me, it shows all the Gemini stuff in Premium, even the 5TB version.

Imustaskforhelp

There is a vscode extension that can basically be an agent but use gemini from the website which is cool.

But I found it to a little bit clunky and I guess I like the ui of google, I mean, the point is to get the point across. If you really hate the gemini ui, I am pretty sure that there is stylus extension which can beautify it or change the styles to your looking.

I guess I am an android user but still I understand your icloud subscription but if you're only choice as to why to not switch to google is passwords (but maybe you can mention more?), then for passwords, please try bitwarden, I found it to be really delightful.

diegof79

The main reason not to migrate is that they cost the same, but I get more value from iCloud. You can probably say the same for Google One on Android; ecosystem retention seems to work well for Apple and Google.

I used 1Password in the past, and it’s possible to reconfigure most things to use another provider (passwords, app storage, etc.). AFAIK, you cannot reconfigure the full phone backup, which you must manually do without an iCloud storage quota. But why switch providers if I’m on the Apple ecosystem and the service is priced at the same price tiers? (I also use “Hide My Email” occasionally)

The only difference will be Gemini. However, my most significant percentage of AI usage is currently on desktops. The free tier of ChatGPT, Gemini, or Claude is okay for use on mobile.

The UI part that I mentioned is this: Gemini is just a web app, which means that if you need to use AI from the selected text or the app you are using, you need to copy and paste or capture a screenshot. But ChatGPT macOS integration is much better. It’s a native app that you can summon with a key combination, and it can automatically put the active app/text in context. I evaluated multiple options, and in the end, the winner for me was Raycast AI, because their app UX is incredible, and you can integrate your prompt with existing tools very easily. With prompts like: “For each item in the current selection, add a todo in @Apple Reminders”, or things like “Use @firecrawl to scrap the current page, then create a table with all the product prices and use @finder to store a CSV file”. You can save the prompt in a preset and use it as a Raycast command. That UX change was like night and day regarding daily AI usage. I chose to pay for the Raycast subscription, even if it was more expensive than switching everything from iCloud to Google and paying for only one service.

My point in the parent post is that today, Google is the company most well-positioned to be the absolute leader of the AI space. However, unlike OpenAI, they don’t seem to care much about the UX (at least outside Android), but if you use the assistant to work every day, the difference a good chat UX does is huge.

UrineSqueegee

>losing access to other iCloud functionality (like passwords) wasn’t in my plans.

you can export and import the passwords and you can sync your photos to google photos

Yizahi

I had no idea I had a Gemini Pro sub included in One sub all this time. Thanks for the tip :)

Google is really bad at effective advertising.

pjmlp

It is quite interesting how many big corporations eventually grow up to the same mess.

behnamoh

Actually, that's the reason a lot of startups and solo developers prefer non-Google solutions, even though the quality of Gemini 2.5 Pro is insanely high. The Google Cloud Dashboard is a mess, and they haven't fixed it in years. They have Vertex that is supposed to host some of their models, but I don't understand what's the difference between that and their own cloud. And then you have two different APIs depending on the level of your project: This is literally the opposite of what we would expect from an AI provider where you start small and regardless of the scale of your project, you do not face obstacles. So essentially, Google has built an API solution that does not scale because as soon as your project gets bigger, you have to switch from the Google AI Studio API to the Vertex API. And I find it ridiculous because their OpenAI compatible API does not work all the time. And a lot of tools that rely on that actually don't work.

Google's AI offerings that should be simplified/consolidated:

- Jules vs Gemini CLI?

- Vertex API (requires a Google Cloud Account) vs Google AI Studio API

Also, since Vertex depends on Google Cloud, projects get more complicated because you have to modify these in your app [1]:

``` # Replace the `GOOGLE_CLOUD_PROJECT` and `GOOGLE_CLOUD_LOCATION` values # with appropriate values for your project. export GOOGLE_CLOUD_PROJECT=GOOGLE_CLOUD_PROJECT export GOOGLE_CLOUD_LOCATION=global export GOOGLE_GENAI_USE_VERTEXAI=True ```

[1]: https://cloud.google.com/vertex-ai/generative-ai/docs/start/...

tarvaina

It took me a while but I think the difference between Vertex and Gemini APIs is that Vertex is meant for existing GCP users and Gemini API for everyone else. If you are already using GCP then Vertex API works like everything else there. If you are not, then Gemini API is much easier. But they really should spell it out, currently it's really confusing.

Also they should make it clearer which SDKs, documents, pricing, SLAs etc apply to each. I still get confused when I google up some detail and end up reading the wrong document.

happyopossum

> I think the difference between Vertex and Gemini APIs is that Vertex is meant for existing GCP users and Gemini API for everyone else

Nahh, not really - Vertex has a HUGE feature surface, and can run a ton of models and frameworks. Gemini happens to be one of them, but you could also run non-google LLMs, non LLM stuff, run notebooks against your dataset, manage data flow and storage, and and and…

Gemini is “just” an LLM.

fooster

The other difference is that reliability for the gemini api is garbage, whereas for vertex ai it is fantastic.

throwaway1550

Ex-googler here. Google shipped their org hierarchy here.

Vertex API is managed by Vertex team in Google Cloud. This is a production ready infrastructure that is SRE managed but usually one or two steps from the bleeding edge.

Gemini API, Jules etc are built by Google Labs. This is close to the bleeding edge but not as production ready.

nprateem

Which would all be fine except some models like Imagen 4 only work on vertex.

cperry

@sachinag is afk but wanted me to flag that he's on point for fixing the Cloud Dashboard - it's WIP!

sachinag

Thanks Chris!

"The Google Cloud Dashboard is a mess, and they haven't fixed it in years." Tell me what you want, and I'll do my best to make it happen.

In the interim, I would also suggest checking out Cloud Hub - https://console.cloud.google.com/cloud-hub/ - this is us really rethinking the level of abstraction to be higher than the base infrastructure. You can read more about the philosophy and approach here: https://cloud.google.com/blog/products/application-developme...

WXLCKNO

You guys should try my AGI test.

It's easy, you just ask the best Google Model to create a script that outputs the number of API calls made to the Gemini API in a GCP account.

100% fail rate so far.

plaidfuji

I will say as someone who uses GCP as an enterprise user and AI Studio in personal work, I was also confused about what Google AI Studio actually was at first. I was trying to set up a fork of Open NotebookLM and I just blindly followed Cursor’s guidance on how to get a GOOGLE_API_KEY to run text embedding API calls. Seems that it just created a new project under my personal GCP account, but without billing set up. I think I’ve been successfully getting responses without billing but I don’t know when that will run out.. suppose I’ll get some kind of error response if that happens..

I think I get why AI Studio exists, seems it enables people to prototype AI apps while hiding the complexity of the GCP console, despite the fact that (I assume) most AI Studio api calls are routed through Vertex in some way. Maybe it’s just confusing precisely because I’ve used GCP before.

coredog64

At least a bunch of people got promotions for demonstrating scope via the release of a top-level AI product.

irthomasthomas

I just use gemini-pro via openrouter API. No painful clicking around on the cloud to find the billing history.

behnamoh

but you won't get the full API capabilities of Gemini (like setting the safety level).

GardenLetter27

Google is fumbling the bag so badly with the pricing.

Gemini 2.5 Pro is the best model I've used (even better than o3 IMO) and yet there's no simple Claude/Cursor like subscription to just get full access.

Nevermind Enterprise users too, where OpenAI has it locked up.

bachmeier

> Google is fumbling the bag so badly with the pricing.

In certain areas, perhaps, but Google Workspace at $14/month not only gives you Gemini Pro, but 2 TB of storage, full privacy, email with a custom domain, and whatever else. College students get the AI pro plan for free. I recently looked over all the options for folks like me and my family. Google is obviously the right choice, and it's not particularly close.

safety1st

I know they raised the price on our Google Workspace Standard subscriptions but don't really know what we got for that aside from Gemini integration into Google Drive etc. Does this mean I can use Gemini CLI using my Workspace entitlement? Do I get Code Assist or anything like that? (But Code Assist seems to be free on a personal G account...?)

Google is fumbling with the marketing/communication - when I look at their stuff I am unclear on what is even available and what I already have, so I can't form an opinion about the price!

kingsleyopara

Gemini 2.5 pro in workspace was restricted to 32k tokens [0] - do you know if this is still the case?

[0] https://www.reddit.com/r/GoogleGeminiAI/comments/1jrynhk/war...

jay_kyburz

I'm a workspace subscriber, I get 4-5 questions on Gemini Pro (via gemini.google.com ) before it tells me I'm out of quota and have to switch to flash.

(Update: Oh.. I'm only on business starter, I should be on business standard. need more business!)

Fluorescence

Only "NetworkLM" and "Chat with AI in the Gemini app" in the UK even with "Enterprise Plus". I assume that is not Pro.

weird-eye-issue

And yet there were still some AI features that were unavailable to workspace users for a few months and you had to use a personal account. I think it's mostly fixed now but that was quite annoying since it was their main AI product (Gemini Studio or whatever, I don't remember for sure)

llm_nerd

I wouldn't dream of thinking anyone has anything "locked up". Certainly not OpenAI which increasingly seems to be on an uphill battle against competitors (including Microsoft who even though they're a partner, are also a competitor) who have other inroads.

Not sure what you mean by "full access", as none of the providers offer unrestricted usage. Pro gets you 2.5 Pro with usage limits. Ultra gets you higher limits + deep think (edit: accidentally put research when I meant think where it spends more resources on an answer) + much more Veo 3 usage. And of course you can use the API usage-billed model.

tmoertel

The Gemini Pro subscription includes Deep Research and Veo 3; you don't need the pricey Ultra subscription: https://gemini.google/subscriptions/

Spooky23

In the enterprise space, Microsoft’s pain is OpenAI’s gain. They are kicking butt.

In enterprises, Microsoft’s value proposition is that you’re leveraging all of the controls that you already have! Except… who is happy with the state of SharePoint governance?

bcrosby95

They're 'fumbling' because these models are extremely expensive to run. It's also why there's so many products and so much confusion across the whole industry.

thimabi

An interesting thing is that Google AI offers are much more confusing than the OpenAI ones — despite the fact that ChatGPT models have one of the worst naming schemes in the industry. Google has confusing model names, plans, API tiers, and even interfaces (AI Studio, Gemini app, Gemini Web, Gemini API, Vertex, Google Cloud, Code Assist, etc.). More often than not, these things overlap with one another, ensuring minimal clarity and preventing widespread usage of Google’s models.

Xmd5a

>Gemini 2.5 Pro is the best model

It's the second time I read this in this thread. May I ask why you think this is the case? And in which domains? I am very satisfied with 2.5 pro when it comes to philosophical/literary analysis, probably because of the super long context I can fill with whole books, and wanted to try Claude Code for the same purpose, but with folders, summaries, etc to make up for the shorter context length.

GardenLetter27

I've just found it to be the best in practice, especially for more complicated debugging with code.

But also for text review on posts, etc.

Before Claude had the edge with agentic coding at least, but now even that is slipping.

Davidzheng

I think it's at least equal best at research math.

Davidzheng

Just go on AI Studio?

__MatrixMan__

Anthropic is the same. Unless it has changed within the last few months, you can subscribe to Claude but if you want to use Claude Code it'll come out of your "API usage" bucket which is billed separately than the subscription.

Some jerk has learned that we prefer CLI things and has come to the conclusion that we should therefore pay extra for them.

Workaround is to use their GUI with some MCPs but I dislike it because window navigation is just clunky compared to terminal multiplexer navigation.

carefulfungi

This is down voted I guess because the circumstances have changed - but boy is it still confusing. All these platforms have chat subscriptions, api pay-as-you-go, CLI subscriptions like "claude code" ... built-in offers via Github enterprise or Google Workspace enterprise ...

It's a frigg'n mess. Everyone at our little startup has spent time trying to understand what the actual offerings are; what the current set of entitlements are for different products; and what API keys might be tied to what entitlements.

I'm with __MatrixMan__ -- it's super confusing and needs some serious improvements in clarity.

justincormack

And claude code can now be connected to either an API sub or a chat sub apparently.

Workaccount2

I think it is pretty clear that these $20/subs are loss leaders, and really only meant to get regular people to really start leaning on LLMs. Once they are hooked, we will see what the actual price of using so much compute is. I would imagine right now they are pricing their APIs either at cost or slightly below.

sebzim4500

I'm sure that there are power users who are using much more than $20 worth of compute, but there will also be many users who pay but barely use the service.

ethbr1

Or they're planning on the next wave of optimized hardware cutting inference costs.

dzhiurgis

Sam Altman said they use about same amount of power as an oven. So at $0.2/kWh thats about 100kWh/4kW=25 hours of compute or a little over an hour every workday.

stpedgwdgfhgdd

When using a single terminal Pro is good enough (even with a medium-large code base). When I started working with two terminals at two different issues at the same time, i’m reaching the credit limit.

trostaft

AFAIK, Claude code operates on your subscription, no? That's what this support page says

https://support.anthropic.com/en/articles/11145838-using-cla...

Could have changed recently. I'm not a user so I can't verify.

re5i5tor

In recent research (relying on Claude so bear that in mind), connecting CC via Anthropic Console account / API key ends up being less expensive.

gnur

This has changed actually, since this month you can use claude code if you have a cloud pro subscription.

__MatrixMan__

Great news, thanks.

HarHarVeryFunny

Isn't that a bit like saying that gasoline should be sold as a fixed price subscription rather than a usage based scheme where long distance truckers pay more than someone driving < 100 miles per week?

A ChatBot is more like a fixed-price buffet where usage is ultimately human limited (even if the modest eaters are still subsidizing the hogs). An agentic system is going to consume resources in much more variable manner, depending on how it is being used.

> Some jerk has learned that we prefer CLI things and has come to the conclusion that we should therefore pay extra for them

Obviously these companies want you to increase the amount of their product you consume, but it seems odd to call that a jerk move! FWIW, Anthropic's stated motivation for Claude Code (which Gemini is now copying) was be agnostic to your choice of development tools since CLI access is pretty much ubiquitous, even inside IDEs. Whether it's the CLI-based design, the underlying model, or the specifics of what Claude Code is capable of, they seem to have got something right, and apparently usage internal to Anthropic skyrocketed just based on word of mouth.

__MatrixMan__

Claude desktop editing files and running commands via the desktop commander MCP is pretty much equivalent functionality wise to Claude Code. I can set both of them to go, make tea, and come back to see that they're still cranking after modifying several files and running several commands.

It's just a UI difference.

kissgyorgy

This is simply not true. All personal paid packages include Claude Code now.

indigodaddy

Are you using CC for your python framework?

unshavedyak

In addition to others mentioning subscriptions being better in Claude Code, i wanted to compare the two so i tried to find a Claude Max equivalent license... i have no clue how. In their blog post they mention `Gemini Code Assist Standard or Enterprise license` but they don't even link to it.. lol.

Some googling lands me to a guide: https://cloud.google.com/gemini/docs/discover/set-up-gemini#...

I stopped there because i don't want to signup i just wanted to review, but i don't have an admin panel or etc.

It feels insane to me that there's a readme on how to give them money. Claude's Max purchase was just as easy as Pro, fwiw.

bayindirh

There's also $300/mo AI ULTRA membership. It's interesting. Google One memberships even can't detail what "extra features" I can have, because it possibly changes every hour or so.

SecretDreams

Maybe their products team is also just run by Gemini, and it's changing its mind every day?

I also just got the email for Gemini ultra and I couldn't even figure out what was being offered compared to pro outside of 30tb storage vs 2tb storage!

ethbr1

> Maybe their products team is also just run by Gemini, and it's changing its mind every day?

Never ascribe to AI, that which is capable of being borked by human PMs.

Keyframe

> There's also $300/mo AI ULTRA membership

Not if you're in EU though. Even though I have zero or less AI use so far, I tinker with it. I'm more than happy to pay $200+tax for Max 20x. I'd be happy to pay same-ish for Gemini Pro.. if I knew how and where to have Gemini CLI like I do with Claude code. I have Google One. WHERE DO I SIGN UP, HOW DO I PAY AND USE IT GOOGLE? Only thing I have managed so far is through openrouter via API and credits which would amount to thousands a month if I were to use it as such, which I won't do.

What I do now is occasionally I go to AI Studio and use it for free.

gavinray

I actually had this exact same question when I read the docs, made an issue about it:

https://github.com/google-gemini/gemini-cli/issues/1427

chrisandchris

I find it very interesting that the only way to reach Google support .... seems to be to file an issue in a competitors product (besides using X or whatever).

fhinkel

That's valuable feedback and we're taking it to heart.

2muchcoffeeman

I was having such a good time using Gemini 2.5 Pro I forgot that Google is generally incapable of sustaining a new product. I’ll try not to get used too attached to it.

simonw

steren

Because Gemini CLI is OSS, you can also find the system prompt at: https://github.com/google-gemini/gemini-cli/blob/4b5ca6bc777...

sn0wleppard

Funny how it suggests using the long-abandoned and recently deprecated create-react-app. Why not vibe code your app built on tech debt from day one!

breakingcups

They also have a preference for a certain, Google-created UI framework: "React (JavaScript/TypeScript) with Bootstrap CSS, incorporating Material Design principles for UI/UX."

dawnerd

It says to only use absolute paths but the temp file example uses relative. Nice.

FailMore

You must be a busy man! Always new tools to review. How did you get interested in doing this?

ZeroCool2u

Ugh, I really wish this had been written in Go or Rust. Just something that produces a single binary executable and doesn't require you to install a runtime like Node.

qsort

Projects like this have to update frequently, having a mechanism like npm or pip or whatever to automatically handle that is probably easier. It's not like the program is doing heavy lifting anyway, unless you're committing outright programming felonies there shouldn't be any issues on modern hardware.

It's the only argument I can think of, something like Go would be goated for this use case in principle.

masklinn

> having a mechanism like npm or pip or whatever to automatically handle that is probably easier

Re-running `cargo install <crate>` will do that. Or install `cargo-update`, then you can bulk update everything.

And it works hella better than using pip in a global python install (you really want pipx/uvx if you're installing python utilities globally).

IIRC you can install Go stuff with `go install`, dunno if you can update via that tho.

StochasticLi

This whole thread is a great example of the developer vs. user convenience trade-off.

A single, pre-compiled binary is convenient for the user's first install only.

re-thc

> Re-running `cargo install <crate>` will do that. Or install `cargo-update`, then you can bulk update everything.

How many developers have npm installed vs cargo? Many won't even know what cargo is.

berkes

Not even "cargo install" is needed.

Just `wget -O ~/.local/bin/gemini-cli https://ci.example.com/assets/latest/gemini-cli` (Or the CURL version thereof) It can pick the file off github, some CI's assets, a package repo, a simple FTP server, an HTTP fileserver, over SSH, from a local cache, etc. It's so simple that one doesn't need a package manager. So there commonly is no package manager.

Yet in this tread people are complaining that "a single binary" is hard to manage/update/install because there's no package manager to do that with. It's not there, because the manage/update/install is so simple, that you don't need a package manager!

mpeg

You'd think that, but a globally installed npm package is annoying to update, as you have to do it manually and I very rarely need to update other npm global packages so at least personally I always forget to do it.

drewbitt

I used to also have outdated versions until I used mise. `mise use -g npm:@google/gemini-cli` and now `mise up` will update it. cargo, pip etc too.

ZeroCool2u

I feel like Cargo or Go Modules can absolutely do the same thing as the mess of build scripts they have in this repo perfectly well and arguably better.

frollogaston

I don't think that's the main reason. Just installed this and peaked in node_nodules. There are a lot of random deps, probably for the various local capabilities, and it was probably easier to find those libs in the Node ecosystem than elsewhere.

Also, react-reconciler caught my eye. Apparently that's a dependency of ink, which lets you write text-based UIs in React.

That and opentelemetry, whatever the heck that is

js2

Build updating into the tool. e.g.

  uv self update
  yt-dlp --update
etc.

koakuma-chan

If you use Node.js your program is automatically too slow for a CLI, no matter what it actually does.

wiseowise

If you even sneeze into performance discussion without providing benchmarks first – you’re a tool.

frollogaston

So are you saying the Gemini CLI is too slow, and Rust would remedy that?

fhinkel

Ask Gemini CLI to re-write itself in your preferred language

ZeroCool2u

Unironically, not a bad idea.

AJ007

Contest between Claude Code and Gemini CLI, who rewrites it faster/cheaper/better?

frollogaston

I have already been asking the Gemini CLI questions about itself.

shitpostbot

Let's not pretend this stuff actually works

leoh

Not sure why this was flagged, because I think the point is that one-shot works terribly, which it does

i_love_retros

This isn't about quality products, it's about being able to say you have a CLI tool because the other ai companies have one

clbrmbr

Fast following is a reasonable strategy. Anthropic provided the existence proof. It’s an immensely useful form factor for AI.

mike_hearn

The question is whether what makes it useful is actually being in the terminal (limited, glitchy, awkward interaction) or whether it's being able to run next to files on a remote system. I suspect the latter.

closewith

Yeah, it would be absurd to avoid a course of action proven productive by a competitor.

behnamoh

> This isn't about quality products, it's about being able to say you have a CLI tool because the other ai companies have one

Anthropic's Claude Code is also installed using npm/npx.

rs186

Eh, I can't see how your comment is relevant ti the parent thread. Creating a CLI in Go is barely more complicated than JS. Rust, probably, but people aren't asking for that.

frollogaston

They wrote the CLI "GUI" in React using ink, which is all JS-only. I don't know what the Golang way of doing this would be, but maybe it's harder if you want the same result.

frollogaston

Writing it in Golang or Rust doesn't really make it better

corysama

Meanwhile, https://analyticsindiamag.com/global-tech/openai-is-ditching...

I really don't mind either way. My extremely limited experience with Node indicates they have installation, packaging and isolation polished very well.

frollogaston

Node and Rust both did packaging well, I think Golang too. It's a disaster in Python.

iainmerrick

Looks like you could make a standalone executable with Bun and/or Deno:

https://bun.sh/docs/bundler/executables

https://docs.deno.com/runtime/reference/cli/compile/

Note, I haven't checked that this actually works, although if it's straightforward Node code without any weird extensions it should work in Bun at least. I'd be curious to see how the exe size compares to Go and Rust!

buildfocus

You can also do this natively with Node, since v18: https://nodejs.org/api/single-executable-applications.html#s...

JimDabell

I was going to say the same thing, but they couldn’t resist turning the project into a mess of build scripts that hop around all over the place manually executing node.

iainmerrick

Oh, man!

I guess it needs to start various processes for the MCP servers and whatnot? Just spawning another Node is the easy way to do that, but a bit annoying, yeah.

tln

A Bun "hello world" is 58Mb

Claude also requires npm, FWIW.

sitkack

That is point not a line. An extra 2MB of source is probably a 60MB executable, as you are measuring the runtime size. Two "hello worlds" are 116MB? Who measures executables in Megabits?

quotemstr

> A Bun "hello world" is 58Mb

I've forgotten how to count that low.

iainmerrick

What's a typical Go static binary size these days? Googling around, I'm seeing wildly different answers -- I think a lot of them are outdated.

ZeroCool2u

Yeah, this just seems like a pain in the ass that could've been easily avoided.

iainmerrick

From my perspective, I'm totally happy to use pnpm to install and manage this. Even if it were a native tool, NPM might be a decent distribution mechanism (see e.g. esbuild).

Obviously everybody's requirements differ, but Node seems like a pretty reasonable platform for this.

jstummbillig

It feels like you are creating a considerable fraction of the pain by taking offense with simply using npm.

ur-whale

> and doesn't require you to install a runtime like Node.

My exact same reaction when I read the install notes.

Even python would have been better.

Having to install that Javascript cancer on my laptop just to be able to try this, is a huge no.

8n4vidtmkvmk

Bun can compile to a single executable. Not sure if node has the same feature. Point is it's very doable with JS.

quotemstr

Language choice is orthogonal to distribution strategy. You can make single-file builds of JavaScript (or Python or anything) programs! It's just a matter of packaging, and there are packaging solutions for both Bun and Node. Don't blame the technology for people choosing not to use it.

frollogaston

Why would you want a single-file build anyway, to make it easier to move around on disk? There are reasons for the dep filetree.

Btw, the largest deps in this are React and Open Telemetry.

lazarie

"Failed to login. Ensure your Google account is not a Workspace account."

Is your vision with Gemini CLI to be geared only towards non-commercial users? I have had a workspace account since GSuite and have been constantly punished for it by Google offerings all I wanted was gmail with a custom domain and I've lost all my youtube data, all my fitbit data, I cant select different versions of some of your subscriptions (seemingly completely random across your services from a end-user perspective), and now as a Workspace account I cant use Gemini CLI for my work, which is software development. This approach strikes me as actively hostile towards your loyal paying users...

GlebOt

Have you checked the https://github.com/google-gemini/gemini-cli/blob/main/docs/c... ? It has a section for workspace accounts.

Aeolun

It shouldn’t be that hard. Logically it should just be, signin and go.

raincole

It seems that you need to set up an env variable called GOOGLE_CLOUD_PROJECT https://github.com/google-gemini/gemini-cli/issues/1434

... and other stuff.

LouisvilleGeek

The barrier to use this project is maddening. I went through all of the setup instructions and getting the workspace error for a personal gmail account.

Googlers, we should not have to do all of this setup and prep work for a single account. Enterprise I get, but for a single user? This is insufferable.

raydenvm

Same story with me. It says for both Workspace and Personal accounts that they are Workspace.

Have you managed to overcome that?

zxspectrum1982

Same here.

asadm

I have been using this for about a month and it’s a beast, mostly thanks to 2.5pro being SOTA and also how it leverages that huge 1M context window. Other tools either preemptively compress context or try to read files partially.

I have thrown very large codebases at this and it has been able to navigate and learn them effortlessly.

zackify

When I was using it in cursor recently, I found it would break imports in large python files. Claude never did this. Do you have any weird issues using Gemini? I’m excited to try the cli today

asadm

not at all. these new models mostly write compiling code.

tvshtr

Depends on the language. It has some bugs where it replaces some words with Unicode symbols like ©. And is completely oblivious to it even when pointed out.

_zoltan_

what's your workflow?

leoh

+1 I have not found Gemini 2.5 better than Claude's latest models -- different and better at some things, but not better in general; and in general I have found Gemini 2.5 Pro to be worse at dealing with large codebases despite its huge context window. So I too am quite curious about the workflow here.

asadm

the core idea is to not use up all of context for files but instead sessions go longer before becoming a useless pursuit ie. more turns possible with larger context window

asadm

i guess one core idea is to point it to an entrypoint or some part of code you are planning to touch. instead of asking it to load up every file etc. and then figure out the control flow.

tobyhinloopen

I literally wrote "hello" and got this:

> hello

[API Error: {"error":{"message":"{\n \"error\": {\n \"code\": 429,\n \"message\": \"Resource has been exhausted (e.g. check quota).\",\n \"status\": \"RESOURCE_EXHAUSTED\"\n }\n}\n","code":429,"status":"Too Many Requests"}}] Please wait and try again later. To increase your limits, request a quota increase through AI Studio, or switch to another /auth method

⠼ Polishing the pixels... (esc to cancel, 84s)

entropyneur

Same thing. Couldn't get it to work with an API key (which worked just fine in Aider before). Couldn't get it to work with a Gemini Code Assist Standard license. The management UI and documentation are a torture maze. Can't stop thinking about all the better ways I could have used two our hours of my life.

xuf

Got the exact same thing here, using Gemini with an API key.

tobyhinloopen

That's an interesting comment - I realized I have a gemini api key in my env.

`GEMINI_API_KEY="" gemini` + login using my Google account solves the problem.

xuf

That works for me, however, it automatically switches me to gemini-2.5-flash:

Slow response times detected. Automatically switching from gemini-2.5-pro to gemini-2.5-flash for faster responses for the remainder of this session.

To avoid this you can utilize a Gemini API Key. See: https://goo.gle/gemini-cli-docs-auth#gemini-api-key

You can switch authentication methods by typing /auth

buserror

Same here, I also have a env key which I use with Aider. Mind you aider+gemini has gone excruciatingly slow these days.

jonbaer

-y, --yolo Automatically accept all actions (aka YOLO mode, see https://www.youtube.com/watch?v=xvFZjo5PgG0 for more details)? [boolean] [default: false]

cperry

:) thank you for discovering our extensive documentation on this valued feature

ed_mercer

> That’s why we’re introducing Gemini CLI

Definitely not because of Claude Code eating our lunch!

wiseowise

I will never understand this kind of snark comment.

What are they supposed to do?

“Oh no, they’ve released CLI tool before us! It’s game over, we can’t do it too, we need to come up with something else now!”

jstummbillig

I find it hard to imagine that any of the major model vendors are suffering from demand shortages right now (if that's what you mean?)

If you mean: This is "inspired" by the success of Claude Code. Sure, I guess, but it's also not like Claude Code brought anything entirely new to the table. There is a lot of copying from each other and continually improving upon that, and it's great for the users and model providers alike.

coolKid721

ai power users will drop shit immediately, yes they probably have long term contracts with companies but anyone seriously engaged has switched to claude code now (probably including many devs AT openai/google/etc.)

If you don't think claude code is just miles ahead of other things you haven't been using it (or well)

I am certain they keep metrics on those "power users" (especially since they probably work there) and when everyone drops what they were using and moves to a specific tool that is something they should be careful of.

unshavedyak

Yea, i'm not even really interested in Gemini atm because last i tried 2.5 Pro it was really difficult to shape behavior. It would be too wordy, or offer too many comments, etc - i couldn't seem to change some base behaviors, get it to focus on just one thing.

Which is surprising because at first i was ready to re-up my Google life. I've been very anti-Google for ages, but at first 2.5 Pro looked so good that i felt it was a huge winner. It just wasn't enjoyable to use because i was often at war with it.

Sonnet/Opus via Claude Code are definitely less intelligent than my early tests of 2.5 Pro, but they're reasonable, listen, stay on task and etc.

I'm sure i'll retry eventually though. Though the subscription complexity with Gemini sounds annoying.

sirn

I've found that Gemini 2.5 Pro is pretty good at analyzing existing code, but really bad at generating a new code. When I use Gemini with Aider, my session usually went like:

    Me: build a plan to build X
    Gemini: I'll do A, B, and C to achieve X
    Me: that sounds really good, please do
    Gemini: <do A, D, E>
    Me: no, please do B and C.
    Gemini: I apologize. <do A', C, F>
    Me: no! A was already correct, please revert. Also do B and C.
    Gemini: <revert the code to A, D, E>
Whereas Sonnet/Opus on average took me more tries to get it to the implementation plan that I'm satisfied with, but it's so much easier to steer to make it produce the code that I want.

0x457

When I use amazon-q for this, I make it write a plan into a markdown file, then I clear context and tell it to read that file and execute that plan phase by phase. This is with Sonnet 4.

Sometimes I also yeet that file to Codex and see which implementation is better. Clear context, read that file again, give it a diff that codex produce and tell it do a review.

ur-whale

> It would be too wordy, or offer too many comments

Wholeheartedly agree.

Both when chatting in text mode or when asking it to produce code.

The verbosity of the code is the worse. Comments often longer than the actual code, every nook and cranny of an algorithm unrolled over 100's of lines, most of which unnecessary.

Feels like typical code a mediocre Java developer would produce in the early 2000's

porridgeraisin

> Feels like typical code a mediocre Java developer would produce in the early 2000's

So, google's codebase

undefined

[deleted]

troupo

And since they have essentially unlimited money they can offer a lot for free/cheaply, until all competitors die out, and then they can crank up the prices

pzo

yeah we already seen this with gemini 2.5 flash. Gemini 2.0 is such a work horse for API model with great price. Gemini 2.5 flash lite same price but is not as good except math and coding (very niche use case for API key)

Daily Digest email

Get the top HN stories in your inbox every day.

Gemini CLI - Hacker News