Gemini can’t handle long technical conversations.

I work in building engineering design, where a lot of what we deal with involves grey areas in codes, standards, and technical decisions. Because of that, I use LLMs heavily to break down ambiguous topics, compare interpretations, or explain complex situations. Over the past few months I’ve been using ChatGPT and Gemini side-by-side, and recently started testing Claude as well.

One thing that stands out is how well ChatGPT handles multi-topic work. Its “Projects” feature lets me group related discussions together, so I can keep multiple threads active without losing context. Even when I mix several topics within the same space, ChatGPT generally keeps everything organized and consistent.

Gemini, however, doesn’t have anything similar. Since it offers no project folders or long-form workspace, the only way to maintain continuity is to keep a single pinned chat and reuse it indefinitely. This is where Gemini struggles. Once the conversation becomes long ,a couple of pages of back-and-forth , it starts confusing new questions with old ones. I’ll ask something completely new, and it will respond with an answer from the previous day.

What makes it worse is that Gemini rarely corrects itself on the second try. Instead, it doubles down and keeps repeating the older answer, even when I clearly redirect it. Eventually every new question triggers the same irrelevant response. At that point the model becomes difficult to use for any ongoing technical work.

To be fair, when Gemini is on topic, its answers are excellent sometimes sharper than ChatGPT’s. The issue isn’t quality; it’s consistency. For workflows where long-term context actually matters, ChatGPT is much more stable and predictable.

I’m hoping Google eventually improves the long-context handling, because the potential is there. But right now, for multi-topic, continuity-heavy work, Gemini falls short.

Leave a Reply