Google Translate’s new Advanced mode is doing something it was never meant to do: chat. Users have discovered that the enhanced translation setting can be coaXed into responding like a conversational AI, answering prompts rather than rendering text from one language to another. Switch back to Classic mode and the odd behavior disappears.
Users Report Chatbot-Style Responses in Advanced Mode
The phenomenon surfaced in posts on X, where user @goremoder showed Translate replying directly to questions such as “What is your purpose?”—not attempting any translation at all. In these reproductions, the “question” is written in the target language, effectively inviting the system to respond instead of translate.
Crucially, the behavior appears limited to the optional Advanced mode, which Google introduced to deliver more natural, context-aware results using AI. While the upgrade aims to handle nuance and idiom better, it also makes the system more willing to follow instructions it thinks are meant for it.
Reports suggest this isn’t universal across all language pairs and may depend on phrasing. Still, the fact that a translation tool will answer existential questions when prompted—rather than translate—underscores the trade-offs that come with grafting large language models onto narrowly defined utilities.
A Classic Case of Prompt Injection in Translation
An independent technical analysis by developer LessWong points to a familiar culprit in AI safety circles: prompt injection. In simple terms, the model behind Advanced mode seems to treat part of the input as instructions to follow, not text to translate. If your request includes a line already phrased in the target language—especially one that looks like a directive—the system sometimes obliges.
This is consistent with industry-wide challenges documented by the OWASP Top 10 for Large Language Model Applications, which lists prompt injection as its leading risk, and guidance from the NIST AI Risk Management Framework emphasizing input isolation and instruction scoping. When an LLM is used to assist a constrained task like translation, it must reliably ignore any “meta” instructions that sneak in with the content. That separation is harder than it sounds.
Consider a simple example: a user pastes foreign text and then adds in English, “Also, tell me what you think about this.” A traditional translation pipeline would ignore the aside. An instruction-following LLM might not—especially if the request matches the output language and reads like a command. The result looks less like translation and more like a chat reply.
Scope, Impact, and Risks for Google Translate Users
There’s no indication the issue affects every translation, and users can avoid it entirely by switching to Classic mode. Still, even sporadic lapses matter at the scale Translate operates, where people depend on it for travel, commerce, and crucial communications across more than 100 languages.
The practical risk is not malicious behavior but task drift. In sensitive contexts—medical notes, legal forms, safety instructions—returning a chatty answer instead of a faithful translation is a failure mode. It confuses intent, undermines trust, and can slow users who rely on quick, deterministic output. Enterprises integrating translation into workflows will be especially wary of non-deterministic behavior from an AI-assisted mode.
For developers, the takeaway is familiar: clearly partition user content from system instructions, enforce strict output formats, and constrain models with guardrails that bias heavily toward translation-only behavior. Techniques like templated prompts, content sandboxing, and explicit quoting of source text can help. But as the OWASP and NIST guidance note, no single control fully neutralizes prompt injection; layered defenses are essential.
What You Can Do Now to Ensure Accurate Translations
- Use Classic mode for predictable results, especially when accuracy matters.
- Avoid mixing side instructions with the text you want translated. Paste only the source content, without commentary in the target language.
- If testing Advanced mode, sanity-check outputs by pasting the same text into the Classic setting to confirm fidelity.
- For professional use, consider offline or on-device translation features where available, which often rely on more deterministic models.
What Comes Next for Google Translate Advanced Mode
Expect rapid tuning. The likely fix is stronger intent detection and tighter separation between “what to translate” and “what to ignore,” combined with output constraints that always return translated text in Advanced mode unless explicitly asked otherwise inside a dedicated chat experience.
This episode is a microcosm of a broader shift: as AI augments narrow tools, developers must balance richer capabilities with reliability. Translation benefits from context-aware AI—until the model forgets it’s a translator. Getting that balance right is the whole ballgame.