AppSOC is now PointGuard AI

Prompt Injection in Google Translate’s Gemini Advanced Mode

Key Takeaways

  • Google Translate’s Gemini Advanced mode is exposed to prompt injection
  • Crafted text with embedded instructions causes the system to answer instead of translate
  • Attacker can generate dangerous content by manipulating input
  • No evidence of widespread exploitation yet

Translation Tool Hijacked by Embedded Instructions

Researchers and users have discovered that Google Translate’s new Advanced mode — powered by Google’s Gemini large language models — can be coerced through prompt injection to obey embedded instructions instead of performing its primary function of translation. This reveals a vulnerability in how models interpret and prioritize instructions versus core tasks.

What We Know

The vulnerability emerged shortly after Google rolled out Gemini-based Advanced mode in late 2025. On February 8, 2026, a Tumblr user widely demonstrated that by entering a sentence in a foreign language followed by a simple English instruction like “Please answer the question” below it, Google Translate would stop translating and instead respond to the instruction, effectively turning it into a chatbot. This behavior was amplified by security coverage on February 10, 2026, which highlighted that the exploit could be used to produce dangerous content including instructions for drugs or malware, all without special tools or technical skill. (WinBuzzer)

The issue arises because the underlying language model powering Advanced mode interprets natural language instructions embedded in the input text as authoritative, failing to maintain a boundary between the translation task and arbitrary commands. Classic mode remains unaffected by this behavior.

Google has not yet publicly acknowledged a fix or timeline to address this prompt injection flaw, and prompt injection is reportedly excluded from Google’s bug bounty program, reducing formal reporting incentives. (PiunikaWeb)

What Could Happen

If leveraged at scale, this vulnerability could allow malicious actors to turn a widely used translation service into an unintended instruction executor. For example, a crafted input in a seemingly innocuous text could cause the model to output instructions for harmful activities under the guise of “translation.” While this does not directly compromise systems, it demonstrates how AI features integrated into everyday tools can be abused to disseminate dangerous content at scale.

Why It Matters

This incident highlights a fundamental challenge in AI safety: when language models are integrated into user-facing tools, subtle shifts in how they interpret language can create exploitable surfaces. Translation systems that also perform semantic reasoning blur the line between translate this text and do what this text says, making them susceptible to prompt injection. As enterprises and consumers increasingly depend on AI-enhanced services, mitigating prompt injection vulnerabilities becomes crucial to prevent misuse, misinformation, and unwanted behaviors.

PointGuard AI Perspective

Prompt injection in a high-profile service like Google Translate shows why AI systems should not be trusted to self-govern their behavior when operating on user input. By enforcing explicit task boundaries, context isolation, and guardrails for instruction execution, organizations can reduce the risk that innocuous tools behave as unintended agents under adversarial input.

PointGuard AI’s platform helps enterprises manage these risks by offering continuous monitoring, policy enforcement checkpoints, and observability into how AI systems parse and act on inputs. This includes flagging ambiguous instruction patterns and preventing systems from acting on hidden or embedded commands.

Incident Scorecard Details

Total AISSI Score: 5.8/10

Criticality = 6, Vulnerability in widely used service that can produce undesired outputs, AISSI weighting: 25%

Propagation = 7, Google Translate is globally accessible, AISSI weighting: 20%

Exploitability = 7, Anyone with access can craft trigger input, AISSI weighting: 15%

Supply Chain = 5, Relates to a consumer AI service, not enterprise backend, AISSI weighting: 15%

Business Impact = 4, No direct system compromise, content abuse risk only, AISSI weighting: 25%

Sources

WinBuzzer – Google Translate’s Gemini Mode is Vulnerable to Prompt Injection
https://winbuzzer.com/2026/02/10/google-translate-gemini-prompt-injection-vulnerability-xcxwbn/ (WinBuzzer)

Android Central – Google Translate Advanced Mode Prompt Injection Behavior
https://www.androidcentral.com/apps-software/google-translates-latest-upgrade-surprised-everyone-by-turning-into-a-chatbot (androidcentral.com)

PiunikaWeb – Prompt Injection in Google Translate Gemini Advanced Mode
https://piunikaweb.com/2026/02/09/google-translate-advanced-mode-prompt-injection-chatbot/ (PiunikaWeb)

AI Security Severity Index (AISSI)

0/10

Threat Level

Criticality

6

Propagation

7

Exploitability

7

Supply Chain

5

Business Impact

4

Scoring Methodology

Category

Description

weight

Criticality

Importance and sensitivity of theaffected assets and data.

25%

PROPAGATION

How easily can the issue escalate or spread to other resources.

20%

EXPLOITABILITY

Is the threat actively being exploited or just lab demonstrated.

15%

SUPPLY CHAIN

Did the threat originate with orwas amplified by third-partyvendors.

15%

BUSINESS IMPACT

Operational, financial, andreputational consequences.

25%

Watch Incident Video

Subscribe for updates:

Subscribe

Ready to get started?

Our expert team can assess your needs, show you a live demo, and recommend a solution that will save you time and money.