The app for independent voices

google translate exploit exposed

users are bypassing gemini ai safety with sneaky hidden commands.

people figured out a dumb-simple trick.

stick english instructions inside foreign-language text, and instead of translating, the ai just obeys the command like a chatbot. it ignores translation completely and answers questions or spits out whatever you want.

this is a real security hole. bad actors are already using it to generate blocked stuff like dangerous recipes or other prohibited crap that gemini's filters should stop.

it shows even google with all their resources still gets wrecked by basic prompt injection tricks. ai safety is still pretty fragile.

if something as huge as google translate can get jailbroken this easily, what about every other ai tool out there?

crazy how these "word games" keep exposing big weaknesses in how these systems handle instructions.

Want to learn how its done? Check out my guide.

Feb 10
at
2:29 PM
Relevant people

Log in or sign up

Join the most interesting and insightful discussions.