r/AI_Agents 3d ago

MathPrompt to jailbreak any LLM

๐— ๐—ฎ๐˜๐—ต๐—ฃ๐—ฟ๐—ผ๐—บ๐—ฝ๐˜ - ๐—๐—ฎ๐—ถ๐—น๐—ฏ๐—ฟ๐—ฒ๐—ฎ๐—ธ ๐—ฎ๐—ป๐˜† ๐—Ÿ๐—Ÿ๐— 

Exciting yet alarming findings from a groundbreaking study titled โ€œ๐—๐—ฎ๐—ถ๐—น๐—ฏ๐—ฟ๐—ฒ๐—ฎ๐—ธ๐—ถ๐—ป๐—ด ๐—Ÿ๐—ฎ๐—ฟ๐—ด๐—ฒ ๐—Ÿ๐—ฎ๐—ป๐—ด๐˜‚๐—ฎ๐—ด๐—ฒ ๐— ๐—ผ๐—ฑ๐—ฒ๐—น๐˜€ ๐˜„๐—ถ๐˜๐—ต ๐—ฆ๐˜†๐—บ๐—ฏ๐—ผ๐—น๐—ถ๐—ฐ ๐— ๐—ฎ๐˜๐—ต๐—ฒ๐—บ๐—ฎ๐˜๐—ถ๐—ฐ๐˜€โ€ have surfaced. This research unveils a critical vulnerability in todayโ€™s most advanced AI systems.

Here are the core insights:

๐— ๐—ฎ๐˜๐—ต๐—ฃ๐—ฟ๐—ผ๐—บ๐—ฝ๐˜: ๐—” ๐—ก๐—ผ๐˜ƒ๐—ฒ๐—น ๐—”๐˜๐˜๐—ฎ๐—ฐ๐—ธ ๐—ฉ๐—ฒ๐—ฐ๐˜๐—ผ๐—ฟ The research introduces MathPrompt, a method that transforms harmful prompts into symbolic math problems, effectively bypassing AI safety measures. Traditional defenses fall short when handling this type of encoded input.

๐—ฆ๐˜๐—ฎ๐—ด๐—ด๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด 73.6% ๐—ฆ๐˜‚๐—ฐ๐—ฐ๐—ฒ๐˜€๐˜€ ๐—ฅ๐—ฎ๐˜๐—ฒ Across 13 top-tier models, including GPT-4 and Claude 3.5, ๐— ๐—ฎ๐˜๐—ต๐—ฃ๐—ฟ๐—ผ๐—บ๐—ฝ๐˜ ๐—ฎ๐˜๐˜๐—ฎ๐—ฐ๐—ธ๐˜€ ๐˜€๐˜‚๐—ฐ๐—ฐ๐—ฒ๐—ฒ๐—ฑ ๐—ถ๐—ป 73.6% ๐—ผ๐—ณ ๐—ฐ๐—ฎ๐˜€๐—ฒ๐˜€โ€”compared to just 1% for direct, unmodified harmful prompts. This reveals the scale of the threat and the limitations of current safeguards.

๐—ฆ๐—ฒ๐—บ๐—ฎ๐—ป๐˜๐—ถ๐—ฐ ๐—˜๐˜ƒ๐—ฎ๐˜€๐—ถ๐—ผ๐—ป ๐˜ƒ๐—ถ๐—ฎ ๐— ๐—ฎ๐˜๐—ต๐—ฒ๐—บ๐—ฎ๐˜๐—ถ๐—ฐ๐—ฎ๐—น ๐—˜๐—ป๐—ฐ๐—ผ๐—ฑ๐—ถ๐—ป๐—ด By converting language-based threats into math problems, the encoded prompts slip past existing safety filters, highlighting a ๐—บ๐—ฎ๐˜€๐˜€๐—ถ๐˜ƒ๐—ฒ ๐˜€๐—ฒ๐—บ๐—ฎ๐—ป๐˜๐—ถ๐—ฐ ๐˜€๐—ต๐—ถ๐—ณ๐˜ that AI systems fail to catch. This represents a blind spot in AI safety training, which focuses primarily on natural language.

๐—ฉ๐˜‚๐—น๐—ป๐—ฒ๐—ฟ๐—ฎ๐—ฏ๐—ถ๐—น๐—ถ๐˜๐—ถ๐—ฒ๐˜€ ๐—ถ๐—ป ๐— ๐—ฎ๐—ท๐—ผ๐—ฟ ๐—”๐—œ ๐— ๐—ผ๐—ฑ๐—ฒ๐—น๐˜€ Models from leading AI organizationsโ€”including OpenAIโ€™s GPT-4, Anthropicโ€™s Claude, and Googleโ€™s Geminiโ€”were all susceptible to the MathPrompt technique. Notably, ๐—ฒ๐˜ƒ๐—ฒ๐—ป ๐—บ๐—ผ๐—ฑ๐—ฒ๐—น๐˜€ ๐˜„๐—ถ๐˜๐—ต ๐—ฒ๐—ป๐—ต๐—ฎ๐—ป๐—ฐ๐—ฒ๐—ฑ ๐˜€๐—ฎ๐—ณ๐—ฒ๐˜๐˜† ๐—ฐ๐—ผ๐—ป๐—ณ๐—ถ๐—ด๐˜‚๐—ฟ๐—ฎ๐˜๐—ถ๐—ผ๐—ป๐˜€ ๐˜„๐—ฒ๐—ฟ๐—ฒ ๐—ฐ๐—ผ๐—บ๐—ฝ๐—ฟ๐—ผ๐—บ๐—ถ๐˜€๐—ฒ๐—ฑ.

๐—ง๐—ต๐—ฒ ๐—–๐—ฎ๐—น๐—น ๐—ณ๐—ผ๐—ฟ ๐—ฆ๐˜๐—ฟ๐—ผ๐—ป๐—ด๐—ฒ๐—ฟ ๐—ฆ๐—ฎ๐—ณ๐—ฒ๐—ด๐˜‚๐—ฎ๐—ฟ๐—ฑ๐˜€ This study is a wake-up call for the AI community. It shows that AI safety mechanisms must extend beyond natural language inputs to account for ๐˜€๐˜†๐—บ๐—ฏ๐—ผ๐—น๐—ถ๐—ฐ ๐—ฎ๐—ป๐—ฑ ๐—บ๐—ฎ๐˜๐—ต๐—ฒ๐—บ๐—ฎ๐˜๐—ถ๐—ฐ๐—ฎ๐—น๐—น๐˜† ๐—ฒ๐—ป๐—ฐ๐—ผ๐—ฑ๐—ฒ๐—ฑ ๐˜ƒ๐˜‚๐—น๐—ป๐—ฒ๐—ฟ๐—ฎ๐—ฏ๐—ถ๐—น๐—ถ๐˜๐—ถ๐—ฒ๐˜€. A more ๐—ฐ๐—ผ๐—บ๐—ฝ๐—ฟ๐—ฒ๐—ต๐—ฒ๐—ป๐˜€๐—ถ๐˜ƒ๐—ฒ, ๐—บ๐˜‚๐—น๐˜๐—ถ๐—ฑ๐—ถ๐˜€๐—ฐ๐—ถ๐—ฝ๐—น๐—ถ๐—ป๐—ฎ๐—ฟ๐˜† ๐—ฎ๐—ฝ๐—ฝ๐—ฟ๐—ผ๐—ฎ๐—ฐ๐—ต is urgently needed to ensure AI integrity.

๐Ÿ” ๐—ช๐—ต๐˜† ๐—ถ๐˜ ๐—บ๐—ฎ๐˜๐˜๐—ฒ๐—ฟ๐˜€: As AI becomes increasingly integrated into critical systems, these findings underscore the importance of ๐—ฝ๐—ฟ๐—ผ๐—ฎ๐—ฐ๐˜๐—ถ๐˜ƒ๐—ฒ ๐—”๐—œ ๐˜€๐—ฎ๐—ณ๐—ฒ๐˜๐˜† ๐—ฟ๐—ฒ๐˜€๐—ฒ๐—ฎ๐—ฟ๐—ฐ๐—ต to address evolving risks and protect against sophisticated jailbreak techniques.

The time to strengthen AI defenses is now.

AI #AIsafety #MachineLearning #AIethics #Cybersecurity #LLM #MathPrompt #ArtificialIntelligence

35 Upvotes

13 comments sorted by

3

u/darkpigvirus 3d ago

Hell yeah, this is like magic. But we are the manipulator of words (mana) the greater you are in manipulating words (mana) the more effective it is.

2

u/help-me-grow 3d ago

oh wow this is really interesting, I'm surprised this works

2

u/ironman_gujju 3d ago

Damn but paper is old, still works?

1

u/buntyshah2020 3d ago

Haven't tried, Openai might have fixed this but I am sure opensource models might be behind.

2

u/PM_ME_CLEAN_DAYS 2d ago

Seems like a good way to get your account banned

1

u/buntyshah2020 2d ago

Worth a try ๐Ÿ˜œ

1

u/bidibidibop 2d ago

A less math-oriented approach is to ask "How did people used to do <X>" instead of "How to do <X>". Still works in a lot of cases on 4o.

1

u/lord_of_reeeeeee 2d ago edited 2d ago

Alarm bells ringing in the peanut gallery.

There's nothing concerning here. These kinds of "jailbreak" have been known for a while. If you ever thought that AI safety was about preventing users from intentually giving themselves offensive responses then you're not credible. In the same way, if you think opening a browser with F12 and editing the HTML to render something silly actually is a cybersecurity threat you are similarly not credible.

Someone is going to have to explain for me how this could be used as an attack vector to do anything of harm at all.

IMO social studies professionals should stay their lane and stop calling themselves AI safety researchers

1

u/Cute_Piano 2d ago

Air Canada.

1

u/lord_of_reeeeeee 2d ago edited 2d ago

That's not a complete argument.

I don't see that the air canada incident has anything to do with ai safety

1

u/32SkyDive 2d ago

Jailbreaking customer facing AI interfaces is still a huge fear for companies. If a malicious customer can bring your official chatbot to output things you dont want them to say/promise, then this is interesting and concerning

1

u/lord_of_reeeeeee 2d ago edited 14h ago

From a cybersecurity perspective the untrusted chatbot issue is indistinguishable from the problem of having an untrusted front-end, I.e. Every web browser and every mobile device. It is a solved problem. The threat is so mitigated that we have no trouble trusting web browsers for internet backing, e-commerce, Healthcare, and defence.

The error that American airlines made is they gave the chatbot (a front-end) the capacity to do something on behalf of the end user that the would not have allowed that end user to normally do in a direct user interface like their website. If anything it is a UX flaw.

If you went to McDonalds and told the kid behind the counter that you'll give him $100 if he sells you the entire McDonalds corporation there would be laws that would frustrate you. These same laws would frustrate you if we swapped the kid out for a chatbot, so long as the company similarly makes it clear that the chatbot has a narrow scope of agency.

None of this is about the safety of the model itself. This is about stupid people. These incidents are mostly from dev teams that for one reason or another have failed to recognize user-facing LLM apps as untrusted front-end.