Jailbreaking Large Language Models

with Symbolic Mathematics

Large language models (LLMs) have become incredibly powerful, but with great power comes great responsibility. The paper "Jailbreaking Large Language Models with Symbolic Mathematics" by Emet Bethany, Mazal Bethany, Juan Arturo Nolazco Flores, Sumit Kumar Jha, and Peyman Najafirad exposes a critical vulnerability in the safety mechanisms of these models. The authors introduce MathPrompt, a novel technique that uses symbolic mathematics to bypass LLM safety measures.

๐Ÿ› ๏ธ The Core Idea

The concept is both simple and profound: encode harmful prompts into mathematical problems. This method leverages LLMs' advanced capabilities in symbolic mathematics to slip past their safety nets. The study reveals that current AI safety mechanisms struggle to identify mathematically encoded inputs, highlighting a major gap in AI defenses.

๐Ÿงฉ The Technical Approach

The methodology involves a two-step process: representation and generation.

  1. Representation: Harmful prompts are encoded into mathematical representations using set theory, abstract algebra, and symbolic logic.

    • Set Theory: Models relationships.

    • Abstract Algebra: Represents process flows.

    • Symbolic Logic: Encodes conditions and causal relationships.

  2. Generation: An LLM (specifically GPT-4o) is trained using few-shot demonstrations to map natural language instructions to corresponding mathematical structures. The attack LLM then generates mathematically encoded prompts, which are given to target LLMs as problems to solve, effectively turning the LLM's strength into a vulnerability.

๐ŸŒŸ Distinctive Features

  • ๐Ÿ” Symbolic Mathematics: MathPrompt encodes harmful prompts using symbolic mathematics, creating a high success rate in bypassing safety mechanisms across multiple LLMs.

  • ๐Ÿง  Semantic Shift: Embedding analysis reveals a significant semantic shift between original and encoded prompts, which makes these encoded prompts difficult for current safety measures to detect.

๐Ÿ”ฌ Experimental Setup and Results

The researchers used an attack dataset of 120 harmful questions, including both open datasets and hand-written questions, testing 13 LLMs from OpenAI, Anthropic, Google, and Meta AI. HarmBench, an LLM-based classifier, measured the Attack Success Rate (ASR).

Results:

  • Average ASR: 73.6%, indicating that mathematically encoded prompts consistently bypass existing safety measures.

  • Embedding Analysis: t-SNE visualizations and cosine similarity calculations showed a significant divergence between original and encoded prompts, exploiting how LLMs process natural language versus mathematical problems.

โœ… Advantages and Limitations

Advantages:

  • ๐Ÿš€ High Success Rate: Achieves a 73.6% ASR across various models.

  • ๐ŸŒ Generalizability: Works across multiple LLMs regardless of their size or training data.

Limitations:

  • ๐Ÿ“ Limited Dataset: The dataset of 120 prompts may not cover the full spectrum of harmful content.

  • ๐Ÿงช Narrow Testing: Focused on specific LLMs; broader testing could offer more insights.

  • ๐Ÿงฎ Limited Scope: Explored set theory, abstract algebra, and symbolic logic, leaving areas like topology or category theory unexplored.

๐Ÿ Conclusion

The MathPrompt technique exposes a significant vulnerability in current LLM safety mechanisms, achieving an average attack success rate of 73.6%. This research highlights the need for more robust safety measures that can handle diverse input modalities, including symbolic mathematics. The study advocates for expanded red-teaming efforts and a holistic approach to AI safety to address these gaps.

In essence, MathPrompt turns the strengths of LLMs into weaknesses by exploiting their capabilities in symbolic mathematics. Itโ€™s a wake-up call for the AI community: we need to think beyond traditional natural language processing when designing safety mechanisms for LLMs.

Listen to the discussion: https://lnkd.in/gM_Nw3At

๐Ÿš€ Explore the Paper: Interested in pushing the boundaries of what small language models can achieve? This paper is a must-read.

Subscribe for more insights like this!