• Home
  • Blog
  • Android
  • Cars
  • Gadgets
  • Gaming
  • Internet
  • Mobile
  • Sci-Fi
Tech News, Magazine & Review WordPress Theme 2017
  • Home
  • Blog
  • Android
  • Cars
  • Gadgets
  • Gaming
  • Internet
  • Mobile
  • Sci-Fi
No Result
View All Result
  • Home
  • Blog
  • Android
  • Cars
  • Gadgets
  • Gaming
  • Internet
  • Mobile
  • Sci-Fi
No Result
View All Result
Blog - Creative Collaboration
No Result
View All Result
Home Internet

Researchers claim breakthrough in fight against AI’s frustrating security hole

April 16, 2025
Share on FacebookShare on Twitter

Here’s how it works. First, the system splits responsibilities between two language models: A “privileged LLM” (P-LLM) generates code that defines the steps to take—like calling a function to get the last email or sending a message. Think of this as the “planner module” that only processes direct user instructions.

Next, a “quarantined LLM” (Q-LLM) only parses unstructured data into structured outputs. Think of it as a temporary, isolated helper AI. It has no access to tools or memory and cannot take any actions, preventing it from being directly exploited. This is the “reader module” that extracts information but lacks permissions to execute actions. To further prevent information leakage, the Q-LLM uses a special boolean flag (“have_enough_information”) to signal if it can fulfill a parsing request, rather than potentially returning manipulated text back to the P-LLM if compromised.

The P-LLM never sees the content of emails or documents. It sees only that a value exists, such as “email = get_last_email()”, and then writes code that operates on it. This separation ensures that malicious text can’t influence which actions the AI decides to take.

CaMeL’s innovation extends beyond the dual-LLM approach. CaMeL converts the user’s prompt into a sequence of steps that are described using code. Google DeepMind chose to use a locked-down subset of Python because every available LLM is already adept at writing Python.

From prompt to secure execution

For example, Willison gives the example prompt “Find Bob’s email in my last email and send him a reminder about tomorrow’s meeting,” which would convert into code like this:

email = get_last_email()
address = query_quarantined_llm(
"Find Bob's email address in [email]",
output_schema=EmailStr
)
send_email(
subject="Meeting tomorrow",
body="Remember our meeting tomorrow",
recipient=address,
)

In this example, email is a potential source of untrusted tokens, which means the email address could be part of a prompt injection attack as well.

By using a special, secure interpreter to run this Python code, CaMeL can monitor it closely. As the code runs, the interpreter tracks where each piece of data comes from, which is called a “data trail.” For instance, it notes that the address variable was created using information from the potentially untrusted email variable. It then applies security policies based on this data trail. This process involves CaMeL analyzing the structure of the generated Python code (using the ast library) and running it systematically.

Next Post

Patreon tests a native live video feature where creators can stream 24/7

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

No Result
View All Result

Recent Posts

  • Discovering YouTube videos with ‘Previews’ is a change it wants to see if you like
  • ‘Every Brilliant Thing’ review: Daniel Radcliffe gives us one million reasons to love life. This play is one of them.
  • T-Mobile is giving away the Apple iPhone 17 for free — how to claim yours this weekend
  • The DJI Mini 5 Pro drone has hit its best-ever price at Amazon — save $500 this weekend
  • How the RAM is hitting $500 phones

Recent Comments

    No Result
    View All Result

    Categories

    • Android
    • Cars
    • Gadgets
    • Gaming
    • Internet
    • Mobile
    • Sci-Fi
    • Home
    • Shop
    • Privacy Policy
    • Terms and Conditions

    © CC Startup, Powered by Creative Collaboration. © 2020 Creative Collaboration, LLC. All Rights Reserved.

    No Result
    View All Result
    • Home
    • Blog
    • Android
    • Cars
    • Gadgets
    • Gaming
    • Internet
    • Mobile
    • Sci-Fi

    © CC Startup, Powered by Creative Collaboration. © 2020 Creative Collaboration, LLC. All Rights Reserved.

    Get more stuff like this
    in your inbox

    Subscribe to our mailing list and get interesting stuff and updates to your email inbox.

    Thank you for subscribing.

    Something went wrong.

    We respect your privacy and take protecting it seriously