Forgot your password?
typodupeerror
AI Privacy Security

Are AI Agents Compromised By Design? 38

Longtime Slashdot reader Gadi Evron writes: Bruce Schneier and Barath Raghavan say agentic AI is already broken at the core. In their IEEE Security & Privacy essay, they argue that AI agents run on untrusted data, use unverified tools, and make decisions in hostile environments. Every part of the OODA loop (observe, orient, decide, act) is open to attack. Prompt injection, data poisoning, and tool misuse corrupt the system from the inside. The model's strength, treating all input as equal, also makes it exploitable. They call this the AI security trilemma: fast, smart, or secure. Pick two. Integrity isn't a feature you bolt on later. It has to be built in from the start. "Computer security has evolved over the decades," the authors wrote. "We addressed availability despite failures through replication and decentralization. We addressed confidentiality despite breaches using authenticated encryption. Now we need to address integrity despite corruption."

"Trustworthy AI agents require integrity because we can't build reliable systems on unreliable foundations. The question isn't whether we can add integrity to AI but whether the architecture permits integrity at all."
This discussion has been archived. No new comments can be posted.

Are AI Agents Compromised By Design?

Comments Filter:
  • Next question.

  • Somebody predicting an AI product category is FUBAR instead of wonderful magic productivity?

  • Then you are already behind the curve and should just liquidate your company now rather than continue operating.

  • Why does this even need to be stated? These things are grossly insecure and that cannot be fixed. It does not get more "broken by design" that that.

  • "we can't build reliable systems on unreliable foundations" ... what? xD
    Half the purpose of the entire practice of engineering is exactly that. Making a reliable thing that you need, from unreliable things that you have.
    I am uncertain whether these people are engineers. Neither in the broader, nor narrower, sense.
    I'm not even getting into the matter of whether LLM-based tools are worth it or not.

    • We will probably end up with AI security agents looking over the shoulder of the AI agent to identify nefarious behaviour and intervene by stopping the process.
    • by stooo ( 2202012 )

      Adding more broken AI ove5r a broken AI is like when you shot yourself in the left foot, find out you can't stand up,
        then you shoot again in the right foot, so you are not in imbalance.

      And that's exactly what AI would do.

    • Engineering is about building FROM, not ON, unreliable things. Sand is unreliable to build a house from, but when mixed into concrete it becomes reliable. Building a house on sand isn't reliable, and no amount of engineering will fix that other than by replacing the sand.

    • Half the purpose of the entire practice of engineering is exactly that. Making a reliable thing that you need, from unreliable things that you have.

      TCP/IP being the obvious example.

  • by Mirnotoriety ( 10462951 ) on Wednesday October 15, 2025 @12:54AM (#65725802)
    > Bruce Schneier and Barath Raghavan say .. AI agents run on untrusted data, use unverified tools, and make decisions in hostile environments.

    That's the most authentic description of AI I've ever seen.
  • by ctilsie242 ( 4841247 ) on Wednesday October 15, 2025 @01:36AM (#65725828)

    I wonder if one can state the same about a web browser. Pulling untested stuff, on an unknown platform, displaying to an unknown user, etc.

  • I dislike shopping, part of the dislike is suspicion I am not getting the best deal, and it feels like a waste of time browsing. I am more interested in utility and durability.

    What if advertised prices were guide prices and the actual price was negotiated privately between agents. Purchase agent has a brief of what to look for. Sales agent has oversight of targets for the day, stock and order levels and all active prospects. A sale at 5% profit rather than 10% is worth more than no sale, or a sale to a comp

  • Schneier and Raghavan argue that the same feedback loops that make human agents powerful—observe, orient, decide, act—also make agentic AI vulnerable when fed poisoned or adversarial data. In their words, “The adversary isn’t inside the loop by accident; it’s there by architecture.” Their point is well-taken: modern LLM agents have no privilege separation between data and control. That’s a security nightmare, and their proposed “integrity-first” rethink

  • If you give your agent untrusted data and unlimited access, then you have the same problem as every software that uses untrusted data to do something with unlimited access. If you don't let your users inject SQL, then you also won't use your user's text unfiltered as input for a LLM with database access, will you?

    You should filter the input, you should filter the LLM tool calls, you should filter the tool input. The same for the output. There are so many places where you can put the access control, so choos

    • by Tatsh ( 893946 )

      The problem is "we" want to be able to turn a blind eye for AI to do huge amounts of work or there's no point. But since LLMs are nondeterministic pretty much their design, there's just no way we can ever be sure that the output will be good without continuously checking it.

      • by allo ( 1728082 )

        It's not nondeterministic (if you don't choose nondeterministic sampling) but you don't know the outcome before. In the end you don't with humans either, but most humans are more reliable, or at least the humans who would get a root shell on my PC.

        People just giving the AI full access and a simple prompt are giving up the control themselves. I like the word "centaur" model. Let the AI do (only) the legwork while you tell it the way.

VMS is like a nightmare about RXS-11M.

Working...