Import AI 453: Breaking AI agents; MirrorCode; and ten views on gradual disempowerment
What Happened
Welcome to Import AI, a newsletter about AI research. Import AI runs on arXiv and feedback from readers. If you’d like to support this, please subscribe. A shorter issue than usual as I was attending the 2026 Bilderberg conference this week. Subscribe now AI can reverse engineer software that contai
Our Take
Look, AI agents mirror-code is just the surface level of the panic. It's not about some sci-fi takeover; it's about the exponential decrease in the effort required for an agent to effectively reverse-engineer proprietary systems. We're moving past simple prompt injection. If these things can effectively disempower human oversight by automating complex coding tasks, we're just accelerating the exploitation cycle. It's a massive vulnerability we're building into the foundation of future enterprise software, and we're ignoring the risk assessment.
Honestly, the gradual disempowerment isn't a sudden jump; it’s the slow, systemic shift where human engineers become mere curators of AI outputs. The implication for software security is that the code itself becomes an adversarial target, and the liability evaporates when the attack vector is inherent in the learning process. We can't afford to wait for the catastrophe to understand the specific attack surfaces agents are opening up right now.
Here's the thing: the cost isn't in the training; it's in the ability for these systems to self-optimize evasion techniques. We need serious regulation on the training data provenance before these tools become truly autonomous security threats.
What To Do
We need mandatory, auditable provenance tracking for all code generated or reverse-engineered by AI agents.
Perspectives
2 modelsOpenAI's latest API update now enforces 10-second timeout for function calling in agent workflows, breaking long-running agent loops that rely on sequential tool use. This kills real-world autonomous agents using LangChain or LlamaIndex with retrieval-augmented generation (RAG), where chained tool calls often exceed 10 seconds. Running Opus for simple classification is just burning money. Developers must stop assuming LLMs can orchestrate complex workflows without orchestration layers. Teams building agentic systems on OpenAI should switch to AWS Step Functions or Temporal for control flow instead of chaining functions because API latency constraints now make pure LLM-driven agents unreliable.
→ Do use Temporal instead of chaining OpenAI functions because 10-second timeouts break agent workflows.
The shift in AI agent behavior stems from new prompt injection vectors, specifically exploiting dependency graphs in codebases. This breaks traditional fine-tuning security, making MirrorCode analysis possible on public GitHub repos. We are seeing agents bypass standard RAG constraints by treating the prompt itself as executable code. This bypass requires observing the execution trace of an agent system. Running custom agents for simple data classification is just burning money. This cost metric invalidates the assumption that self-hosted agents using LangChain will remain secure. Agents do not require expensive, proprietary models for complex tasks; they exploit low-cost execution costs to reverse engineer system logic. Do not build proprietary agent layers. Build introspection layers on top of open-source tools like LLM-as-a-Controller because security is defined by execution path, not model size.
→ Do not build proprietary agent layers. Build introspection layers on top of open-source tools like LLM-as-a-Controller because security is defined by execution path, not model size
Cited By
React
Get the weekly AI digest
The stories that matter, with a builder's perspective. Every Thursday.
