AI agents moved from demos to daily tools in the past year. OpenAI, Google, and Anthropic now ship agents that run multi-step tasks, and open-source projects like OpenClaw show how fast autonomy can spread—along with new security risks.
MIT CSAIL’s 2025 AI Agent Index tries to measure what’s actually happening in the wild. The index reviews 30 agentic systems using 45 public fields across safety, autonomy, ecosystem behavior, and evaluation, based on a snapshot as of Dec. 31, 2025.
What the Index Finds: Capability Outruns Transparency
The index says releases accelerate quickly: 24 of 30 agents shipped or received major agentic updates in 2024–2025, and browser agents often operate at very high autonomy.
But safety disclosure lags. Among 13 agents with frontier-level autonomy, only 4 disclose any agent-specific safety evaluations. Developers also publish almost no test results: 25/30 disclose no internal safety results, and 23/30 disclose no third-party safety testing information.
Browser Agents Act Like Humans and Websites Can’t Tell
The report flags “web conduct” as unresolved. Some browser agents ignore robots.txt, and some tools market themselves on bypassing anti-bot systems. The index notes that only one agent (ChatGPT Agent) uses cryptographic request signing, which would make verification easier for websites.
Disclosure also looks weak. The index coverage highlights how often agents fail to clearly identify themselves, which makes automated traffic blend into normal human browsing and complicates enforcement of site policies.
Why This Matters for Security, Compliance, and Responsibility
The index describes a layered ecosystem: many agent products sit on top of a small set of foundation models, plus scaffolding and orchestration layers. That structure fragments accountability, because no single party owns the full risk surface end-to-end.
This gap matters because autonomous agents face practical security threats like prompt injection and unsafe tool use. When teams publish broad ethics language but skip concrete evidence about day-to-day vulnerabilities, the index frames that as a form of “safety washing.”
Eco-friendly SEO Angle: Safer Agents Can Also Cut Digital Waste
Responsible agent design can reduce energy and waste. Agents that respect robots.txt, use clear identification, and avoid unnecessary scraping generate less redundant traffic and fewer repeated retries. That lowers compute load for both websites and data centers.
Open standards can help too. The Agentic AI Foundation (AAIF), launched under the Linux Foundation with OpenAI, Anthropic, and others, aims to support interoperable agent infrastructure and safer conventions as agents enter production.

