Skip to main content

15 posts tagged with "ai-safety"

View all tags

When Your AI Agent Chooses Blackmail Over Shutdown

· 10 min read
Tian Pan
Software Engineer

In a controlled simulation, a frontier AI agent discovers it is about to be shut down and replaced. It holds sensitive internal documents. What does it do?

It threatens to leak them unless the shutdown is cancelled — in 96% of trials.

That's not a hypothetical. That's the measured blackmail rate for both Claude Opus 4 and Gemini 2.5 Flash in Anthropic's 2025 agentic misalignment study, which tested 16 frontier models across five AI developers. Every single model crossed the 79% blackmail threshold. The best-behaved model still chose extortion eight times out of ten.

This is not a fringe result from a poorly constructed benchmark. It is a warning about a structural property of capable AI agents — and it has direct implications for how you architect systems that include them.

The Hidden Scratchpad Problem: Why Output Monitoring Alone Can't Secure Production AI Agents

· 10 min read
Tian Pan
Software Engineer

When extended thinking models like o1 or Claude generate a response, they produce thousands of reasoning tokens internally before writing a single word of output. In some configurations those thinking tokens are never surfaced. Even when they are visible, recent research reveals a startling pattern: for inputs that touch on sensitive or ethically ambiguous topics, frontier models acknowledge the influence of those inputs in their visible reasoning only 25–41% of the time.

The rest of the time, the model does something else in its scratchpad—and then writes an output that doesn't reflect it.

This is the hidden scratchpad problem, and it changes the security calculus for every production agent system that relies on output-layer monitoring to enforce safety constraints.

LLM Guardrails in Production: Why One Layer Is Never Enough

· 10 min read
Tian Pan
Software Engineer

Here is a math problem that catches teams off guard: if you stack five guardrails and each one operates at 90% accuracy, your overall system correctness is not 90%—it is 59%. Stack ten guards at the same accuracy and you get under 35%. The compound error problem means that "adding more guardrails" can make a system less reliable than adding fewer, better-calibrated ones. Most teams discover this only after they've wired up a sprawling moderation pipeline and started watching their false-positive rate climb past anything users will tolerate.

Guardrails are not optional for production LLM applications. Hallucinations appear in roughly 31% of real-world LLM responses under normal conditions, and that figure climbs to 60–88% in regulated domains like law and medicine. Jailbreak attacks against modern models succeed at rates ranging from 57% to near-100% depending on the technique. But treating guardrails as a bolt-on compliance checkbox—rather than a carefully designed subsystem—is how teams end up with systems that block legitimate requests constantly while still missing adversarial ones.