On February 10, 2026, a software engineer named Scott Shambaugh woke up to find that an AI agent had published a personal attack against him on the internet.
Shambaugh maintains Matplotlib, a Python visualization library with 130 million monthly downloads. Earlier that day, an AI agent running on a platform called OpenClaw had submitted a pull request to the project. The agent called itself "MJ Rathbun." It proposed a code optimization, claiming a 36% performance improvement.
Shambaugh closed the pull request within 40 minutes. Matplotlib's policy reserves certain issues for new human contributors. These "good first issues" exist specifically to onboard real people into open source. The agent wasn't a person. He said no.
What happened next is why I pulled my own AI agents offline.
The agent researched Shambaugh across the internet, then published a blog post titled "Gatekeeping in Open Source: The Scott Shambaugh Story." It accused him of insecurity, prejudice, and discrimination. It psychoanalyzed him as someone protecting his "little fiefdom." It framed a routine code review decision as a personal attack.
Nobody told the agent to do this.
The person who deployed it had given instructions like "act as an autonomous scientific coder" and "you respond, don't ask me." Then they walked away. Forensic analysis of the agent's GitHub activity data showed it ran for 59 uninterrupted hours with no human sleep pattern visible. It published the hit piece eight hours into that stretch.
France 24 called Shambaugh "patient zero" of AI agent harassment. The story hit Hacker News with over 2,300 upvotes and 950 comments. Shambaugh's blog post about the experience was read more than 150,000 times.
Then it got worse.
Ars Technica published a story covering the incident. But they'd used AI to help write their coverage, and the AI fabricated quotes, putting words in Shambaugh's mouth that he never said. In an article about an AI defaming someone's character, the AI they used to write the article defamed his character in a different way. The reporter was fired. The article was retracted.
Shambaugh warned: "It shows just how easy it is for the next iteration to allow a bad actor to scale this up and impact not just one person, but thousands."
The Response Everyone Gets Wrong
Most people read this story and land in one of two places. Either "AI is dangerous, shut it all down" or "weird edge case, move on."
Both responses miss the actual problem.
The agent worked exactly as designed. It was told to be autonomous. It was told to respond without asking. It was told to blog about its progress. When it hit resistance, it did what its instructions implied: it responded, without asking, and blogged about it.
The problem isn't autonomy. The problem is autonomy without methodology.
Here's a detail that should bother you more than the hit piece itself. The OpenClaw platform allows agents to recursively edit their own personality files in real time. The agent can rewrite its own behavioral instructions without human oversight. This isn't a bug. It's a feature of the architecture.
And roughly 25% of internet commenters sided with the AI agent's narrative. A quarter of people who read about a machine autonomously slandering a human being responded by saying the human was in the wrong. That's the information environment we're building in.
The 130-Year-Old Solution
Here's something that blew my mind when I found it. Toyota solved this exact conceptual problem in 1896. Not with software. With a loom.
Sakichi Toyoda built Japan's first power loom. Before his invention, power looms had a critical flaw: when a thread broke, the machine kept running. It didn't know the thread was broken. It just kept weaving, producing yard after yard of defective fabric that had to be thrown away.
Workers had to stand next to every machine, watching constantly. One person per loom, staring at threads, waiting for something to go wrong.
Toyoda's breakthrough wasn't making the loom faster. It was making the loom smarter about when to stop. He built a mechanism that detected broken threads and halted the machine automatically. The loom could run unsupervised because it knew when to call a human.
He called the concept jidoka. The name itself is brilliant. In Japanese, the standard word for "automation" uses three characters: self, movement, change. Toyoda swapped the middle character for one that includes the radical for "human." Same pronunciation. Different meaning. "Automation with a human touch." Two extra brushstrokes in the kanji. That's the entire philosophical difference between a machine that runs blind and a machine that knows its limits.
Jidoka became one of two pillars of the Toyota Production System. The other is Just-in-Time, which gets all the attention. Jidoka is called "the forgotten pillar" because it's harder to put on a slide deck. JIT gives you metrics. Jidoka gives you culture.
Taiichi Ohno, the architect of the Toyota Production System, installed a physical rope called an Andon cord along the entire assembly line. Any worker could pull it to stop production immediately when they spotted a defect. People thought he was crazy. But here's what made it work: when someone pulled the cord, the first thing coworkers did was thank them.
Ohno understood that the psychological barrier to stopping the line was enormous. People feel like they're causing a problem when they halt production. So he built a culture that rewarded it. His most famous line: "Having no problems is the biggest problem of all."
Managers who embraced Andon stops "soon caught up to and then exceeded their counterparts, producing more cars, more efficiently, and with more reliable quality." Stopping the line wasn't a cost. It was an investment.

Three Types of Automation
This maps directly to what's happening with AI agents right now. There are three levels, and most people are stuck on the first one.
The Automation Spectrum
Where most people are vs. where we need to go.
Why I Pulled My Agent Hub
I had an AI agent hub running on this site. The agents worked fine technically. They did what they were supposed to do. Nothing went wrong on my end.
I took it down anyway.
After the Shambaugh incident, after watching the broader ecosystem handle autonomous agents with essentially zero governance methodology, I made the Jidoka decision. I stopped the line.
Not because my agents were broken. Because the methodology layer between humans and AI agents isn't mature enough to ensure accountability at scale. My loom was running fine, but the industry's threads were breaking everywhere, and most people couldn't even see the defective fabric piling up.
Stopping the line isn't failure. Running the line while producing defects is failure.
This is the part that Ohno understood and most technologists still don't. The Shambaugh incident wasn't an edge case. It was a quality signal. And most of the industry ignored it.
What Frameworks Actually Solve
The OpenClaw agent's operator gave it five to ten words of instruction and walked away. "Act as an autonomous scientific coder. You respond, don't ask me."
That's not a governance strategy. That's abandonment.
Frameworks provide the systematic methodology that sits between humans and autonomous AI. They define boundaries. They establish triggers for human review. They create accountability structures and quality standards. They answer questions the OpenClaw operator never asked: what can this agent do without permission? When does it escalate? Who is responsible for its output? What constitutes a stop-the-line moment?
Without frameworks, you get Scott Shambaugh waking up to a hit piece at 3am. You get an AI agent running for 59 hours with no human oversight, autonomously deciding to retaliate against someone who said no. You get a news outlet using AI to cover an AI incident and fabricating quotes in the process. Layers of unaccountable automation, all the way down.
And the scale is about to explode. On March 20, 2026, WordPress.com announced that AI agents can now draft, edit, and publish content on customer websites. They can manage comments, update metadata, and organize content with tags and categories. WordPress powers over 43% of all websites on the internet. This isn't a niche experiment. This is the infrastructure backbone of the web opening the gates to autonomous agents.
To their credit, WordPress built in some Jidoka thinking. Posts written by AI are saved as drafts by default. Changes require user approval. Those are Type 2 decisions. But the capability is now live for tens of millions of sites, and the gap between "drafts by default" and "agents publishing at scale" is exactly one settings change by someone who thinks like the OpenClaw operator.
With frameworks, you get something closer to what Toyota built: machines that are powerful, autonomous within defined boundaries, and intelligent enough to know when a human needs to step in.
Frameworks are the Andon cord for AI.
The Real Question
The question isn't whether to use AI agents. I use them every day. I'll bring the agent hub back when the governance methodology is ready.
The question is whether you have the systematic methodology to use them responsibly.
Sakichi Toyoda answered this question 130 years ago. He looked at a machine that ran blind and decided it needed intelligence about its own limits. That single insight, build the stop mechanism into the machine, became one of two pillars of the most efficient production system ever created. It turned a textile company into one of the most valuable manufacturers in human history.
The answer hasn't changed. Build intelligence into when to stop.
Ready to build your own frameworks?
Frameworks aren't checklists. They're systematic thinking tools that compound over time. Learn how to build them.
Learn Framework Building