<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Incident Response on Stack Research</title><link>https://stackresearch.org/tags/incident-response/</link><description>Recent content in Incident Response on Stack Research</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Fri, 17 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://stackresearch.org/tags/incident-response/index.xml" rel="self" type="application/rss+xml"/><item><title>Agent Incident Response Needs a Measurable Drill</title><link>https://stackresearch.org/research/agent-incident-drill/</link><pubDate>Fri, 17 Apr 2026 00:00:00 +0000</pubDate><guid>https://stackresearch.org/research/agent-incident-drill/</guid><description>&lt;p&gt;Agent incident response needs a clock, a journal, and a stopping point.&lt;/p&gt;
&lt;p&gt;Without those three things, failure remains theatrical. A bad action happens, someone opens logs, someone reconstructs intent, someone asks whether the system could have been stopped sooner. The answers arrive after the important interval has already passed.&lt;/p&gt;
&lt;p&gt;The useful question is narrower: can a controlled agent failure be made measurable while it is happening?&lt;/p&gt;
&lt;p&gt;&lt;a href="https://stackresearch.org/research/control-ops/"&gt;ControlOps&lt;/a&gt; built the parts: scope validation, decision lineage, blast-radius assessment, and kill-path auditing. The drill described here connects those parts around one small incident. It does not prove that agent systems are safe. It proves something more modest and more useful: one proposed action can be checked, stopped, recorded, scored, and prepared for rollback before it becomes an invisible state change.&lt;/p&gt;</description></item><item><title>Build for the Hour After Failure</title><link>https://stackresearch.org/editorial/build-for-the-hour-after-failure/</link><pubDate>Sun, 08 Mar 2026 00:00:00 +0000</pubDate><guid>https://stackresearch.org/editorial/build-for-the-hour-after-failure/</guid><description>&lt;p&gt;At 4 a.m., the model is rarely the whole problem. The missing recovery path is.&lt;/p&gt;
&lt;p&gt;Agent systems are often designed around the moment before action: the prompt, the tool schema, the evaluator, the approval check, the confidence score. Those pieces matter. They shape whether the system should act at all. But the harder question arrives after a bad action has already crossed the boundary into production.&lt;/p&gt;
&lt;p&gt;What stops next? What is still allowed to run? Which identity was used? Which records changed? Which downstream systems trusted the result? Which part can be reversed, and which part can only be compensated for?&lt;/p&gt;</description></item></channel></rss>