<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Incident Response on Stack Research</title><link>https://stackresearch.org/categories/incident-response/</link><description>Recent content in Incident Response on Stack Research</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Sun, 08 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://stackresearch.org/categories/incident-response/index.xml" rel="self" type="application/rss+xml"/><item><title>Build for the Hour After Failure</title><link>https://stackresearch.org/editorial/build-for-the-hour-after-failure/</link><pubDate>Sun, 08 Mar 2026 00:00:00 +0000</pubDate><guid>https://stackresearch.org/editorial/build-for-the-hour-after-failure/</guid><description>&lt;p&gt;At 4 a.m., the model is rarely the whole problem. The missing recovery path is.&lt;/p&gt;
&lt;p&gt;Agent systems are often designed around the moment before action: the prompt, the tool schema, the evaluator, the approval check, the confidence score. Those pieces matter. They shape whether the system should act at all. But the harder question arrives after a bad action has already crossed the boundary into production.&lt;/p&gt;
&lt;p&gt;What stops next? What is still allowed to run? Which identity was used? Which records changed? Which downstream systems trusted the result? Which part can be reversed, and which part can only be compensated for?&lt;/p&gt;</description></item></channel></rss>