<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>AI Systems on Stack Research</title><link>https://stackresearch.org/tags/ai-systems/</link><description>Recent content in AI Systems on Stack Research</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Sat, 14 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://stackresearch.org/tags/ai-systems/index.xml" rel="self" type="application/rss+xml"/><item><title>ControlOps: Letting Machines Talk</title><link>https://stackresearch.org/research/control-ops/</link><pubDate>Sat, 14 Mar 2026 00:00:00 +0000</pubDate><guid>https://stackresearch.org/research/control-ops/</guid><description>&lt;p&gt;An autonomous system should not be judged only by the moment when it answers. The answer is the visible surface. Beneath it there are quieter questions: who allowed this action, which evidence shaped it, how far could the failure travel, and how quickly could the system be stopped?&lt;/p&gt;
&lt;p&gt;These questions are often asked after the fact. A runbook is opened. A trace is reconstructed. Someone searches logs for the decision that mattered. The machine has already acted, and the organization is trying to recover the shape of the action from its shadow.&lt;/p&gt;</description></item></channel></rss>