<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Accountability on Stack Research</title><link>https://stackresearch.org/tags/accountability/</link><description>Recent content in Accountability on Stack Research</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Tue, 24 Feb 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://stackresearch.org/tags/accountability/index.xml" rel="self" type="application/rss+xml"/><item><title>NHI and Agentic Risk: When Humans Use Machine Credentials</title><link>https://stackresearch.org/research/nhi-asi-series-05-human-use-of-nhi/</link><pubDate>Tue, 24 Feb 2026 00:00:00 +0000</pubDate><guid>https://stackresearch.org/research/nhi-asi-series-05-human-use-of-nhi/</guid><description>&lt;p&gt;The audit log says the machine acted. The real question is who meant for it to act.&lt;/p&gt;
&lt;p&gt;An engineer uses an automation token to run a one-off maintenance task. The token already has the right access. The work is urgent. The safer path takes longer. Later, an agent uses the same token to approve a sensitive action because the credential still works and the tool accepts it. When the action is questioned, the log shows the non-human identity. It does not show the human intent that first bent the identity out of shape.&lt;/p&gt;</description></item></channel></rss>