<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Spoiledlunch</title><link>https://26ebd17c.spoiledlunch.pages.dev/</link><description>Nerdy Stuff. Tech Talk. Zero Freshness. Analysis and commentary on GRC, security, and AI.</description><generator>Hugo 0.160.1</generator><language>en-us</language><lastBuildDate>Fri, 24 Apr 2026 08:30:00 -0400</lastBuildDate><atom:link href="https://26ebd17c.spoiledlunch.pages.dev/tags/safety/" rel="self" type="application/rss+xml"/><item><title>AI Governance Gets Real Only After Deployment</title><link>https://26ebd17c.spoiledlunch.pages.dev/articles/2026-04-24-ai-governance-gets-real-only-after-deployment/</link><pubDate>Fri, 24 Apr 2026 08:30:00 -0400</pubDate><guid>https://26ebd17c.spoiledlunch.pages.dev/articles/2026-04-24-ai-governance-gets-real-only-after-deployment/</guid><description>
&lt;![CDATA[<p><strong>Article</strong> • April 24, 2026 • 2 min read</p><p><strong>Topics:</strong> AI</p><p>The industry still talks about AI governance like the hardest part is agreeing on principles before launch. Recent work from NIST and OpenAI points to a different reality: the difficult part starts …</p><p><a href="https://26ebd17c.spoiledlunch.pages.dev/articles/2026-04-24-ai-governance-gets-real-only-after-deployment/">Read full analysis →</a></p>
]]></description><author>@spoiledlunch</author><category>AI</category><category>ai governance</category><category>monitoring</category><category>nist</category><category>safety</category></item></channel></rss>