The Week of the Moltbots - OpenClaw and autonomous agents

The Week of the Moltbots

I wrote this article for Ippon on February 10, 2026. The enthusiasm around projects like OpenClaw—an open-source framework enabling autonomous task execution across messaging platforms, file systems, and enterprise APIs—reveals a critical blind spot in enterprise technology governance. For an entire week, I heard nothing but discussions of moltbots, OpenClaw, Clawd, and various implementations being created with these tools. But when I realized the permissions being exposed and how enterprises were approaching agent governance, I felt compelled to document this moment in AI history. ...

February 10, 2026 · 4 min · Luke Little
Reference Architecture for Safe Generative AI on AWS for Regulated Environments

From Prompt to Production: Designing Safe Generative AI on AWS for Regulated Environments

The Real Problem: Production, Not Prototypes Everyone can demo generative AI. Almost no one can run it safely in production. Enterprises in finance, healthcare, and the public sector aren’t blocked by technology capabilities—they’re blocked by governance requirements that today’s AI implementations rarely satisfy. These organizations face three critical blockers: Data leakage risk: Sensitive information, from PII to trade secrets, flowing through public model APIs Lack of auditability: No reliable record of prompts, responses, or who accessed what information Unclear ownership: Ambiguous rights over prompt engineering IP, training data, and generated outputs AWS customers don’t want AI that behaves like a chatbot toy. They need AI that behaves like enterprise infrastructure: secured, monitored, audited, governed, and compliant with their existing security posture. ...

February 1, 2026 · 5 min · Luke Little