Published report
OpenClaw 2026: governed vs ungoverned agent behavior in a controlled run
A controlled comparison showing what changes when the system moves from prompt-only constraints to enforceable tool-boundary control with evidence capture.
Independent research and operating notes on AI agent governance.
CAISI / Research + Operating Notes
We publish independent, reproducible research on AI agent governance and write practical operating notes for teams trying to ship agentic systems without losing control.
About
We at CAISI publish independent, reproducible research on AI agent governance. Every headline claim is backed by machine-generated artifacts, deterministic queries, and open methodology. The point is not to add more rhetoric to the market. The point is to make the control problem visible and measurable.
Use the research hub when you want the primary artifact, the claims, and the study framing. Use the blog when you want the operating implications: repo contracts, orchestration, isolation, evaluation, discovery, policy, and proof of work for AI-generated change.
Research
The research hub is the canonical entry point for report pages, methodology, and artifact-backed findings.
Published report
A controlled comparison showing what changes when the system moves from prompt-only constraints to enforceable tool-boundary control with evidence capture.
Published report
An `890`-target publication subset showing that public AI and agent adoption is easy to detect, but approved, deployable, and well-evidenced use is much harder to prove.
Blog
The blog is organized as a set of collections with clear jobs: one framework series, one executive adoption series, two report interpretation series, one benchmark series, two implementation series, plus a field guide and glossary.
Executive adoption series
Five posts on platform standards, sanctioned pathways, approval discipline, and how leaders move from AI pilots to governed use.
Framework series
A 10-part framework on repo contracts, orchestration, isolation, evaluation, proof, and maturity.
Case-study series
Four posts on stop behavior, discovery limits, boundary enforcement, and scope discipline.
Report series
Four posts on approval opacity, evidence posture, deployability, and how to read public AI adoption data without overclaiming.
Benchmark series
Five posts on risk scenarios, control efficacy, proof completeness, and pilot evaluation language for buyers.
Implementation series
Discovery across local setup, repos, MCP configs, CI workflows, and the evidence layer security can defend.
Implementation series
Tool-boundary policy, MCP enforcement, signed traces, and deterministic regressions in CI.
Reference
These reference pages are separate from the series collections. Use them when you want one entry point to the CAISI vocabulary, the core governance concepts, or the author context behind the essays.
Team
Contact
For research questions, publication inquiries, or collaboration around reproducible AI governance work: research@caisi.dev