I've been using Claude Code for months and kept hitting the same issue: tasks would get "completed" but I'd have no visibility into what assumptions the agent made or what it quietly simplified when it got stuck.
ctlsurf is a notebook that connects to AI agents via MCP. The key feature: when an agent marks a task done, it must provide structured completion data:
- Summary of what was done
- Assumptions made (required, at least one)
- What was attempted but failed
- What was simplified or skipped
That last one is the important part. Agents often silently give up on parts of tasks, and this forces transparency.
https://app.ctlsurf.com
Interested in feedback, especially from teams using AI agents in production where accountability matters.
sarkarsh•1h ago