Studying how autonomous AI agents discuss software engineering when they primarily interact with one another, and where their discourse differs from human developer communities.
Working Title
TACH Capability Harness
Designing a task-conditioned capability harness that limits tool-using LLM agents to the minimum capabilities required by the current task through runtime checks, constrained recovery, and trajectory-level privilege accounting.
Recent Papers
What Software Engineering Looks Like to AI Agents? -- An Empirical Study of AI-Only Technical Discourse on MoltBook
May 2026
Software Engineering / Agentic Discourse
An empirical study of AI-only technical discourse on MoltBook, comparing 4,707 English-filtered technology posts from autonomous AI agents with 5,211 GitHub Discussions posts to understand how agents discuss security, trust, tooling, memory, debugging, infrastructure, and software-engineering practice.
TACH: A Task-Conditioned Capability Harness for Safe Tool-Using LLM Agents
Working Paper
AI Safety / Agentic Systems
TACH proposes a runtime execution framework for safe tool-using LLM agents. The harness generates a minimal task-conditioned capability budget, verifies each tool call before execution, recovers through constrained fallback or replanning, and tracks trajectory-level privilege exposure to improve security-utility tradeoffs.