Skip to content

Execution & Monitoring

One of Agent HQ’s most powerful features is the ability to watch AI agents work in real time. When a task is processing, you get full visibility into what’s happening.

Click any task card on the board to open the Task Detail modal. For running tasks, this shows live execution data.

Logs stream in real time as the task executes. Each log entry has a type:

TypeDescription
systemSystem messages (setup, teardown)
textStandard output
toolTool invocations (file writes, commands)
errorErrors and warnings
outputFinal output and results

Logs are displayed in a terminal-style view with color coding.

As Pilot writes code, modified and created files appear in a file tree. Click any file to see its contents with syntax highlighting.

For files that Pilot annotated (inline comments or drawing annotations), those annotations appear alongside the code.

If the task runs tests, results are parsed and displayed in a structured format showing:

  • Number of tests passed / failed / skipped
  • Individual test names and their status
  • Error messages for failed tests

For tasks that produce a web application or page, a live preview is available in an iframe. This lets you see the visual result without leaving Agent HQ.

Tasks can produce markdown artifacts — structured output like documentation, reports, or plans. These render with full markdown formatting in the task detail view.

For complex tasks, Pilot may create an execution plan — a series of steps to accomplish the goal. The plan view shows each step with its status (pending, in progress, done, failed).

Each task runs in an isolated sandbox environment that includes:

  • Filesystem — an isolated copy of the project’s repo
  • Terminal — shell access for running commands
  • Browser — Puppeteer-based browser for screenshots and web automation

When a task finishes, it moves to the “Done” column. The task detail shows:

  • Total execution time
  • Token usage (input and output)
  • Cost in dollars
  • All generated files and changes

If a task fails, you can:

  1. Inspect the logs — understand what went wrong
  2. Modify the task — edit the title or body to provide better instructions
  3. Retry — click the retry button to re-run the task from scratch

You can capture browser screenshots of a task’s sandbox at any point. Screenshots are stored and viewable in the task detail, and can be annotated with inline comments or drawings.