Bounded worker execution
Delegate narrow coding tasks to local or cheap workers while the supervisor stays focused on planning and review.
OSS agent execution harness
Use one strong supervisor model as the architect, then route bounded coding work to cheap local workers with deterministic verification, policy guardrails, and operator control.
Standalone OSS, optimized for solo power users running local or mixed-model execution.
harness init --profile local harness doctor harness demo --list-tasks harness demo --dry-run --task ts-isUUID # then run one bounded task harness run tasks/ts-lib/ts-isUUID.json \ --repo test-repos/ts-lib
What it does
Agent Harness is the execution layer between a high-skill supervisor and a pool of cheaper workers. The point is not raw autonomy. The point is repeatable bounded execution with evidence, routing, verification, and recovery you can actually trust.
Delegate narrow coding tasks to local or cheap workers while the supervisor stays focused on planning and review.
Check expected text, file scope, policy, and test commands before a task is allowed to pass.
Generate current-schema configs with local, distributed, or semantic profiles instead of hand-assembling JSON from source code.
Doctor checks, reports, approvals, audit logs, and structured artifacts make failures and recoveries explainable.
Mix Ollama, LM Studio, Anthropic, OpenAI, and OpenAI-compatible endpoints behind one model registry and router.
Run a local single-user harness today, then move to controller plus workers without changing the product story.
How it works
Start with npm, Docker Compose, or local-source dev mode. Those are the canonical supported paths.
Use `harness init --profile ...` to create a current config, then `harness doctor` to verify paths, git, and model connectivity.
Point the harness at a repo and one or more task definitions. A supervisor can route, verify, escalate, or require approval.
Use reports, audit logs, approvals, and demo flows to understand outcomes and recover from the common failure cases.
First-run shape
Start with `harness init`, `harness doctor`, and `harness demo`. Expand to distributed workers, semantic context, and approvals only when your workflow actually needs them.
Positioning
| Capability | Agent Harness | Aider | OpenHands | Cursor |
|---|---|---|---|---|
| Standalone OSS first | Yes | No | No | No |
| Local-worker orchestration | Yes | No | Yes | No |
| Structured approvals and policy | Yes | No | No | No |
| Profile-driven init and doctor | Yes | No | No | No |
| Distributed controller and workers | Yes | No | Yes | No |
| Soft-stable reports and CLI surface | Yes | No | No | No |
Current product path
The roadmap prioritizes reliable weekly usage, benchmarkable quality, and operator trust before any hosted-control-plane story.
OSS Local
Best-supported path today. Single-user standalone operation with local or mixed-model execution.
Power User
Phase 12 and 13 work: release hardening, evidence-rich reliability, and first-run trust surfaces.
Team-Ready Later
Shared-controller and multi-operator features are planned, but not allowed to distort the standalone product path.
FAQ
Yes. The current product direction is standalone OSS first, with npm, Docker Compose, and source installs as the canonical supported paths.
It gives supervisors a reliable way to delegate bounded coding work to cheaper workers while keeping verification, policy, routing, and recovery under operator control.
No, but that is the primary wedge. Agent Harness also supports hosted and OpenAI-compatible providers through the shared adapter layer.
Generate a config with `harness init --profile local`, run `harness doctor`, then inspect or run the built-in `harness demo` scenario.
Not in this cycle. Team-ready operation is planned later, but the current roadmap explicitly keeps standalone local use as the primary product path.
Use the install guide, quickstart, and examples docs. They map directly to the current first-run CLI and tested example configs.
Start with the tested quickstart flow, inspect one demo task, and validate the example configs before you wire it into a bigger agent stack.