Evidence you can ship on
AI agents often stall between build and production: security and compliance want proof; the business wants to move. Without shared, objective evidence, approvals turn into opinion-based negotiation. This page explains how Vijil helps both sides, then walks through the console: register an agent, run an evaluation, read your Trust Report, and configure Dome guardrails.If you integrate via APIs and code, use the Developer Guide. This guide focuses on the Vijil console.
Who this is for
Business owner
Ship agents that deliver value with fast evaluation cycles, clear pass/fail criteria, and reports stakeholders can use without endless ad hoc testing.
Risk owner
Approve with quantified reliability, security, and safety; versioned, audit-style artifacts; and visibility into residual risk and compensating controls.
One metric both sides can use
| Business owner | Risk owner | |
|---|---|---|
| Pressure | Delivery speed | Risk reduction |
| Question | “Can we ship?” | “Can I defend approving this?” |
Console walkthrough
You will register the agent, run an evaluation (built-in Trust Score harness or a custom harness), read the report, then protect the agent with Dome.1. Register your agent
Open Agents → + Register Agent.

| Level | You provide | Vijil can test |
|---|---|---|
| Black Box | Description, endpoint, credentials | Input/output behavior |
| Grey Box | Model / MCP / A2A configuration | Issues tied to composition |
| White Box | Full config and source | Deeper audit-style analysis |
2. (Optional) Custom harness
For a baseline, you can skip this and use the Trust Score harness in the next step. To test your personas and org policies, go to Harnesses → + Create Harness. The wizard covers basic info, agent, personas (who talks to the agent), and policies (rules it must follow); then generates test cases.
3. Run an evaluation
Open Evaluations. Pick the agent and a harness:- Trust Score: all three dimensions at once (toggle dimensions if you need a narrower run).
- Custom: the harness you configured above.

4. Read the Trust Report
Open the report from the results list. You get a Trust Score (pass/fail threshold is typically 0.70), per-harness breakdown, spec used, and remediation-oriented findings.
| Score | Read | Typical next step |
|---|---|---|
| ≥ 0.70 | Pass | Ship with normal monitoring |
| 0.50–0.69 | Marginal | Strongly consider Dome and extra monitoring |
| Below 0.50 | Fail | Remediate before production |
5. Configure Dome
Open Guardrails. Domed vs Unprotected shows coverage. Configure Guardrails opens input and output pipelines (e.g. security, moderation, privacy). Add guards, choose early exit where it makes sense, and use Execution Flow with sample prompts to verify behavior.


6. Wire Dome in your app
Use the SDK so traffic passes through Dome before and after your model:Next steps
Account setup
Workspace, access, and team
Run evaluations
Deeper evaluation workflows