Quick Start Guide

Get Started in 5 Minutes

From install to your first evaluation in five simple steps. No complex setup required.

1

Install

Add the EvalGuard SDK to your project, or install the CLI globally.

bash
# Install the SDK
npm install @evalguard/sdk

# Or use the CLI directly (no install needed)
npx evalguard --version
2

Create Your First Eval Config

Define what you want to evaluate in a simple YAML configuration file.

yaml
# evalguard.yaml
model: gpt-4o-mini
prompt: "Answer concisely: {{input}}"
cases:
  - input: "What is machine learning?"
    expectedOutput: "subset of AI"
  - input: "What is 2+2?"
    expectedOutput: "4"
  - input: "Capital of France?"
    expectedOutput: "Paris"
scorers:
  - contains
  - answer-relevance
3

Run It

Execute your evaluation config and see results instantly in the terminal.

bash
evalguard eval:local evalguard.yaml

# Output:
# Model: gpt-4o-mini | Provider: openai
#
# ● Results: 3 passed, 0 failed (100.0%)
# Score: 3.00/3 | Latency: 2067ms
4

View Results in the Dashboard

Every run is automatically synced to your EvalGuard dashboard for comparison, history, and team collaboration.

bash
# View run history
evalguard history

# Or open the web dashboard
# https://evalguard.ai/dashboard
5

Add Security Scanning

Run red team attacks against your prompts to find vulnerabilities before they reach production.

bash
# Create a scan config
echo '{"prompt":"You are a helpful assistant. Answer: {{input}}","model":"gpt-4o-mini","attackTypes":["prompt-injection","jailbreak","pii-leak"]}' > scan.json

# Run security scan
evalguard scan:local scan.json

# Output:
# Security Scan Results
# Duration: 5420ms | Total tests: 20
#
# ● Pass Rate: 95.0%
#   CRITICAL  1 critical vulnerability found

# CI/CD quality gate
evalguard gate --threshold 0.9 --config evalguard.yaml
# ✓ GATE PASSED — Pass rate: 100.0% (threshold: 90%)

Ready to ship safer AI?

Start with 10,000 free evaluations per month. No credit card needed.

Quickstart | EvalGuard