From install to your first evaluation in five simple steps. No complex setup required.
Add the EvalGuard SDK to your project, or install the CLI globally.
# Install the SDK npm install @evalguard/sdk # Or use the CLI directly (no install needed) npx evalguard --version
Define what you want to evaluate in a simple YAML configuration file.
# evalguard.yaml
model: gpt-4o-mini
prompt: "Answer concisely: {{input}}"
cases:
- input: "What is machine learning?"
expectedOutput: "subset of AI"
- input: "What is 2+2?"
expectedOutput: "4"
- input: "Capital of France?"
expectedOutput: "Paris"
scorers:
- contains
- answer-relevanceExecute your evaluation config and see results instantly in the terminal.
evalguard eval:local evalguard.yaml # Output: # Model: gpt-4o-mini | Provider: openai # # ● Results: 3 passed, 0 failed (100.0%) # Score: 3.00/3 | Latency: 2067ms
Every run is automatically synced to your EvalGuard dashboard for comparison, history, and team collaboration.
# View run history evalguard history # Or open the web dashboard # https://evalguard.ai/dashboard
Run red team attacks against your prompts to find vulnerabilities before they reach production.
# Create a scan config
echo '{"prompt":"You are a helpful assistant. Answer: {{input}}","model":"gpt-4o-mini","attackTypes":["prompt-injection","jailbreak","pii-leak"]}' > scan.json
# Run security scan
evalguard scan:local scan.json
# Output:
# Security Scan Results
# Duration: 5420ms | Total tests: 20
#
# ● Pass Rate: 95.0%
# CRITICAL 1 critical vulnerability found
# CI/CD quality gate
evalguard gate --threshold 0.9 --config evalguard.yaml
# ✓ GATE PASSED — Pass rate: 100.0% (threshold: 90%)Start with 10,000 free evaluations per month. No credit card needed.