pramana docs
Crowdsourced drift detection for LLM APIs.
LLM providers update, fine-tune, and swap models behind stable identifiers. Same name, different behavior. Pramana measures that.
Install
pip install pramana-ai
30-second demo
$ export OPENAI_API_KEY=sk-...
$ pramana run --tier cheap --model gpt-5.2
Running cheap suite against gpt-5.2...
✓ 10/10 passed
Pass rate: 100.0%
$ pramana submit results.json
✓ Submitted 10 results
What it does
- Runs fixed prompts against provider APIs with deterministic parameters (
temperature=0.0,seed=42) - Hashes results (SHA-256 of model + prompt + output) for deduplication
- Submits to a public dataset where hash-based consistency tracking detects changes across users and time
Pages
| Page | Content |
|---|---|
| Quick Start | Install, configure API keys, run, submit |
| CLI Reference | All commands, options, and examples |
| Architecture | System design, modules, data flow |
| Providers | Provider comparison, adding new providers |
| Test Suites | JSONL format, assertions, tiers |
| Reproducibility | Provider matrix, recommendations |
| Contributing | Setup, PR guidelines, test cases |
Scope
Pramana targets API providers (OpenAI, Anthropic, Google) where silent updates occur. Local models (Ollama, vLLM) are out of scope — you control the weights, so there's no drift to detect.
pramana (प्रमाण) — Sanskrit for proof, evidence, valid knowledge.