Test and compare LLM prompts with one command. Measure response time, token usage, and cost.
export OPENAI_API_KEY=sk-...
pip install requests pyyaml
# Test a single prompt
python promptlab.py "Summarize this text: {{input}}" --var input="The quick brown fox..."
# Compare multiple prompts from a YAML file
python promptlab.py templates/summarization.yaml --var input="Your long text here"- Template variables — Use
{{variable}}placeholders in prompts - Side-by-side comparison — Test up to 3 prompt variations in one run
- Metrics — Response time, token count, estimated cost per call
- YAML templates — Define and reuse prompt collections
- Cost estimation — Per-model pricing for accurate cost tracking
summarization.yaml— 3 summarization strategiesextraction.yaml— 3 data extraction approachesclassification.yaml— 3 classification methodscode_review.yaml— 3 code review stylesrewriting.yaml— 3 rewriting techniques
PromptLab Pro ($24) adds:
- Multi-model comparison — OpenAI, Anthropic, Google Gemini, Ollama (local)
- Batch testing — Test prompts against CSV datasets
- Auto-scoring — LLM judge rates accuracy, completeness, clarity
- A/B test significance — Welch's t-test for statistical confidence
- Cost optimization — Recommendations for cheaper models
- Prompt chains — Multi-step prompt pipelines
- HTML reports — Beautiful dark-theme reports with visualizations
- Unlimited variations — No cap on prompt comparisons
- CSV Cleaner — Fix messy CSV files in one command
- JSONKit — Swiss Army knife for JSON: format, validate, query, diff, flatten, convert
- Polymarket Scanner — Scan prediction markets for mispricings
- All Tools — Full product catalog
MIT — free for personal and commercial use.