@llm-dev-ops/test-bench-cli
v0.2.0
Published
CLI wrapper for LLM Test Bench - A production-grade framework for testing and benchmarking Large Language Models
Maintainers
Readme
@llm-dev-ops/test-bench-cli
CLI wrapper for LLM Test Bench - A production-grade framework for testing and benchmarking Large Language Models.
Installation
# Install globally
npm install -g @llm-dev-ops/test-bench-cli
# Or use with npx (no installation required)
npx @llm-dev-ops/test-bench-cli --helpUsage
After installation, you can use the ltb command:
# Show help
ltb --help
# Show version
ltb --version
# Run a benchmark
ltb benchmark --provider openai --model gpt-4 --prompt "Explain quantum computing"
# Compare models
ltb compare --providers openai:gpt-4,anthropic:claude-opus-4 --prompt "Write a poem"
# Run evaluation
ltb evaluate --file responses.json --evaluator coherenceCommands
ltb benchmark- Run benchmarks on LLM modelsltb compare- Compare multiple modelsltb evaluate- Evaluate model responsesltb analyze- Analyze benchmark resultsltb visualize- Generate visualization dashboards
Prerequisites
This CLI requires the Rust-based LLM Test Bench binary to be installed:
# Install via Cargo
cargo install llm-test-bench
# Or download from releases
# https://github.com/LLM-Dev-Ops/test-bench/releasesSDK Package
For programmatic access in TypeScript/JavaScript, use the SDK package:
npm install @llm-dev-ops/test-benchimport { LLMTestBench } from '@llm-dev-ops/test-bench';
const bench = new LLMTestBench();
const results = await bench.benchmark({
provider: 'openai',
model: 'gpt-4',
prompts: ['Test prompt']
});Documentation
License
MIT
Repository
https://github.com/LLM-Dev-Ops/test-bench
