llmperf

A CLI tool for LLM performance testing


Keywords
azure, benchmarking, deepseek, latency, llm, openai, qps, test, tps
License
MIT
Install
pip install llmperf==0.0.1

Documentation

LLM Perf

llmperf is a benchmarking tool for stress testing self-hosted LLMs, Azure OpenAI, and OpenAI APIs. It provides efficient concurrent request management, detailed performance analysis, and helps developers optimize inference latency and throughput. With customizable test parameters, it is suitable for various LLM workload evaluation scenarios.

1. Configure your env

Copy and set up the environment variables from .env.example.

cp .env.example .env

2. Launch your App

Copy and set up the environment variables.

./start.sh