Files
vllm/benchmarks
2025-10-08 10:40:42 -07:00
..

Benchmarks

This directory used to contain vLLM's benchmark scripts and utilities for performance testing and evaluation.

Contents

  • Serving benchmarks: Scripts for testing online inference performance (latency, throughput)
  • Throughput benchmarks: Scripts for testing offline batch inference performance
  • Specialized benchmarks: Tools for testing specific features like structured output, prefix caching, long document QA, request prioritization, and multi-modal inference
  • Dataset utilities: Framework for loading and sampling from various benchmark datasets (ShareGPT, HuggingFace datasets, synthetic data, etc.)

Usage

For detailed usage instructions, examples, and dataset information, see the Benchmark CLI documentation.

For full CLI reference see: