vllm.entrypoints.cli
Modules:
Name | Description |
---|---|
benchmark |
|
collect_env |
|
main |
The CLI entrypoints of vLLM |
openai |
|
run_batch |
|
serve |
|
types |
|
__all__
module-attribute
¶
__all__: list[str] = [
"BenchmarkLatencySubcommand",
"BenchmarkServingSubcommand",
"BenchmarkThroughputSubcommand",
]
BenchmarkLatencySubcommand
¶
Bases: BenchmarkSubcommandBase
The latency
subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/latency.py
help
class-attribute
instance-attribute
¶
add_cli_args
classmethod
¶
add_cli_args(parser: ArgumentParser) -> None
BenchmarkServingSubcommand
¶
Bases: BenchmarkSubcommandBase
The serve
subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/serve.py
add_cli_args
classmethod
¶
add_cli_args(parser: ArgumentParser) -> None
BenchmarkThroughputSubcommand
¶
Bases: BenchmarkSubcommandBase
The throughput
subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/throughput.py
add_cli_args
classmethod
¶
add_cli_args(parser: ArgumentParser) -> None