llm --help

Let's help help help devs.

Target: <200ms. Several important LLM CLI tools take multiple seconds. PRs welcome ❤️

library cold warm (10 runs) version measured on
vllm --help15610ms7473ms0.14.1+cpu2026-01-28T21:50Z
sglang --help13875ms5490msv0.5.72026-01-05T00:09Z
VLMEvalKit --help14551ms4990msv0.22026-01-05T00:28Z
transformers --help8799ms3102ms4.57.32026-01-05T00:23Z
tensorrt-llm --help8208ms2372ms1.0.02026-01-05T00:04Z
datasets --help3351ms842ms4.4.22026-01-05T00:21Z
llm --help1257ms573ms0.282026-01-22T16:52Z
openai --help1591ms505ms2.14.02026-01-05T00:11Z
langchain-cli --help749ms253ms0.0.372026-01-05T00:03Z
hf --help805ms218ms1.2.32026-01-05T00:14Z
lm-eval --help49ms43ms0.4.9.22026-01-05T00:08Z
llama.cpp --help26ms17msb76262026-01-05T00:12Z
ollama --help14ms14ms0.13.52026-01-05T00:30Z