docs/api/README.md
API documentation for vLLM's configuration classes.
LLM Class.
Prompt schema for LLM APIs.
Engine classes for offline and online inference.
Inference parameters for vLLM APIs.
vLLM provides experimental support for multi-modal models through the [vllm.multimodal][] package.
Multi-modal inputs can be passed alongside text and token prompts to supported models
via the multi_modal_data field in [vllm.inputs.PromptType][].
Looking to add your own multi-modal model? Please follow the instructions listed here.