pixano_inference.impls.vllm.vlm
vLLM-based VLM (Vision-Language Model).
VLLMVLMModel(config)
Bases: VLMModel
Native Ray Serve model for vLLM-based VLMs.
model_params contract:
path(str, required): HuggingFace model ID.config(dict, optional): Kwargs forvllm.LLM.processor_config(dict, optional): Kwargs forvllm.LLMprocessor options.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
config
|
ModelDeploymentConfig
|
Model deployment configuration. |
required |
Source code in pixano_inference/impls/vllm/vlm.py
metadata
property
Model metadata.
load_model()
Load the vLLM model.
Source code in pixano_inference/impls/vllm/vlm.py
predict(input)
Run VLM generation via vLLM.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
input
|
VLMInput
|
VLM input with prompt and generation parameters.
Images must be embedded in the prompt for vLLM ( |
required |
Returns:
| Type | Description |
|---|---|
VLMOutput
|
VLM output with generated text, usage info, and generation config. |