Supported Providers
vLLM Provider
The vLLM Provider enables integration with vLLM-deployed language models into Keep.
The vLLM Provider supports querying language models deployed with vLLM for prompt-based interactions.
Inputs
The vLLM Provider supports the following parameters:
prompt
: Interact with vLLM-deployed models by sending prompts and receiving responsesmodel
: The model to be used, defaults toQwen/Qwen1.5-1.8B-Chat
temperature
: Controls randomness in the response, defaults to 0.7max_tokens
: Limit amount of tokens returned by the model, defaults to 1024structured_output_format
: Optional JSON schema for structured output formatting
Outputs
The vLLM Provider returns the model’s response based on the provided prompt. When using structured output format, the response will be formatted according to the provided JSON schema.
Authentication Parameters
To use the vLLM Provider, you’ll need to configure the following authentication parameters:
- api_url (required): The endpoint URL where your vLLM service is deployed
- api_key (optional): API key if your vLLM deployment requires authentication
Connecting with the Provider
To connect to a vLLM deployment:
- Deploy your vLLM instance or obtain the API endpoint of an existing deployment
- Configure the API URL in your provider configuration
- If your deployment requires authentication, configure the API key
Structured Output
Structure output for vLLM should follow Json Schema notation. Example: