Documentation Index
Fetch the complete documentation index at: https://docs.keephq.dev/llms.txt
Use this file to discover all available pages before exploring further.
Cloud Limitation
This provider is disabled for cloud environments and can only be used in local or self-hosted environments.Authentication
This provider requires authentication.- host: Llama.cpp Server Host URL (required: True, sensitive: False)
In workflows
This provider can be used in workflows. As “step” to query data, example:Connecting with the Provider
To use the Llama.cpp Provider:- Install Llama.cpp on your system
- Download or convert your model to GGUF format
- Start the Llama.cpp server with HTTP interface:
- Configure the host URL and model path in your Keep configuration
Prerequisites
- Llama.cpp must be installed and compiled with server support
- A GGUF format model file must be available on your system
- The Llama.cpp server must be running and accessible
- The server must have sufficient resources to load and run your model
Model Compatibility
The provider works with any GGUF format model compatible with Llama.cpp, including:- LLaMA and LLaMA-2 models
- Mistral models
- OpenLLaMA models
- Vicuna models
- And other compatible model architectures

