Models
Models define the language models that power agent conversations. Evonic supports multiple model providers and local inference engines.
Supported Runners
Section titled “Supported Runners”| Runner | Best For | Hardware Requirements |
|---|---|---|
| Ollama | Quick setup, multi-model support | Moderate (CPU/GPU) |
| llama.cpp | Maximum portability, edge devices | Low to Moderate |
| vLLM | High-throughput production workloads | High (GPU recommended) |
Model Configuration
Section titled “Model Configuration”Models are managed via:
- CLI —
evonic model list,evonic model add,evonic model rm - API — REST endpoints for programmatic management
Each model has a provider (e.g., openai, anthropic, llama.cpp), a base URL, API key, and configuration options like max tokens, timeout, and temperature.
Learn More
Section titled “Learn More”- Local Models Overview — why local-first and model selection
- Ollama — setup and configuration
- llama.cpp — setup and configuration
- vLLM — setup and configuration
- Model Selection
- Quantization