Skip to content

Models

Models define the language models that power agent conversations. Evonic supports multiple model providers and local inference engines.

RunnerBest ForHardware Requirements
OllamaQuick setup, multi-model supportModerate (CPU/GPU)
llama.cppMaximum portability, edge devicesLow to Moderate
vLLMHigh-throughput production workloadsHigh (GPU recommended)

Models are managed via:

  • CLIevonic model list, evonic model add, evonic model rm
  • API — REST endpoints for programmatic management

Each model has a provider (e.g., openai, anthropic, llama.cpp), a base URL, API key, and configuration options like max tokens, timeout, and temperature.