Skip to content

LLMs

KARLI-hosted chat models are exposed as one provider among the standard set (OpenAI, Anthropic, etc.) and can be selected from any component that accepts a model — most commonly the Language Model and Agent components.

Under the hood, the KARLI provider speaks an OpenAI-compatible API: components select a KARLI model the same way they select an OpenAI model, but requests are routed to the KARLI endpoint.

Available Models

Model Notes
karli/KARLI-MoE-latest Default KARLI chat model; supports tool calling.

Selecting a KARLI Model

Open any model-aware component (e.g. Language Model or Agent), choose KARLI as the provider, and pick a model from the dropdown. The provider surfaces the shared KARLI_API_KEY / KARLI_BASE_URL credentials described in the Models overview.

agent-component-karli-llm-select.png

Provider Filtering

Some Agentlab dropdowns are scoped to a specific provider — for example, the extraction-model picker on the Read File component is filtered to KARLI only (see Data Extraction). Broader, deployment-level provider filtering (for data sovereignty or compliance) is enforced by the surrounding Karli Studio platform rather than by Agentlab itself.