This guide provides detailed information on how to configure each provider in Kilo Code CLI. Each provider has specific configuration requirements and optional settings that can be customized to suit your needs.
Kilo Code CLI supports multiple AI providers, each with their own configuration requirements. This document details the configuration fields for each provider, including required and optional parameters.
You can configure providers using:
kilocode config to configure providers interactivelyThe official Kilo Code provider for accessing Kilo Code's managed AI services.
Description: Access Kilo Code's managed AI infrastructure with support for multiple models and organizations.
Required Fields:
kilocodeToken (password): Your Kilo Code authentication tokenkilocodeModel (text): The model to use (default: anthropic/claude-sonnet-4.5)Optional Fields:
kilocodeOrganizationId (text): Organization ID for team accounts (leave empty for personal use)Example Configuration:
{
"id": "default",
"provider": "kilocode",
"kilocodeToken": "your-token-here",
"kilocodeModel": "anthropic/claude-sonnet-4",
"kilocodeOrganizationId": "org-123456"
}
Default Model: anthropic/claude-sonnet-4.5
Direct integration with Anthropic's Claude API.
Description: Use Claude models directly from Anthropic with your own API key.
Required Fields:
apiKey (password): Your Anthropic API keyapiModelId (text): The Claude model to use (default: claude-sonnet-4.5)Optional Fields:
anthropicBaseUrl (text): Custom base URL for API requests (leave empty for default)Example Configuration:
{
"id": "default",
"provider": "anthropic",
"apiKey": "sk-ant-...",
"apiModelId": "claude-sonnet-4.5",
"anthropicBaseUrl": ""
}
Default Model: claude-sonnet-4.5
Notes:
Native OpenAI API integration.
Description: Use OpenAI's models with native API support.
Required Fields:
openAiNativeApiKey (password): Your OpenAI API keyapiModelId (text): The OpenAI model to use (default: gpt-5-chat-latest)Optional Fields:
openAiNativeBaseUrl (text): Custom base URL for API requests (leave empty for default)Example Configuration:
{
"id": "default",
"provider": "openai-native",
"openAiNativeApiKey": "sk-...",
"apiModelId": "gpt-5-chat-latest",
"openAiNativeBaseUrl": ""
}
Default Model: gpt-5-chat-latest
Notes:
Access multiple AI models through OpenRouter's unified API.
Description: Use OpenRouter to access various AI models from different providers through a single API.
Required Fields:
openRouterApiKey (password): Your OpenRouter API keyopenRouterModelId (text): The model identifier (default: anthropic/claude-3-5-sonnet)Optional Fields:
openRouterBaseUrl (text): Custom base URL (leave empty for default)Example Configuration:
{
"id": "default",
"provider": "openrouter",
"openRouterApiKey": "sk-or-...",
"openRouterModelId": "anthropic/claude-3-5-sonnet",
"openRouterBaseUrl": ""
}
Default Model: anthropic/claude-3-5-sonnet
Notes:
AWS Bedrock for accessing foundation models on AWS infrastructure.
Description: Use AWS Bedrock to access various foundation models with AWS security and compliance.
Required Fields:
awsAccessKey (password): Your AWS access key IDawsSecretKey (password): Your AWS secret access keyawsRegion (text): AWS region (default: us-east-1)apiModelId (text): The model to use (default: anthropic.claude-sonnet-4.5-20250929-v1:0)Optional Fields:
awsSessionToken (password): AWS session token for temporary credentialsawsUseCrossRegionInference (boolean): Enable cross-region inferenceExample Configuration:
{
"id": "default",
"provider": "bedrock",
"awsAccessKey": "AKIA...",
"awsSecretKey": "...",
"awsRegion": "us-east-1",
"apiModelId": "anthropic.claude-sonnet-4.5-20250929-v1:0",
"awsSessionToken": "",
"awsUseCrossRegionInference": false
}
Default Model: anthropic.claude-sonnet-4.5-20250929-v1:0
Notes:
Google's Gemini AI models via direct API access.
Description: Access Google's Gemini models directly with your API key.
Required Fields:
geminiApiKey (password): Your Google AI API keyapiModelId (text): The model to use (default: gemini-2.5-flash-preview-04-17)Optional Fields:
googleGeminiBaseUrl (text): Custom base URL (leave empty for default)Example Configuration:
{
"id": "default",
"provider": "gemini",
"geminiApiKey": "AIza...",
"apiModelId": "gemini-2.5-flash-preview-04-17",
"googleGeminiBaseUrl": ""
}
Default Model: gemini-2.5-flash-preview-04-17
Notes:
Google Cloud Vertex AI for enterprise-grade AI deployment.
Description: Use Google Cloud's Vertex AI platform for accessing AI models with enterprise features.
Required Fields:
vertexProjectId (text): Your Google Cloud project IDvertexRegion (text): Google Cloud region (default: us-central1)apiModelId (text): The model to use (default: claude-4.5-sonnet)Authentication (choose one):
vertexJsonCredentials (password): JSON service account credentialsvertexKeyFile (text): Path to service account key fileExample Configuration:
{
"id": "default",
"provider": "vertex",
"vertexProjectId": "my-project-123",
"vertexRegion": "us-central1",
"apiModelId": "claude-4.5-sonnet",
"vertexJsonCredentials": "{...}",
"vertexKeyFile": ""
}
Default Model: claude-4.5-sonnet
Notes:
Local Claude Code CLI integration.
Description: Use the Claude Code CLI tool for local AI interactions.
Required Fields:
claudeCodePath (text): Path to the Claude Code executableapiModelId (text): The model to use (default: claude-sonnet-4-5)claudeCodeMaxOutputTokens (text): Maximum output tokens (default: 8000)Example Configuration:
{
"id": "default",
"provider": "claude-code",
"claudeCodePath": "/usr/local/bin/claude-code",
"apiModelId": "claude-sonnet-4-5",
"claudeCodeMaxOutputTokens": "8000"
}
Default Model: claude-sonnet-4-5
Notes:
Mistral AI's language models.
Description: Access Mistral's powerful language models including Codestral for code generation.
Required Fields:
mistralApiKey (password): Your Mistral API keyapiModelId (text): The model to use (default: magistral-medium-latest)Optional Fields:
mistralCodestralUrl (text): Custom Codestral base URL (leave empty for default)Example Configuration:
{
"id": "default",
"provider": "mistral",
"mistralApiKey": "...",
"apiModelId": "magistral-medium-latest",
"mistralCodestralUrl": ""
}
Default Model: magistral-medium-latest
Notes:
Groq's ultra-fast LPU inference.
Description: Use Groq's Language Processing Unit (LPU) for extremely fast inference.
Required Fields:
groqApiKey (password): Your Groq API keyapiModelId (text): The model to use (default: llama-3.3-70b-versatile)Example Configuration:
{
"id": "default",
"provider": "groq",
"groqApiKey": "gsk_...",
"apiModelId": "llama-3.3-70b-versatile"
}
Default Model: llama-3.3-70b-versatile
Notes:
DeepSeek's AI models.
Description: Access DeepSeek's language models optimized for coding and reasoning.
Required Fields:
deepSeekApiKey (password): Your DeepSeek API keyapiModelId (text): The model to use (default: deepseek-chat)Example Configuration:
{
"id": "default",
"provider": "deepseek",
"deepSeekApiKey": "...",
"apiModelId": "deepseek-chat"
}
Default Model: deepseek-chat
Notes:
xAI's Grok models.
Description: Access xAI's Grok language models.
Required Fields:
xaiApiKey (password): Your xAI API keyapiModelId (text): The model to use (default: grok-code-fast-1)Example Configuration:
{
"id": "default",
"provider": "xai",
"xaiApiKey": "...",
"apiModelId": "grok-code-fast-1"
}
Default Model: grok-code-fast-1
Notes:
Cerebras AI inference platform.
Description: Use Cerebras' wafer-scale AI inference platform.
Required Fields:
cerebrasApiKey (password): Your Cerebras API keyapiModelId (text): The model to use (default: qwen-3-coder-480b-free)Example Configuration:
{
"id": "default",
"provider": "cerebras",
"cerebrasApiKey": "...",
"apiModelId": "qwen-3-coder-480b-free"
}
Default Model: qwen-3-coder-480b-free
Notes:
Local Ollama instance for running models locally.
Description: Run AI models locally using Ollama.
Required Fields:
ollamaBaseUrl (text): Ollama server URL (default: http://localhost:11434)ollamaModelId (text): Model identifier (default: llama3.2)Optional Fields:
ollamaApiKey (password): API key if authentication is enabledExample Configuration:
{
"id": "default",
"provider": "ollama",
"ollamaBaseUrl": "http://localhost:11434",
"ollamaModelId": "llama3.2",
"ollamaApiKey": ""
}
Default Model: llama3.2
Notes:
LM Studio for local model inference.
Description: Use LM Studio to run models locally with a user-friendly interface.
Required Fields:
lmStudioBaseUrl (text): LM Studio server URL (default: http://localhost:1234/v1)lmStudioModelId (text): Model identifier (default: local-model)Optional Fields:
lmStudioSpeculativeDecodingEnabled (boolean): Enable speculative decoding for faster inferenceExample Configuration:
{
"id": "default",
"provider": "lmstudio",
"lmStudioBaseUrl": "http://localhost:1234/v1",
"lmStudioModelId": "local-model",
"lmStudioSpeculativeDecodingEnabled": false
}
Default Model: local-model
Notes:
VSCode's built-in language model API.
Description: Use VSCode's native language model capabilities (e.g., GitHub Copilot).
Required Fields:
vsCodeLmModelSelector (text): Model selector in format vendor/familyExample Configuration:
{
"id": "default",
"provider": "vscode-lm",
"vsCodeLmModelSelector": {
"vendor": "copilot",
"family": "gpt-4o"
}
}
Default Model: copilot-gpt-4o
Notes:
OpenAI API integration (alternative configuration).
Description: Alternative OpenAI integration with simplified configuration.
Required Fields:
openAiApiKey (password): Your OpenAI API keyapiModelId (text): The model to use (default: gpt-4o)Optional Fields:
openAiBaseUrl (text): Custom base URL (leave empty for default)Example Configuration:
{
"id": "default",
"provider": "openai",
"openAiApiKey": "sk-...",
"apiModelId": "gpt-4o",
"openAiBaseUrl": ""
}
Default Model: gpt-4o
Notes:
Glama AI platform.
Description: Access AI models through the Glama platform.
Required Fields:
glamaApiKey (password): Your Glama API keyglamaModelId (text): Model identifier (default: llama-3.1-70b-versatile)Example Configuration:
{
"id": "default",
"provider": "glama",
"glamaApiKey": "...",
"glamaModelId": "llama-3.1-70b-versatile"
}
Default Model: llama-3.1-70b-versatile
HuggingFace Inference API.
Description: Access models hosted on HuggingFace's inference infrastructure.
Required Fields:
huggingFaceApiKey (password): Your HuggingFace API tokenhuggingFaceModelId (text): Model identifier (default: meta-llama/Llama-2-70b-chat-hf)huggingFaceInferenceProvider (text): Inference provider (default: auto)Example Configuration:
{
"id": "default",
"provider": "huggingface",
"huggingFaceApiKey": "hf_...",
"huggingFaceModelId": "meta-llama/Llama-2-70b-chat-hf",
"huggingFaceInferenceProvider": "auto"
}
Default Model: meta-llama/Llama-2-70b-chat-hf
Notes:
auto, hf-inference, or specific endpointsLiteLLM proxy for unified model access.
Description: Use LiteLLM as a proxy to access multiple AI providers through a unified interface.
Required Fields:
litellmBaseUrl (text): LiteLLM proxy URLlitellmApiKey (password): API key for the proxylitellmModelId (text): Model identifier (default: gpt-4o)Example Configuration:
{
"id": "default",
"provider": "litellm",
"litellmBaseUrl": "http://localhost:8000",
"litellmApiKey": "...",
"litellmModelId": "gpt-4o"
}
Default Model: gpt-4o
Notes:
Moonshot AI platform.
Description: Access Moonshot AI's language models.
Required Fields:
moonshotBaseUrl (text): Moonshot API base URL (default: https://api.moonshot.ai/v1)moonshotApiKey (password): Your Moonshot API keyapiModelId (text): The model to use (default: kimi-k2-0711-preview)Example Configuration:
{
"id": "default",
"provider": "moonshot",
"moonshotBaseUrl": "https://api.moonshot.ai/v1",
"moonshotApiKey": "...",
"apiModelId": "kimi-k2-0711-preview"
}
Default Model: kimi-k2-0711-preview
Doubao AI platform.
Description: Access Doubao's AI models.
Required Fields:
doubaoApiKey (password): Your Doubao API keyapiModelId (text): The model to use (default: doubao-seed-1-6-250615)Example Configuration:
{
"id": "default",
"provider": "doubao",
"doubaoApiKey": "...",
"apiModelId": "doubao-seed-1-6-250615"
}
Default Model: doubao-seed-1-6-250615
Chutes AI platform.
Description: Access AI models through the Chutes platform.
Required Fields:
chutesApiKey (password): Your Chutes API keyapiModelId (text): The model to use (default: deepseek-ai/DeepSeek-R1-0528)Example Configuration:
{
"id": "default",
"provider": "chutes",
"chutesApiKey": "...",
"apiModelId": "deepseek-ai/DeepSeek-R1-0528"
}
Default Model: deepseek-ai/DeepSeek-R1-0528
SambaNova AI inference platform.
Description: Use SambaNova's AI inference platform for fast model execution.
Required Fields:
sambaNovaApiKey (password): Your SambaNova API keyapiModelId (text): The model to use (default: Meta-Llama-3.1-70B-Instruct)Example Configuration:
{
"id": "default",
"provider": "sambanova",
"sambaNovaApiKey": "...",
"apiModelId": "Meta-Llama-3.1-70B-Instruct"
}
Default Model: Meta-Llama-3.1-70B-Instruct
Fireworks AI platform.
Description: Access models through Fireworks AI's fast inference platform.
Required Fields:
fireworksApiKey (password): Your Fireworks API keyapiModelId (text): The model to use (default: accounts/fireworks/models/kimi-k2-instruct-0905)Example Configuration:
{
"id": "default",
"provider": "fireworks",
"fireworksApiKey": "...",
"apiModelId": "accounts/fireworks/models/kimi-k2-instruct-0905"
}
Default Model: accounts/fireworks/models/kimi-k2-instruct-0905
Notes:
Featherless AI platform.
Description: Access AI models through the Featherless platform.
Required Fields:
featherlessApiKey (password): Your Featherless API keyapiModelId (text): The model to use (default: deepseek-ai/DeepSeek-V3-0324)Example Configuration:
{
"id": "default",
"provider": "featherless",
"featherlessApiKey": "...",
"apiModelId": "deepseek-ai/DeepSeek-V3-0324"
}
Default Model: deepseek-ai/DeepSeek-V3-0324
DeepInfra's serverless AI inference.
Description: Use DeepInfra for serverless access to various AI models.
Required Fields:
deepInfraApiKey (password): Your DeepInfra API keydeepInfraModelId (text): Model identifier (default: meta-llama/Meta-Llama-3.1-70B-Instruct)Example Configuration:
{
"id": "default",
"provider": "deepinfra",
"deepInfraApiKey": "...",
"deepInfraModelId": "meta-llama/Meta-Llama-3.1-70B-Instruct"
}
Default Model: meta-llama/Meta-Llama-3.1-70B-Instruct
Notes:
IO Intelligence platform.
Description: Access AI models through the IO Intelligence platform.
Required Fields:
ioIntelligenceApiKey (password): Your IO Intelligence API keyioIntelligenceModelId (text): Model identifier (default: gpt-4o)Example Configuration:
{
"id": "default",
"provider": "io-intelligence",
"ioIntelligenceApiKey": "...",
"ioIntelligenceModelId": "gpt-4o"
}
Default Model: gpt-4o
Qwen Code AI models.
Description: Access Qwen's code-specialized models using OAuth authentication.
Required Fields:
qwenCodeOauthPath (text): Path to OAuth credentials file (default: ~/.qwen/oauth_creds.json)apiModelId (text): The model to use (default: qwen3-coder-plus)Example Configuration:
{
"id": "default",
"provider": "qwen-code",
"qwenCodeOauthPath": "~/.qwen/oauth_creds.json",
"apiModelId": "qwen3-coder-plus"
}
Default Model: qwen3-coder-plus
Notes:
Gemini CLI integration.
Description: Use Google's Gemini models through CLI with OAuth authentication.
Required Fields:
geminiCliOAuthPath (text): Path to OAuth credentials file (default: ~/.gemini/oauth_creds.json)geminiCliProjectId (text): Google Cloud project IDapiModelId (text): The model to use (default: gemini-2.5-flash-preview-04-17)Example Configuration:
{
"id": "default",
"provider": "gemini-cli",
"geminiCliOAuthPath": "~/.gemini/oauth_creds.json",
"geminiCliProjectId": "my-project-123",
"apiModelId": "gemini-2.5-flash-preview-04-17"
}
Default Model: gemini-2.5-flash-preview-04-17
Notes:
ZAI AI platform.
Description: Access AI models through the ZAI platform with support for both international and China-based API endpoints.
Required Fields:
zaiApiKey (password): Your ZAI API keyzaiApiLine (text): API line identifier (default: international_coding)apiModelId (text): The model to use (default: glm-4.6)Available API Lines:
The zaiApiLine parameter determines which API endpoint and region to use:
international_coding (default): International Coding Plan
https://api.z.ai/api/coding/paas/v4international: International Standard
https://api.z.ai/api/paas/v4china_coding: China Coding Plan
https://open.bigmodel.cn/api/coding/paas/v4china: China Standard
https://open.bigmodel.cn/api/paas/v4Example Configuration:
{
"id": "default",
"provider": "zai",
"zaiApiKey": "...",
"zaiApiLine": "international_coding",
"apiModelId": "glm-4.6"
}
Default Model: glm-4.6
Notes:
Unbound AI platform.
Description: Access AI models through the Unbound platform.
Required Fields:
unboundApiKey (password): Your Unbound API keyunboundModelId (text): Model identifier (default: gpt-4o)Example Configuration:
{
"id": "default",
"provider": "unbound",
"unboundApiKey": "...",
"unboundModelId": "gpt-4o"
}
Default Model: gpt-4o
Requesty AI platform.
Description: Access AI models through the Requesty platform.
Required Fields:
requestyApiKey (password): Your Requesty API keyrequestyModelId (text): Model identifier (default: gpt-4o)Optional Fields:
requestyBaseUrl (text): Custom base URL (leave empty for default)Example Configuration:
{
"id": "default",
"provider": "requesty",
"requestyApiKey": "...",
"requestyBaseUrl": "",
"requestyModelId": "gpt-4o"
}
Default Model: gpt-4o
Roo AI platform.
Description: Access AI models through the Roo platform.
Required Fields:
apiModelId (text): Model identifier (default: deepseek-ai/DeepSeek-R1-0528)Example Configuration:
{
"id": "default",
"provider": "roo",
"apiModelId": "deepseek-ai/DeepSeek-R1-0528"
}
Default Model: deepseek-ai/DeepSeek-R1-0528
Notes:
Vercel AI Gateway for unified model access.
Description: Use Vercel's AI Gateway to access multiple AI providers.
Required Fields:
vercelAiGatewayApiKey (password): Your Vercel AI Gateway API keyvercelAiGatewayModelId (text): Model identifier (default: gpt-4o)Example Configuration:
{
"id": "default",
"provider": "vercel-ai-gateway",
"vercelAiGatewayApiKey": "...",
"vercelAiGatewayModelId": "gpt-4o"
}
Default Model: gpt-4o
Notes:
Virtual quota management with automatic fallback.
Description: Manage multiple provider profiles with automatic fallback when quotas are exceeded.
Required Fields:
profiles (text): Array of provider profiles with quota configurationsExample Configuration:
{
"id": "default",
"provider": "virtual-quota-fallback",
"profiles": [
{
"provider": "anthropic",
"quota": 1000000,
"config": {
"apiKey": "...",
"apiModelId": "claude-3-5-sonnet-20241022"
}
},
{
"provider": "openai",
"quota": 500000,
"config": {
"openAiApiKey": "...",
"apiModelId": "gpt-4o"
}
}
]
}
Default Model: gpt-4o
Notes:
Human-in-the-loop relay for manual responses.
Description: Route requests to a human operator for manual responses.
Required Fields:
apiModelId (text): Model identifier (fixed value: human)Example Configuration:
{
"id": "default",
"provider": "human-relay",
"apiModelId": "human"
}
Default Model: human
Notes:
Fake AI provider for testing and development.
Description: Mock AI provider for testing purposes without making actual API calls.
Required Fields:
apiModelId (text): Model identifier (fixed value: fake-model)Example Configuration:
{
"id": "default",
"provider": "fake-ai",
"apiModelId": "fake-model"
}
Default Model: fake-model
Notes:
For issues or questions about provider configuration: