| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705 |
- import type { APICallError, ModelMessage } from "ai"
- import { unique } from "remeda"
- import type { JSONSchema } from "zod/v4/core"
- import type { Provider } from "./provider"
- import type { ModelsDev } from "./models"
- import { iife } from "@/util/iife"
- type Modality = NonNullable<ModelsDev.Model["modalities"]>["input"][number]
- function mimeToModality(mime: string): Modality | undefined {
- if (mime.startsWith("image/")) return "image"
- if (mime.startsWith("audio/")) return "audio"
- if (mime.startsWith("video/")) return "video"
- if (mime === "application/pdf") return "pdf"
- return undefined
- }
- export namespace ProviderTransform {
- // Maps npm package to the key the AI SDK expects for providerOptions
- function sdkKey(npm: string): string | undefined {
- switch (npm) {
- case "@ai-sdk/github-copilot":
- case "@ai-sdk/openai":
- case "@ai-sdk/azure":
- return "openai"
- case "@ai-sdk/amazon-bedrock":
- return "bedrock"
- case "@ai-sdk/anthropic":
- return "anthropic"
- case "@ai-sdk/google-vertex":
- case "@ai-sdk/google":
- return "google"
- case "@ai-sdk/gateway":
- return "gateway"
- case "@openrouter/ai-sdk-provider":
- return "openrouter"
- }
- return undefined
- }
- function normalizeMessages(
- msgs: ModelMessage[],
- model: Provider.Model,
- options: Record<string, unknown>,
- ): ModelMessage[] {
- // Anthropic rejects messages with empty content - filter out empty string messages
- // and remove empty text/reasoning parts from array content
- if (model.api.npm === "@ai-sdk/anthropic") {
- msgs = msgs
- .map((msg) => {
- if (typeof msg.content === "string") {
- if (msg.content === "") return undefined
- return msg
- }
- if (!Array.isArray(msg.content)) return msg
- const filtered = msg.content.filter((part) => {
- if (part.type === "text" || part.type === "reasoning") {
- return part.text !== ""
- }
- return true
- })
- if (filtered.length === 0) return undefined
- return { ...msg, content: filtered }
- })
- .filter((msg): msg is ModelMessage => msg !== undefined && msg.content !== "")
- }
- if (model.api.id.includes("claude")) {
- return msgs.map((msg) => {
- if ((msg.role === "assistant" || msg.role === "tool") && Array.isArray(msg.content)) {
- msg.content = msg.content.map((part) => {
- if ((part.type === "tool-call" || part.type === "tool-result") && "toolCallId" in part) {
- return {
- ...part,
- toolCallId: part.toolCallId.replace(/[^a-zA-Z0-9_-]/g, "_"),
- }
- }
- return part
- })
- }
- return msg
- })
- }
- if (model.providerID === "mistral" || model.api.id.toLowerCase().includes("mistral")) {
- const result: ModelMessage[] = []
- for (let i = 0; i < msgs.length; i++) {
- const msg = msgs[i]
- const nextMsg = msgs[i + 1]
- if ((msg.role === "assistant" || msg.role === "tool") && Array.isArray(msg.content)) {
- msg.content = msg.content.map((part) => {
- if ((part.type === "tool-call" || part.type === "tool-result") && "toolCallId" in part) {
- // Mistral requires alphanumeric tool call IDs with exactly 9 characters
- const normalizedId = part.toolCallId
- .replace(/[^a-zA-Z0-9]/g, "") // Remove non-alphanumeric characters
- .substring(0, 9) // Take first 9 characters
- .padEnd(9, "0") // Pad with zeros if less than 9 characters
- return {
- ...part,
- toolCallId: normalizedId,
- }
- }
- return part
- })
- }
- result.push(msg)
- // Fix message sequence: tool messages cannot be followed by user messages
- if (msg.role === "tool" && nextMsg?.role === "user") {
- result.push({
- role: "assistant",
- content: [
- {
- type: "text",
- text: "Done.",
- },
- ],
- })
- }
- }
- return result
- }
- if (
- model.capabilities.interleaved &&
- typeof model.capabilities.interleaved === "object" &&
- model.capabilities.interleaved.field === "reasoning_content"
- ) {
- return msgs.map((msg) => {
- if (msg.role === "assistant" && Array.isArray(msg.content)) {
- const reasoningParts = msg.content.filter((part: any) => part.type === "reasoning")
- const reasoningText = reasoningParts.map((part: any) => part.text).join("")
- // Filter out reasoning parts from content
- const filteredContent = msg.content.filter((part: any) => part.type !== "reasoning")
- // Include reasoning_content directly on the message for all assistant messages
- if (reasoningText) {
- return {
- ...msg,
- content: filteredContent,
- providerOptions: {
- ...msg.providerOptions,
- openaiCompatible: {
- ...(msg.providerOptions as any)?.openaiCompatible,
- reasoning_content: reasoningText,
- },
- },
- }
- }
- return {
- ...msg,
- content: filteredContent,
- }
- }
- return msg
- })
- }
- return msgs
- }
- function applyCaching(msgs: ModelMessage[], providerID: string): ModelMessage[] {
- const system = msgs.filter((msg) => msg.role === "system").slice(0, 2)
- const final = msgs.filter((msg) => msg.role !== "system").slice(-2)
- const providerOptions = {
- anthropic: {
- cacheControl: { type: "ephemeral" },
- },
- openrouter: {
- cacheControl: { type: "ephemeral" },
- },
- bedrock: {
- cachePoint: { type: "ephemeral" },
- },
- openaiCompatible: {
- cache_control: { type: "ephemeral" },
- },
- }
- for (const msg of unique([...system, ...final])) {
- const shouldUseContentOptions = providerID !== "anthropic" && Array.isArray(msg.content) && msg.content.length > 0
- if (shouldUseContentOptions) {
- const lastContent = msg.content[msg.content.length - 1]
- if (lastContent && typeof lastContent === "object") {
- lastContent.providerOptions = {
- ...lastContent.providerOptions,
- ...providerOptions,
- }
- continue
- }
- }
- msg.providerOptions = {
- ...msg.providerOptions,
- ...providerOptions,
- }
- }
- return msgs
- }
- function unsupportedParts(msgs: ModelMessage[], model: Provider.Model): ModelMessage[] {
- return msgs.map((msg) => {
- if (msg.role !== "user" || !Array.isArray(msg.content)) return msg
- const filtered = msg.content.map((part) => {
- if (part.type !== "file" && part.type !== "image") return part
- // Check for empty base64 image data
- if (part.type === "image") {
- const imageStr = part.image.toString()
- if (imageStr.startsWith("data:")) {
- const match = imageStr.match(/^data:([^;]+);base64,(.*)$/)
- if (match && (!match[2] || match[2].length === 0)) {
- return {
- type: "text" as const,
- text: "ERROR: Image file is empty or corrupted. Please provide a valid image.",
- }
- }
- }
- }
- const mime = part.type === "image" ? part.image.toString().split(";")[0].replace("data:", "") : part.mediaType
- const filename = part.type === "file" ? part.filename : undefined
- const modality = mimeToModality(mime)
- if (!modality) return part
- if (model.capabilities.input[modality]) return part
- const name = filename ? `"${filename}"` : modality
- return {
- type: "text" as const,
- text: `ERROR: Cannot read ${name} (this model does not support ${modality} input). Inform the user.`,
- }
- })
- return { ...msg, content: filtered }
- })
- }
- export function message(msgs: ModelMessage[], model: Provider.Model, options: Record<string, unknown>) {
- msgs = unsupportedParts(msgs, model)
- msgs = normalizeMessages(msgs, model, options)
- if (
- model.providerID === "anthropic" ||
- model.api.id.includes("anthropic") ||
- model.api.id.includes("claude") ||
- model.api.npm === "@ai-sdk/anthropic"
- ) {
- msgs = applyCaching(msgs, model.providerID)
- }
- // Remap providerOptions keys from stored providerID to expected SDK key
- const key = sdkKey(model.api.npm)
- if (key && key !== model.providerID && model.api.npm !== "@ai-sdk/azure") {
- const remap = (opts: Record<string, any> | undefined) => {
- if (!opts) return opts
- if (!(model.providerID in opts)) return opts
- const result = { ...opts }
- result[key] = result[model.providerID]
- delete result[model.providerID]
- return result
- }
- msgs = msgs.map((msg) => {
- if (!Array.isArray(msg.content)) return { ...msg, providerOptions: remap(msg.providerOptions) }
- return {
- ...msg,
- providerOptions: remap(msg.providerOptions),
- content: msg.content.map((part) => ({ ...part, providerOptions: remap(part.providerOptions) })),
- } as typeof msg
- })
- }
- return msgs
- }
- export function temperature(model: Provider.Model) {
- const id = model.id.toLowerCase()
- if (id.includes("qwen")) return 0.55
- if (id.includes("claude")) return undefined
- if (id.includes("gemini")) return 1.0
- if (id.includes("glm-4.6")) return 1.0
- if (id.includes("glm-4.7")) return 1.0
- if (id.includes("minimax-m2")) return 1.0
- if (id.includes("kimi-k2")) {
- if (id.includes("thinking")) return 1.0
- return 0.6
- }
- return undefined
- }
- export function topP(model: Provider.Model) {
- const id = model.id.toLowerCase()
- if (id.includes("qwen")) return 1
- if (id.includes("minimax-m2")) {
- return 0.95
- }
- if (id.includes("gemini")) return 0.95
- return undefined
- }
- export function topK(model: Provider.Model) {
- const id = model.id.toLowerCase()
- if (id.includes("minimax-m2")) {
- if (id.includes("m2.1")) return 40
- return 20
- }
- if (id.includes("gemini")) return 64
- return undefined
- }
- const WIDELY_SUPPORTED_EFFORTS = ["low", "medium", "high"]
- const OPENAI_EFFORTS = ["none", "minimal", ...WIDELY_SUPPORTED_EFFORTS, "xhigh"]
- export function variants(model: Provider.Model): Record<string, Record<string, any>> {
- if (!model.capabilities.reasoning) return {}
- const id = model.id.toLowerCase()
- if (id.includes("deepseek") || id.includes("minimax") || id.includes("glm") || id.includes("mistral")) return {}
- switch (model.api.npm) {
- case "@openrouter/ai-sdk-provider":
- if (!model.id.includes("gpt") && !model.id.includes("gemini-3") && !model.id.includes("grok-4")) return {}
- return Object.fromEntries(OPENAI_EFFORTS.map((effort) => [effort, { reasoning: { effort } }]))
- // TODO: YOU CANNOT SET max_tokens if this is set!!!
- case "@ai-sdk/gateway":
- return Object.fromEntries(OPENAI_EFFORTS.map((effort) => [effort, { reasoningEffort: effort }]))
- case "@ai-sdk/cerebras":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/cerebras
- case "@ai-sdk/togetherai":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/togetherai
- case "@ai-sdk/xai":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/xai
- case "@ai-sdk/deepinfra":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/deepinfra
- case "@ai-sdk/openai-compatible":
- return Object.fromEntries(WIDELY_SUPPORTED_EFFORTS.map((effort) => [effort, { reasoningEffort: effort }]))
- case "@ai-sdk/azure":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/azure
- if (id === "o1-mini") return {}
- const azureEfforts = ["low", "medium", "high"]
- if (id.includes("gpt-5-") || id === "gpt-5") {
- azureEfforts.unshift("minimal")
- }
- return Object.fromEntries(
- azureEfforts.map((effort) => [
- effort,
- {
- reasoningEffort: effort,
- reasoningSummary: "auto",
- include: ["reasoning.encrypted_content"],
- },
- ]),
- )
- case "@ai-sdk/openai":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/openai
- if (id === "gpt-5-pro") return {}
- const openaiEfforts = iife(() => {
- if (id.includes("codex")) {
- if (id.includes("5.2")) return [...WIDELY_SUPPORTED_EFFORTS, "xhigh"]
- return WIDELY_SUPPORTED_EFFORTS
- }
- const arr = [...WIDELY_SUPPORTED_EFFORTS]
- if (id.includes("gpt-5-") || id === "gpt-5") {
- arr.unshift("minimal")
- }
- if (model.release_date >= "2025-11-13") {
- arr.unshift("none")
- }
- if (model.release_date >= "2025-12-04") {
- arr.push("xhigh")
- }
- return arr
- })
- return Object.fromEntries(
- openaiEfforts.map((effort) => [
- effort,
- {
- reasoningEffort: effort,
- reasoningSummary: "auto",
- include: ["reasoning.encrypted_content"],
- },
- ]),
- )
- case "@ai-sdk/anthropic":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/anthropic
- return {
- high: {
- thinking: {
- type: "enabled",
- budgetTokens: 16000,
- },
- },
- max: {
- thinking: {
- type: "enabled",
- budgetTokens: 31999,
- },
- },
- }
- case "@ai-sdk/amazon-bedrock":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/amazon-bedrock
- // For Anthropic models on Bedrock, use reasoningConfig with budgetTokens
- if (model.api.id.includes("anthropic")) {
- return {
- high: {
- reasoningConfig: {
- type: "enabled",
- budgetTokens: 16000,
- },
- },
- max: {
- reasoningConfig: {
- type: "enabled",
- budgetTokens: 31999,
- },
- },
- }
- }
- // For Amazon Nova models, use reasoningConfig with maxReasoningEffort
- return Object.fromEntries(
- WIDELY_SUPPORTED_EFFORTS.map((effort) => [
- effort,
- {
- reasoningConfig: {
- type: "enabled",
- maxReasoningEffort: effort,
- },
- },
- ]),
- )
- case "@ai-sdk/google-vertex":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/google-vertex
- case "@ai-sdk/google":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/google-generative-ai
- if (id.includes("2.5")) {
- return {
- high: {
- thinkingConfig: {
- includeThoughts: true,
- thinkingBudget: 16000,
- },
- },
- max: {
- thinkingConfig: {
- includeThoughts: true,
- thinkingBudget: 24576,
- },
- },
- }
- }
- return Object.fromEntries(
- ["low", "high"].map((effort) => [
- effort,
- {
- includeThoughts: true,
- thinkingLevel: effort,
- },
- ]),
- )
- case "@ai-sdk/mistral":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/mistral
- return {}
- case "@ai-sdk/cohere":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/cohere
- return {}
- case "@ai-sdk/groq":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/groq
- const groqEffort = ["none", ...WIDELY_SUPPORTED_EFFORTS]
- return Object.fromEntries(
- groqEffort.map((effort) => [
- effort,
- {
- includeThoughts: true,
- thinkingLevel: effort,
- },
- ]),
- )
- case "@ai-sdk/perplexity":
- // https://v5.ai-sdk.dev/providers/ai-sdk-providers/perplexity
- return {}
- }
- return {}
- }
- export function options(input: {
- model: Provider.Model
- sessionID: string
- providerOptions?: Record<string, any>
- }): Record<string, any> {
- const result: Record<string, any> = {}
- // openai and providers using openai package should set store to false by default.
- if (input.model.providerID === "openai" || input.model.api.npm === "@ai-sdk/openai") {
- result["store"] = false
- }
- if (input.model.api.npm === "@openrouter/ai-sdk-provider") {
- result["usage"] = {
- include: true,
- }
- if (input.model.api.id.includes("gemini-3")) {
- result["reasoning"] = { effort: "high" }
- }
- }
- if (
- input.model.providerID === "baseten" ||
- (input.model.providerID === "opencode" && ["kimi-k2-thinking", "glm-4.6"].includes(input.model.api.id))
- ) {
- result["chat_template_args"] = { enable_thinking: true }
- }
- if (["zai", "zhipuai"].includes(input.model.providerID) && input.model.api.npm === "@ai-sdk/openai-compatible") {
- result["thinking"] = {
- type: "enabled",
- clear_thinking: false,
- }
- }
- if (input.model.providerID === "openai" || input.providerOptions?.setCacheKey) {
- result["promptCacheKey"] = input.sessionID
- }
- if (input.model.api.npm === "@ai-sdk/google" || input.model.api.npm === "@ai-sdk/google-vertex") {
- result["thinkingConfig"] = {
- includeThoughts: true,
- }
- if (input.model.api.id.includes("gemini-3")) {
- result["thinkingConfig"]["thinkingLevel"] = "high"
- }
- }
- if (input.model.api.id.includes("gpt-5") && !input.model.api.id.includes("gpt-5-chat")) {
- if (input.model.providerID.includes("codex")) {
- result["store"] = false
- }
- if (!input.model.api.id.includes("codex") && !input.model.api.id.includes("gpt-5-pro")) {
- result["reasoningEffort"] = "medium"
- }
- if (input.model.api.id.endsWith("gpt-5.") && input.model.providerID !== "azure") {
- result["textVerbosity"] = "low"
- }
- if (input.model.providerID.startsWith("opencode")) {
- result["promptCacheKey"] = input.sessionID
- result["include"] = ["reasoning.encrypted_content"]
- result["reasoningSummary"] = "auto"
- }
- }
- return result
- }
- export function smallOptions(model: Provider.Model) {
- if (model.providerID === "openai" || model.api.id.includes("gpt-5")) {
- if (model.api.id.includes("5.")) {
- return { reasoningEffort: "low" }
- }
- return { reasoningEffort: "minimal" }
- }
- if (model.providerID === "google") {
- // gemini-3 uses thinkingLevel, gemini-2.5 uses thinkingBudget
- if (model.api.id.includes("gemini-3")) {
- return { thinkingConfig: { thinkingLevel: "minimal" } }
- }
- return { thinkingConfig: { thinkingBudget: 0 } }
- }
- if (model.providerID === "openrouter") {
- if (model.api.id.includes("google")) {
- return { reasoning: { enabled: false } }
- }
- return { reasoningEffort: "minimal" }
- }
- return {}
- }
- export function providerOptions(model: Provider.Model, options: { [x: string]: any }) {
- const key = sdkKey(model.api.npm) ?? model.providerID
- return { [key]: options }
- }
- export function maxOutputTokens(
- npm: string,
- options: Record<string, any>,
- modelLimit: number,
- globalLimit: number,
- ): number {
- const modelCap = modelLimit || globalLimit
- const standardLimit = Math.min(modelCap, globalLimit)
- if (npm === "@ai-sdk/anthropic") {
- const thinking = options?.["thinking"]
- const budgetTokens = typeof thinking?.["budgetTokens"] === "number" ? thinking["budgetTokens"] : 0
- const enabled = thinking?.["type"] === "enabled"
- if (enabled && budgetTokens > 0) {
- // Return text tokens so that text + thinking <= model cap, preferring 32k text when possible.
- if (budgetTokens + standardLimit <= modelCap) {
- return standardLimit
- }
- return modelCap - budgetTokens
- }
- }
- return standardLimit
- }
- export function schema(model: Provider.Model, schema: JSONSchema.BaseSchema) {
- /*
- if (["openai", "azure"].includes(providerID)) {
- if (schema.type === "object" && schema.properties) {
- for (const [key, value] of Object.entries(schema.properties)) {
- if (schema.required?.includes(key)) continue
- schema.properties[key] = {
- anyOf: [
- value as JSONSchema.JSONSchema,
- {
- type: "null",
- },
- ],
- }
- }
- }
- }
- */
- // Convert integer enums to string enums for Google/Gemini
- if (model.providerID === "google" || model.api.id.includes("gemini")) {
- const sanitizeGemini = (obj: any): any => {
- if (obj === null || typeof obj !== "object") {
- return obj
- }
- if (Array.isArray(obj)) {
- return obj.map(sanitizeGemini)
- }
- const result: any = {}
- for (const [key, value] of Object.entries(obj)) {
- if (key === "enum" && Array.isArray(value)) {
- // Convert all enum values to strings
- result[key] = value.map((v) => String(v))
- // If we have integer type with enum, change type to string
- if (result.type === "integer" || result.type === "number") {
- result.type = "string"
- }
- } else if (typeof value === "object" && value !== null) {
- result[key] = sanitizeGemini(value)
- } else {
- result[key] = value
- }
- }
- // Filter required array to only include fields that exist in properties
- if (result.type === "object" && result.properties && Array.isArray(result.required)) {
- result.required = result.required.filter((field: any) => field in result.properties)
- }
- if (result.type === "array" && result.items == null) {
- result.items = {}
- }
- return result
- }
- schema = sanitizeGemini(schema)
- }
- return schema
- }
- export function error(providerID: string, error: APICallError) {
- let message = error.message
- if (providerID.includes("github-copilot") && error.statusCode === 403) {
- return "Please reauthenticate with the copilot provider to ensure your credentials work properly with OpenCode."
- }
- if (providerID.includes("github-copilot") && message.includes("The requested model is not supported")) {
- return (
- message +
- "\n\nMake sure the model is enabled in your copilot settings: https://github.com/settings/copilot/features"
- )
- }
- return message
- }
- }
|