|
|
@@ -0,0 +1,241 @@
|
|
|
+import { Anthropic } from "@anthropic-ai/sdk"
|
|
|
+import OpenAI from "openai"
|
|
|
+
|
|
|
+type ContentPartText = OpenAI.Chat.ChatCompletionContentPartText
|
|
|
+type ContentPartImage = OpenAI.Chat.ChatCompletionContentPartImage
|
|
|
+type UserMessage = OpenAI.Chat.ChatCompletionUserMessageParam
|
|
|
+type AssistantMessage = OpenAI.Chat.ChatCompletionAssistantMessageParam
|
|
|
+type SystemMessage = OpenAI.Chat.ChatCompletionSystemMessageParam
|
|
|
+type ToolMessage = OpenAI.Chat.ChatCompletionToolMessageParam
|
|
|
+type Message = OpenAI.Chat.ChatCompletionMessageParam
|
|
|
+type AnthropicMessage = Anthropic.Messages.MessageParam
|
|
|
+
|
|
|
+/**
|
|
|
+ * Extended assistant message type to support Z.ai's interleaved thinking.
|
|
|
+ * Z.ai's API returns reasoning_content alongside content and tool_calls,
|
|
|
+ * and requires it to be passed back in subsequent requests for preserved thinking.
|
|
|
+ */
|
|
|
+export type ZAiAssistantMessage = AssistantMessage & {
|
|
|
+ reasoning_content?: string
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * Converts Anthropic messages to OpenAI format optimized for Z.ai's GLM-4.7 thinking mode.
|
|
|
+ *
|
|
|
+ * Key differences from standard OpenAI format:
|
|
|
+ * - Preserves reasoning_content on assistant messages for interleaved thinking
|
|
|
+ * - Text content after tool_results (like environment_details) is converted to system messages
|
|
|
+ * instead of user messages, preventing reasoning_content from being dropped
|
|
|
+ *
|
|
|
+ * @param messages Array of Anthropic messages
|
|
|
+ * @param options Optional configuration for message conversion
|
|
|
+ * @param options.convertToolResultTextToSystem If true, convert text content after tool_results
|
|
|
+ * to system messages instead of user messages.
|
|
|
+ * This preserves reasoning_content continuity.
|
|
|
+ * @returns Array of OpenAI messages optimized for Z.ai's thinking mode
|
|
|
+ */
|
|
|
+export function convertToZAiFormat(
|
|
|
+ messages: AnthropicMessage[],
|
|
|
+ options?: { convertToolResultTextToSystem?: boolean },
|
|
|
+): Message[] {
|
|
|
+ const result: Message[] = []
|
|
|
+
|
|
|
+ for (const message of messages) {
|
|
|
+ // Check if the message has reasoning_content (for Z.ai interleaved thinking)
|
|
|
+ const messageWithReasoning = message as AnthropicMessage & { reasoning_content?: string }
|
|
|
+ const reasoningContent = messageWithReasoning.reasoning_content
|
|
|
+
|
|
|
+ if (message.role === "user") {
|
|
|
+ // Handle user messages - may contain tool_result blocks
|
|
|
+ if (Array.isArray(message.content)) {
|
|
|
+ const textParts: string[] = []
|
|
|
+ const imageParts: ContentPartImage[] = []
|
|
|
+ const toolResults: { tool_use_id: string; content: string }[] = []
|
|
|
+
|
|
|
+ for (const part of message.content) {
|
|
|
+ if (part.type === "text") {
|
|
|
+ textParts.push(part.text)
|
|
|
+ } else if (part.type === "image") {
|
|
|
+ imageParts.push({
|
|
|
+ type: "image_url",
|
|
|
+ image_url: { url: `data:${part.source.media_type};base64,${part.source.data}` },
|
|
|
+ })
|
|
|
+ } else if (part.type === "tool_result") {
|
|
|
+ // Convert tool_result to OpenAI tool message format
|
|
|
+ let content: string
|
|
|
+ if (typeof part.content === "string") {
|
|
|
+ content = part.content
|
|
|
+ } else if (Array.isArray(part.content)) {
|
|
|
+ content =
|
|
|
+ part.content
|
|
|
+ ?.map((c) => {
|
|
|
+ if (c.type === "text") return c.text
|
|
|
+ if (c.type === "image") return "(image)"
|
|
|
+ return ""
|
|
|
+ })
|
|
|
+ .join("\n") ?? ""
|
|
|
+ } else {
|
|
|
+ content = ""
|
|
|
+ }
|
|
|
+ toolResults.push({
|
|
|
+ tool_use_id: part.tool_use_id,
|
|
|
+ content,
|
|
|
+ })
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ // Add tool messages first (they must follow assistant tool_use)
|
|
|
+ for (const toolResult of toolResults) {
|
|
|
+ const toolMessage: ToolMessage = {
|
|
|
+ role: "tool",
|
|
|
+ tool_call_id: toolResult.tool_use_id,
|
|
|
+ content: toolResult.content,
|
|
|
+ }
|
|
|
+ result.push(toolMessage)
|
|
|
+ }
|
|
|
+
|
|
|
+ // Handle text/image content after tool results
|
|
|
+ if (textParts.length > 0 || imageParts.length > 0) {
|
|
|
+ // For Z.ai interleaved thinking: when convertToolResultTextToSystem is enabled and we have
|
|
|
+ // tool results followed by text (like environment_details), convert to system message
|
|
|
+ // instead of user message to avoid dropping reasoning_content.
|
|
|
+ const shouldConvertToSystem =
|
|
|
+ options?.convertToolResultTextToSystem && toolResults.length > 0 && imageParts.length === 0
|
|
|
+
|
|
|
+ if (shouldConvertToSystem) {
|
|
|
+ // Convert text content to system message
|
|
|
+ const systemMessage: SystemMessage = {
|
|
|
+ role: "system",
|
|
|
+ content: textParts.join("\n"),
|
|
|
+ }
|
|
|
+ result.push(systemMessage)
|
|
|
+ } else {
|
|
|
+ // Standard behavior: add user message with text/image content
|
|
|
+ let content: UserMessage["content"]
|
|
|
+ if (imageParts.length > 0) {
|
|
|
+ const parts: (ContentPartText | ContentPartImage)[] = []
|
|
|
+ if (textParts.length > 0) {
|
|
|
+ parts.push({ type: "text", text: textParts.join("\n") })
|
|
|
+ }
|
|
|
+ parts.push(...imageParts)
|
|
|
+ content = parts
|
|
|
+ } else {
|
|
|
+ content = textParts.join("\n")
|
|
|
+ }
|
|
|
+
|
|
|
+ // Check if we can merge with the last message
|
|
|
+ const lastMessage = result[result.length - 1]
|
|
|
+ if (lastMessage?.role === "user") {
|
|
|
+ // Merge with existing user message
|
|
|
+ if (typeof lastMessage.content === "string" && typeof content === "string") {
|
|
|
+ lastMessage.content += `\n${content}`
|
|
|
+ } else {
|
|
|
+ const lastContent = Array.isArray(lastMessage.content)
|
|
|
+ ? lastMessage.content
|
|
|
+ : [{ type: "text" as const, text: lastMessage.content || "" }]
|
|
|
+ const newContent = Array.isArray(content)
|
|
|
+ ? content
|
|
|
+ : [{ type: "text" as const, text: content }]
|
|
|
+ lastMessage.content = [...lastContent, ...newContent] as UserMessage["content"]
|
|
|
+ }
|
|
|
+ } else {
|
|
|
+ result.push({ role: "user", content })
|
|
|
+ }
|
|
|
+ }
|
|
|
+ }
|
|
|
+ } else {
|
|
|
+ // Simple string content
|
|
|
+ const lastMessage = result[result.length - 1]
|
|
|
+ if (lastMessage?.role === "user") {
|
|
|
+ if (typeof lastMessage.content === "string") {
|
|
|
+ lastMessage.content += `\n${message.content}`
|
|
|
+ } else {
|
|
|
+ ;(lastMessage.content as (ContentPartText | ContentPartImage)[]).push({
|
|
|
+ type: "text",
|
|
|
+ text: message.content,
|
|
|
+ })
|
|
|
+ }
|
|
|
+ } else {
|
|
|
+ result.push({ role: "user", content: message.content })
|
|
|
+ }
|
|
|
+ }
|
|
|
+ } else if (message.role === "assistant") {
|
|
|
+ // Handle assistant messages - may contain tool_use blocks and reasoning blocks
|
|
|
+ if (Array.isArray(message.content)) {
|
|
|
+ const textParts: string[] = []
|
|
|
+ const toolCalls: OpenAI.Chat.ChatCompletionMessageToolCall[] = []
|
|
|
+ let extractedReasoning: string | undefined
|
|
|
+
|
|
|
+ for (const part of message.content) {
|
|
|
+ if (part.type === "text") {
|
|
|
+ textParts.push(part.text)
|
|
|
+ } else if (part.type === "tool_use") {
|
|
|
+ toolCalls.push({
|
|
|
+ id: part.id,
|
|
|
+ type: "function",
|
|
|
+ function: {
|
|
|
+ name: part.name,
|
|
|
+ arguments: JSON.stringify(part.input),
|
|
|
+ },
|
|
|
+ })
|
|
|
+ } else if ((part as any).type === "reasoning" && (part as any).text) {
|
|
|
+ // Extract reasoning from content blocks (Task stores it this way)
|
|
|
+ extractedReasoning = (part as any).text
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ // Use reasoning from content blocks if not provided at top level
|
|
|
+ const finalReasoning = reasoningContent || extractedReasoning
|
|
|
+
|
|
|
+ const assistantMessage: ZAiAssistantMessage = {
|
|
|
+ role: "assistant",
|
|
|
+ content: textParts.length > 0 ? textParts.join("\n") : null,
|
|
|
+ ...(toolCalls.length > 0 && { tool_calls: toolCalls }),
|
|
|
+ // Preserve reasoning_content for Z.ai interleaved thinking
|
|
|
+ ...(finalReasoning && { reasoning_content: finalReasoning }),
|
|
|
+ }
|
|
|
+
|
|
|
+ // Check if we can merge with the last message (only if no tool calls)
|
|
|
+ const lastMessage = result[result.length - 1]
|
|
|
+ if (lastMessage?.role === "assistant" && !toolCalls.length && !(lastMessage as any).tool_calls) {
|
|
|
+ // Merge text content
|
|
|
+ if (typeof lastMessage.content === "string" && typeof assistantMessage.content === "string") {
|
|
|
+ lastMessage.content += `\n${assistantMessage.content}`
|
|
|
+ } else if (assistantMessage.content) {
|
|
|
+ const lastContent = lastMessage.content || ""
|
|
|
+ lastMessage.content = `${lastContent}\n${assistantMessage.content}`
|
|
|
+ }
|
|
|
+ // Preserve reasoning_content from the new message if present
|
|
|
+ if (finalReasoning) {
|
|
|
+ ;(lastMessage as ZAiAssistantMessage).reasoning_content = finalReasoning
|
|
|
+ }
|
|
|
+ } else {
|
|
|
+ result.push(assistantMessage)
|
|
|
+ }
|
|
|
+ } else {
|
|
|
+ // Simple string content
|
|
|
+ const lastMessage = result[result.length - 1]
|
|
|
+ if (lastMessage?.role === "assistant" && !(lastMessage as any).tool_calls) {
|
|
|
+ if (typeof lastMessage.content === "string") {
|
|
|
+ lastMessage.content += `\n${message.content}`
|
|
|
+ } else {
|
|
|
+ lastMessage.content = message.content
|
|
|
+ }
|
|
|
+ // Preserve reasoning_content from the new message if present
|
|
|
+ if (reasoningContent) {
|
|
|
+ ;(lastMessage as ZAiAssistantMessage).reasoning_content = reasoningContent
|
|
|
+ }
|
|
|
+ } else {
|
|
|
+ const assistantMessage: ZAiAssistantMessage = {
|
|
|
+ role: "assistant",
|
|
|
+ content: message.content,
|
|
|
+ ...(reasoningContent && { reasoning_content: reasoningContent }),
|
|
|
+ }
|
|
|
+ result.push(assistantMessage)
|
|
|
+ }
|
|
|
+ }
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ return result
|
|
|
+}
|