| 12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007 |
- package service
- import (
- "encoding/json"
- "fmt"
- "strings"
- "github.com/QuantumNous/new-api/common"
- "github.com/QuantumNous/new-api/constant"
- "github.com/QuantumNous/new-api/dto"
- "github.com/QuantumNous/new-api/relay/channel/openrouter"
- relaycommon "github.com/QuantumNous/new-api/relay/common"
- "github.com/QuantumNous/new-api/relay/reasonmap"
- "github.com/samber/lo"
- )
- func ClaudeToOpenAIRequest(claudeRequest dto.ClaudeRequest, info *relaycommon.RelayInfo) (*dto.GeneralOpenAIRequest, error) {
- openAIRequest := dto.GeneralOpenAIRequest{
- Model: claudeRequest.Model,
- Temperature: claudeRequest.Temperature,
- }
- if claudeRequest.MaxTokens != nil {
- openAIRequest.MaxTokens = lo.ToPtr(lo.FromPtr(claudeRequest.MaxTokens))
- }
- if claudeRequest.TopP != nil {
- openAIRequest.TopP = lo.ToPtr(lo.FromPtr(claudeRequest.TopP))
- }
- if claudeRequest.TopK != nil {
- openAIRequest.TopK = lo.ToPtr(lo.FromPtr(claudeRequest.TopK))
- }
- if claudeRequest.Stream != nil {
- openAIRequest.Stream = lo.ToPtr(lo.FromPtr(claudeRequest.Stream))
- }
- isOpenRouter := info.ChannelType == constant.ChannelTypeOpenRouter
- if isOpenRouter {
- if effort := claudeRequest.GetEfforts(); effort != "" {
- effortBytes, _ := json.Marshal(effort)
- openAIRequest.Verbosity = effortBytes
- }
- if claudeRequest.Thinking != nil {
- var reasoning openrouter.RequestReasoning
- if claudeRequest.Thinking.Type == "enabled" {
- reasoning = openrouter.RequestReasoning{
- Enabled: true,
- MaxTokens: claudeRequest.Thinking.GetBudgetTokens(),
- }
- } else if claudeRequest.Thinking.Type == "adaptive" {
- reasoning = openrouter.RequestReasoning{
- Enabled: true,
- }
- }
- reasoningJSON, err := json.Marshal(reasoning)
- if err != nil {
- return nil, fmt.Errorf("failed to marshal reasoning: %w", err)
- }
- openAIRequest.Reasoning = reasoningJSON
- }
- } else {
- thinkingSuffix := "-thinking"
- if strings.HasSuffix(info.OriginModelName, thinkingSuffix) &&
- !strings.HasSuffix(openAIRequest.Model, thinkingSuffix) {
- openAIRequest.Model = openAIRequest.Model + thinkingSuffix
- }
- }
- // Convert stop sequences
- if len(claudeRequest.StopSequences) == 1 {
- openAIRequest.Stop = claudeRequest.StopSequences[0]
- } else if len(claudeRequest.StopSequences) > 1 {
- openAIRequest.Stop = claudeRequest.StopSequences
- }
- // Convert tools
- tools, _ := common.Any2Type[[]dto.Tool](claudeRequest.Tools)
- openAITools := make([]dto.ToolCallRequest, 0)
- for _, claudeTool := range tools {
- openAITool := dto.ToolCallRequest{
- Type: "function",
- Function: dto.FunctionRequest{
- Name: claudeTool.Name,
- Description: claudeTool.Description,
- Parameters: claudeTool.InputSchema,
- },
- }
- openAITools = append(openAITools, openAITool)
- }
- openAIRequest.Tools = openAITools
- // Convert messages
- openAIMessages := make([]dto.Message, 0)
- // Add system message if present
- if claudeRequest.System != nil {
- if claudeRequest.IsStringSystem() && claudeRequest.GetStringSystem() != "" {
- openAIMessage := dto.Message{
- Role: "system",
- }
- openAIMessage.SetStringContent(claudeRequest.GetStringSystem())
- openAIMessages = append(openAIMessages, openAIMessage)
- } else {
- systems := claudeRequest.ParseSystem()
- if len(systems) > 0 {
- openAIMessage := dto.Message{
- Role: "system",
- }
- isOpenRouterClaude := isOpenRouter && strings.HasPrefix(info.UpstreamModelName, "anthropic/claude")
- if isOpenRouterClaude {
- systemMediaMessages := make([]dto.MediaContent, 0, len(systems))
- for _, system := range systems {
- message := dto.MediaContent{
- Type: "text",
- Text: system.GetText(),
- CacheControl: system.CacheControl,
- }
- systemMediaMessages = append(systemMediaMessages, message)
- }
- openAIMessage.SetMediaContent(systemMediaMessages)
- } else {
- systemStr := ""
- for _, system := range systems {
- if system.Text != nil {
- systemStr += *system.Text
- }
- }
- openAIMessage.SetStringContent(systemStr)
- }
- openAIMessages = append(openAIMessages, openAIMessage)
- }
- }
- }
- for _, claudeMessage := range claudeRequest.Messages {
- openAIMessage := dto.Message{
- Role: claudeMessage.Role,
- }
- //log.Printf("claudeMessage.Content: %v", claudeMessage.Content)
- if claudeMessage.IsStringContent() {
- openAIMessage.SetStringContent(claudeMessage.GetStringContent())
- } else {
- content, err := claudeMessage.ParseContent()
- if err != nil {
- return nil, err
- }
- contents := content
- var toolCalls []dto.ToolCallRequest
- mediaMessages := make([]dto.MediaContent, 0, len(contents))
- for _, mediaMsg := range contents {
- switch mediaMsg.Type {
- case "text", "input_text":
- message := dto.MediaContent{
- Type: "text",
- Text: mediaMsg.GetText(),
- CacheControl: mediaMsg.CacheControl,
- }
- mediaMessages = append(mediaMessages, message)
- case "image":
- // Handle image conversion (base64 to URL or keep as is)
- imageData := fmt.Sprintf("data:%s;base64,%s", mediaMsg.Source.MediaType, mediaMsg.Source.Data)
- //textContent += fmt.Sprintf("[Image: %s]", imageData)
- mediaMessage := dto.MediaContent{
- Type: "image_url",
- ImageUrl: &dto.MessageImageUrl{Url: imageData},
- }
- mediaMessages = append(mediaMessages, mediaMessage)
- case "tool_use":
- toolCall := dto.ToolCallRequest{
- ID: mediaMsg.Id,
- Type: "function",
- Function: dto.FunctionRequest{
- Name: mediaMsg.Name,
- Arguments: toJSONString(mediaMsg.Input),
- },
- }
- toolCalls = append(toolCalls, toolCall)
- case "tool_result":
- // Add tool result as a separate message
- toolName := mediaMsg.Name
- if toolName == "" {
- toolName = claudeRequest.SearchToolNameByToolCallId(mediaMsg.ToolUseId)
- }
- oaiToolMessage := dto.Message{
- Role: "tool",
- Name: &toolName,
- ToolCallId: mediaMsg.ToolUseId,
- }
- //oaiToolMessage.SetStringContent(*mediaMsg.GetMediaContent().Text)
- if mediaMsg.IsStringContent() {
- oaiToolMessage.SetStringContent(mediaMsg.GetStringContent())
- } else {
- mediaContents := mediaMsg.ParseMediaContent()
- encodeJson, _ := common.Marshal(mediaContents)
- oaiToolMessage.SetStringContent(string(encodeJson))
- }
- openAIMessages = append(openAIMessages, oaiToolMessage)
- }
- }
- if len(toolCalls) > 0 {
- openAIMessage.SetToolCalls(toolCalls)
- }
- if len(mediaMessages) > 0 && len(toolCalls) == 0 {
- openAIMessage.SetMediaContent(mediaMessages)
- }
- }
- if len(openAIMessage.ParseContent()) > 0 || len(openAIMessage.ToolCalls) > 0 {
- openAIMessages = append(openAIMessages, openAIMessage)
- }
- }
- openAIRequest.Messages = openAIMessages
- return &openAIRequest, nil
- }
- func generateStopBlock(index int) *dto.ClaudeResponse {
- return &dto.ClaudeResponse{
- Type: "content_block_stop",
- Index: common.GetPointer[int](index),
- }
- }
- func buildClaudeUsageFromOpenAIUsage(oaiUsage *dto.Usage) *dto.ClaudeUsage {
- if oaiUsage == nil {
- return nil
- }
- cacheCreation5m, cacheCreation1h := NormalizeCacheCreationSplit(
- oaiUsage.PromptTokensDetails.CachedCreationTokens,
- oaiUsage.ClaudeCacheCreation5mTokens,
- oaiUsage.ClaudeCacheCreation1hTokens,
- )
- usage := &dto.ClaudeUsage{
- InputTokens: oaiUsage.PromptTokens,
- OutputTokens: oaiUsage.CompletionTokens,
- CacheCreationInputTokens: oaiUsage.PromptTokensDetails.CachedCreationTokens,
- CacheReadInputTokens: oaiUsage.PromptTokensDetails.CachedTokens,
- }
- if cacheCreation5m > 0 || cacheCreation1h > 0 {
- usage.CacheCreation = &dto.ClaudeCacheCreationUsage{
- Ephemeral5mInputTokens: cacheCreation5m,
- Ephemeral1hInputTokens: cacheCreation1h,
- }
- }
- return usage
- }
- func NormalizeCacheCreationSplit(totalTokens int, tokens5m int, tokens1h int) (int, int) {
- remainder := lo.Max([]int{totalTokens - tokens5m - tokens1h, 0})
- return tokens5m + remainder, tokens1h
- }
- func StreamResponseOpenAI2Claude(openAIResponse *dto.ChatCompletionsStreamResponse, info *relaycommon.RelayInfo) []*dto.ClaudeResponse {
- if info.ClaudeConvertInfo.Done {
- return nil
- }
- var claudeResponses []*dto.ClaudeResponse
- // stopOpenBlocks emits the required content_block_stop event(s) for the currently open block(s)
- // according to Anthropic's SSE streaming state machine:
- // content_block_start -> content_block_delta* -> content_block_stop (per index).
- //
- // For text/thinking, there is at most one open block at info.ClaudeConvertInfo.Index.
- // For tools, OpenAI tool_calls can stream multiple parallel tool_use blocks (indexed from 0),
- // so we may have multiple open blocks and must stop each one explicitly.
- stopOpenBlocks := func() {
- switch info.ClaudeConvertInfo.LastMessagesType {
- case relaycommon.LastMessageTypeText, relaycommon.LastMessageTypeThinking:
- claudeResponses = append(claudeResponses, generateStopBlock(info.ClaudeConvertInfo.Index))
- case relaycommon.LastMessageTypeTools:
- base := info.ClaudeConvertInfo.ToolCallBaseIndex
- for offset := 0; offset <= info.ClaudeConvertInfo.ToolCallMaxIndexOffset; offset++ {
- claudeResponses = append(claudeResponses, generateStopBlock(base+offset))
- }
- }
- }
- // stopOpenBlocksAndAdvance closes the currently open block(s) and advances the content block index
- // to the next available slot for subsequent content_block_start events.
- //
- // This prevents invalid streams where a content_block_delta (e.g. thinking_delta) is emitted for an
- // index whose active content_block type is different (the typical cause of "Mismatched content block type").
- stopOpenBlocksAndAdvance := func() {
- if info.ClaudeConvertInfo.LastMessagesType == relaycommon.LastMessageTypeNone {
- return
- }
- stopOpenBlocks()
- switch info.ClaudeConvertInfo.LastMessagesType {
- case relaycommon.LastMessageTypeTools:
- info.ClaudeConvertInfo.Index = info.ClaudeConvertInfo.ToolCallBaseIndex + info.ClaudeConvertInfo.ToolCallMaxIndexOffset + 1
- info.ClaudeConvertInfo.ToolCallBaseIndex = 0
- info.ClaudeConvertInfo.ToolCallMaxIndexOffset = 0
- default:
- info.ClaudeConvertInfo.Index++
- }
- info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeNone
- }
- if info.SendResponseCount == 1 {
- msg := &dto.ClaudeMediaMessage{
- Id: openAIResponse.Id,
- Model: openAIResponse.Model,
- Type: "message",
- Role: "assistant",
- Usage: &dto.ClaudeUsage{
- InputTokens: info.GetEstimatePromptTokens(),
- OutputTokens: 0,
- },
- }
- msg.SetContent(make([]any, 0))
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Type: "message_start",
- Message: msg,
- })
- //claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- // Type: "ping",
- //})
- if openAIResponse.IsToolCall() {
- info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeTools
- info.ClaudeConvertInfo.ToolCallBaseIndex = 0
- info.ClaudeConvertInfo.ToolCallMaxIndexOffset = 0
- var toolCall dto.ToolCallResponse
- if len(openAIResponse.Choices) > 0 && len(openAIResponse.Choices[0].Delta.ToolCalls) > 0 {
- toolCall = openAIResponse.Choices[0].Delta.ToolCalls[0]
- } else {
- first := openAIResponse.GetFirstToolCall()
- if first != nil {
- toolCall = *first
- } else {
- toolCall = dto.ToolCallResponse{}
- }
- }
- resp := &dto.ClaudeResponse{
- Type: "content_block_start",
- ContentBlock: &dto.ClaudeMediaMessage{
- Id: toolCall.ID,
- Type: "tool_use",
- Name: toolCall.Function.Name,
- Input: map[string]interface{}{},
- },
- }
- resp.SetIndex(0)
- claudeResponses = append(claudeResponses, resp)
- // 首块包含工具 delta,则追加 input_json_delta
- if toolCall.Function.Arguments != "" {
- idx := 0
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Index: &idx,
- Type: "content_block_delta",
- Delta: &dto.ClaudeMediaMessage{
- Type: "input_json_delta",
- PartialJson: &toolCall.Function.Arguments,
- },
- })
- }
- } else {
- }
- // 判断首个响应是否存在内容(非标准的 OpenAI 响应)
- if len(openAIResponse.Choices) > 0 {
- reasoning := openAIResponse.Choices[0].Delta.GetReasoningContent()
- content := openAIResponse.Choices[0].Delta.GetContentString()
- if reasoning != "" {
- if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeThinking {
- stopOpenBlocksAndAdvance()
- }
- idx := info.ClaudeConvertInfo.Index
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Index: &idx,
- Type: "content_block_start",
- ContentBlock: &dto.ClaudeMediaMessage{
- Type: "thinking",
- Thinking: common.GetPointer[string](""),
- },
- })
- idx2 := idx
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Index: &idx2,
- Type: "content_block_delta",
- Delta: &dto.ClaudeMediaMessage{
- Type: "thinking_delta",
- Thinking: &reasoning,
- },
- })
- info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeThinking
- } else if content != "" {
- if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeText {
- stopOpenBlocksAndAdvance()
- }
- idx := info.ClaudeConvertInfo.Index
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Index: &idx,
- Type: "content_block_start",
- ContentBlock: &dto.ClaudeMediaMessage{
- Type: "text",
- Text: common.GetPointer[string](""),
- },
- })
- idx2 := idx
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Index: &idx2,
- Type: "content_block_delta",
- Delta: &dto.ClaudeMediaMessage{
- Type: "text_delta",
- Text: common.GetPointer[string](content),
- },
- })
- info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeText
- }
- }
- // 如果首块就带 finish_reason,需要立即发送停止块
- if len(openAIResponse.Choices) > 0 && openAIResponse.Choices[0].FinishReason != nil && *openAIResponse.Choices[0].FinishReason != "" {
- info.FinishReason = *openAIResponse.Choices[0].FinishReason
- stopOpenBlocks()
- oaiUsage := openAIResponse.Usage
- if oaiUsage == nil {
- oaiUsage = info.ClaudeConvertInfo.Usage
- }
- if oaiUsage != nil {
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Type: "message_delta",
- Usage: buildClaudeUsageFromOpenAIUsage(oaiUsage),
- Delta: &dto.ClaudeMediaMessage{
- StopReason: common.GetPointer[string](stopReasonOpenAI2Claude(info.FinishReason)),
- },
- })
- }
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Type: "message_stop",
- })
- info.ClaudeConvertInfo.Done = true
- }
- return claudeResponses
- }
- if len(openAIResponse.Choices) == 0 {
- // Some OpenAI-compatible upstreams end with a usage-only SSE chunk.
- oaiUsage := openAIResponse.Usage
- if oaiUsage == nil {
- oaiUsage = info.ClaudeConvertInfo.Usage
- }
- if oaiUsage != nil {
- stopOpenBlocks()
- stopReason := stopReasonOpenAI2Claude(info.FinishReason)
- if stopReason == "" {
- stopReason = "end_turn"
- }
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Type: "message_delta",
- Usage: buildClaudeUsageFromOpenAIUsage(oaiUsage),
- Delta: &dto.ClaudeMediaMessage{
- StopReason: common.GetPointer[string](stopReason),
- },
- })
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Type: "message_stop",
- })
- info.ClaudeConvertInfo.Done = true
- }
- return claudeResponses
- } else {
- chosenChoice := openAIResponse.Choices[0]
- doneChunk := chosenChoice.FinishReason != nil && *chosenChoice.FinishReason != ""
- if doneChunk {
- info.FinishReason = *chosenChoice.FinishReason
- oaiUsage := openAIResponse.Usage
- if oaiUsage == nil {
- oaiUsage = info.ClaudeConvertInfo.Usage
- // Some upstreams emit finish_reason first, then send a final usage-only chunk.
- // Defer closing until usage is available so the final message_delta carries it.
- return claudeResponses
- }
- }
- var claudeResponse dto.ClaudeResponse
- var isEmpty bool
- claudeResponse.Type = "content_block_delta"
- if len(chosenChoice.Delta.ToolCalls) > 0 {
- toolCalls := chosenChoice.Delta.ToolCalls
- if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeTools {
- stopOpenBlocksAndAdvance()
- info.ClaudeConvertInfo.ToolCallBaseIndex = info.ClaudeConvertInfo.Index
- info.ClaudeConvertInfo.ToolCallMaxIndexOffset = 0
- }
- info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeTools
- base := info.ClaudeConvertInfo.ToolCallBaseIndex
- maxOffset := info.ClaudeConvertInfo.ToolCallMaxIndexOffset
- for i, toolCall := range toolCalls {
- offset := 0
- if toolCall.Index != nil {
- offset = *toolCall.Index
- } else {
- offset = i
- }
- if offset > maxOffset {
- maxOffset = offset
- }
- blockIndex := base + offset
- idx := blockIndex
- if toolCall.Function.Name != "" {
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Index: &idx,
- Type: "content_block_start",
- ContentBlock: &dto.ClaudeMediaMessage{
- Id: toolCall.ID,
- Type: "tool_use",
- Name: toolCall.Function.Name,
- Input: map[string]interface{}{},
- },
- })
- }
- if len(toolCall.Function.Arguments) > 0 {
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Index: &idx,
- Type: "content_block_delta",
- Delta: &dto.ClaudeMediaMessage{
- Type: "input_json_delta",
- PartialJson: &toolCall.Function.Arguments,
- },
- })
- }
- }
- info.ClaudeConvertInfo.ToolCallMaxIndexOffset = maxOffset
- info.ClaudeConvertInfo.Index = base + maxOffset
- } else {
- reasoning := chosenChoice.Delta.GetReasoningContent()
- textContent := chosenChoice.Delta.GetContentString()
- if reasoning != "" || textContent != "" {
- if reasoning != "" {
- if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeThinking {
- stopOpenBlocksAndAdvance()
- idx := info.ClaudeConvertInfo.Index
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Index: &idx,
- Type: "content_block_start",
- ContentBlock: &dto.ClaudeMediaMessage{
- Type: "thinking",
- Thinking: common.GetPointer[string](""),
- },
- })
- }
- info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeThinking
- claudeResponse.Delta = &dto.ClaudeMediaMessage{
- Type: "thinking_delta",
- Thinking: &reasoning,
- }
- } else {
- if info.ClaudeConvertInfo.LastMessagesType != relaycommon.LastMessageTypeText {
- stopOpenBlocksAndAdvance()
- idx := info.ClaudeConvertInfo.Index
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Index: &idx,
- Type: "content_block_start",
- ContentBlock: &dto.ClaudeMediaMessage{
- Type: "text",
- Text: common.GetPointer[string](""),
- },
- })
- }
- info.ClaudeConvertInfo.LastMessagesType = relaycommon.LastMessageTypeText
- claudeResponse.Delta = &dto.ClaudeMediaMessage{
- Type: "text_delta",
- Text: common.GetPointer[string](textContent),
- }
- }
- } else {
- isEmpty = true
- }
- }
- claudeResponse.Index = common.GetPointer[int](info.ClaudeConvertInfo.Index)
- if !isEmpty && claudeResponse.Delta != nil {
- claudeResponses = append(claudeResponses, &claudeResponse)
- }
- if doneChunk || info.ClaudeConvertInfo.Done {
- stopOpenBlocks()
- oaiUsage := openAIResponse.Usage
- if oaiUsage == nil {
- oaiUsage = info.ClaudeConvertInfo.Usage
- }
- if oaiUsage != nil {
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Type: "message_delta",
- Usage: buildClaudeUsageFromOpenAIUsage(oaiUsage),
- Delta: &dto.ClaudeMediaMessage{
- StopReason: common.GetPointer[string](stopReasonOpenAI2Claude(info.FinishReason)),
- },
- })
- }
- claudeResponses = append(claudeResponses, &dto.ClaudeResponse{
- Type: "message_stop",
- })
- info.ClaudeConvertInfo.Done = true
- return claudeResponses
- }
- }
- return claudeResponses
- }
- func ResponseOpenAI2Claude(openAIResponse *dto.OpenAITextResponse, info *relaycommon.RelayInfo) *dto.ClaudeResponse {
- var stopReason string
- contents := make([]dto.ClaudeMediaMessage, 0)
- claudeResponse := &dto.ClaudeResponse{
- Id: openAIResponse.Id,
- Type: "message",
- Role: "assistant",
- Model: openAIResponse.Model,
- }
- for _, choice := range openAIResponse.Choices {
- stopReason = stopReasonOpenAI2Claude(choice.FinishReason)
- if choice.FinishReason == "tool_calls" {
- for _, toolUse := range choice.Message.ParseToolCalls() {
- claudeContent := dto.ClaudeMediaMessage{}
- claudeContent.Type = "tool_use"
- claudeContent.Id = toolUse.ID
- claudeContent.Name = toolUse.Function.Name
- var mapParams map[string]interface{}
- if err := common.Unmarshal([]byte(toolUse.Function.Arguments), &mapParams); err == nil {
- claudeContent.Input = mapParams
- } else {
- claudeContent.Input = toolUse.Function.Arguments
- }
- contents = append(contents, claudeContent)
- }
- } else {
- claudeContent := dto.ClaudeMediaMessage{}
- claudeContent.Type = "text"
- claudeContent.SetText(choice.Message.StringContent())
- contents = append(contents, claudeContent)
- }
- }
- claudeResponse.Content = contents
- claudeResponse.StopReason = stopReason
- claudeResponse.Usage = buildClaudeUsageFromOpenAIUsage(&openAIResponse.Usage)
- return claudeResponse
- }
- func stopReasonOpenAI2Claude(reason string) string {
- return reasonmap.OpenAIFinishReasonToClaudeStopReason(reason)
- }
- func toJSONString(v interface{}) string {
- b, err := json.Marshal(v)
- if err != nil {
- return "{}"
- }
- return string(b)
- }
- func GeminiToOpenAIRequest(geminiRequest *dto.GeminiChatRequest, info *relaycommon.RelayInfo) (*dto.GeneralOpenAIRequest, error) {
- openaiRequest := &dto.GeneralOpenAIRequest{
- Model: info.UpstreamModelName,
- Stream: lo.ToPtr(info.IsStream),
- }
- // 转换 messages
- var messages []dto.Message
- for _, content := range geminiRequest.Contents {
- message := dto.Message{
- Role: convertGeminiRoleToOpenAI(content.Role),
- }
- // 处理 parts
- var mediaContents []dto.MediaContent
- var toolCalls []dto.ToolCallRequest
- for _, part := range content.Parts {
- if part.Text != "" {
- mediaContent := dto.MediaContent{
- Type: "text",
- Text: part.Text,
- }
- mediaContents = append(mediaContents, mediaContent)
- } else if part.InlineData != nil {
- mediaContent := dto.MediaContent{
- Type: "image_url",
- ImageUrl: &dto.MessageImageUrl{
- Url: fmt.Sprintf("data:%s;base64,%s", part.InlineData.MimeType, part.InlineData.Data),
- Detail: "auto",
- MimeType: part.InlineData.MimeType,
- },
- }
- mediaContents = append(mediaContents, mediaContent)
- } else if part.FileData != nil {
- mediaContent := dto.MediaContent{
- Type: "image_url",
- ImageUrl: &dto.MessageImageUrl{
- Url: part.FileData.FileUri,
- Detail: "auto",
- MimeType: part.FileData.MimeType,
- },
- }
- mediaContents = append(mediaContents, mediaContent)
- } else if part.FunctionCall != nil {
- // 处理 Gemini 的工具调用
- toolCall := dto.ToolCallRequest{
- ID: fmt.Sprintf("call_%d", len(toolCalls)+1), // 生成唯一ID
- Type: "function",
- Function: dto.FunctionRequest{
- Name: part.FunctionCall.FunctionName,
- Arguments: toJSONString(part.FunctionCall.Arguments),
- },
- }
- toolCalls = append(toolCalls, toolCall)
- } else if part.FunctionResponse != nil {
- // 处理 Gemini 的工具响应,创建单独的 tool 消息
- toolMessage := dto.Message{
- Role: "tool",
- ToolCallId: fmt.Sprintf("call_%d", len(toolCalls)), // 使用对应的调用ID
- }
- toolMessage.SetStringContent(toJSONString(part.FunctionResponse.Response))
- messages = append(messages, toolMessage)
- }
- }
- // 设置消息内容
- if len(toolCalls) > 0 {
- // 如果有工具调用,设置工具调用
- message.SetToolCalls(toolCalls)
- } else if len(mediaContents) == 1 && mediaContents[0].Type == "text" {
- // 如果只有一个文本内容,直接设置字符串
- message.Content = mediaContents[0].Text
- } else if len(mediaContents) > 0 {
- // 如果有多个内容或包含媒体,设置为数组
- message.SetMediaContent(mediaContents)
- }
- // 只有当消息有内容或工具调用时才添加
- if len(message.ParseContent()) > 0 || len(message.ToolCalls) > 0 {
- messages = append(messages, message)
- }
- }
- openaiRequest.Messages = messages
- if geminiRequest.GenerationConfig.Temperature != nil {
- openaiRequest.Temperature = geminiRequest.GenerationConfig.Temperature
- }
- if geminiRequest.GenerationConfig.TopP != nil && *geminiRequest.GenerationConfig.TopP > 0 {
- openaiRequest.TopP = lo.ToPtr(*geminiRequest.GenerationConfig.TopP)
- }
- if geminiRequest.GenerationConfig.TopK != nil && *geminiRequest.GenerationConfig.TopK > 0 {
- openaiRequest.TopK = lo.ToPtr(int(*geminiRequest.GenerationConfig.TopK))
- }
- if geminiRequest.GenerationConfig.MaxOutputTokens != nil && *geminiRequest.GenerationConfig.MaxOutputTokens > 0 {
- openaiRequest.MaxTokens = lo.ToPtr(*geminiRequest.GenerationConfig.MaxOutputTokens)
- }
- // gemini stop sequences 最多 5 个,openai stop 最多 4 个
- if len(geminiRequest.GenerationConfig.StopSequences) > 0 {
- openaiRequest.Stop = geminiRequest.GenerationConfig.StopSequences[:4]
- }
- if geminiRequest.GenerationConfig.CandidateCount != nil && *geminiRequest.GenerationConfig.CandidateCount > 0 {
- openaiRequest.N = lo.ToPtr(*geminiRequest.GenerationConfig.CandidateCount)
- }
- // 转换工具调用
- if len(geminiRequest.GetTools()) > 0 {
- var tools []dto.ToolCallRequest
- for _, tool := range geminiRequest.GetTools() {
- if tool.FunctionDeclarations != nil {
- functionDeclarations, err := common.Any2Type[[]dto.FunctionRequest](tool.FunctionDeclarations)
- if err != nil {
- common.SysError(fmt.Sprintf("failed to parse gemini function declarations: %v (type=%T)", err, tool.FunctionDeclarations))
- continue
- }
- for _, function := range functionDeclarations {
- openAITool := dto.ToolCallRequest{
- Type: "function",
- Function: dto.FunctionRequest{
- Name: function.Name,
- Description: function.Description,
- Parameters: function.Parameters,
- },
- }
- tools = append(tools, openAITool)
- }
- }
- }
- if len(tools) > 0 {
- openaiRequest.Tools = tools
- }
- }
- // gemini system instructions
- if geminiRequest.SystemInstructions != nil {
- // 将系统指令作为第一条消息插入
- systemMessage := dto.Message{
- Role: "system",
- Content: extractTextFromGeminiParts(geminiRequest.SystemInstructions.Parts),
- }
- openaiRequest.Messages = append([]dto.Message{systemMessage}, openaiRequest.Messages...)
- }
- return openaiRequest, nil
- }
- func convertGeminiRoleToOpenAI(geminiRole string) string {
- switch geminiRole {
- case "user":
- return "user"
- case "model":
- return "assistant"
- case "function":
- return "function"
- default:
- return "user"
- }
- }
- func extractTextFromGeminiParts(parts []dto.GeminiPart) string {
- var texts []string
- for _, part := range parts {
- if part.Text != "" {
- texts = append(texts, part.Text)
- }
- }
- return strings.Join(texts, "\n")
- }
- // ResponseOpenAI2Gemini 将 OpenAI 响应转换为 Gemini 格式
- func ResponseOpenAI2Gemini(openAIResponse *dto.OpenAITextResponse, info *relaycommon.RelayInfo) *dto.GeminiChatResponse {
- geminiResponse := &dto.GeminiChatResponse{
- Candidates: make([]dto.GeminiChatCandidate, 0, len(openAIResponse.Choices)),
- UsageMetadata: dto.GeminiUsageMetadata{
- PromptTokenCount: openAIResponse.PromptTokens,
- CandidatesTokenCount: openAIResponse.CompletionTokens,
- TotalTokenCount: openAIResponse.PromptTokens + openAIResponse.CompletionTokens,
- },
- }
- for _, choice := range openAIResponse.Choices {
- candidate := dto.GeminiChatCandidate{
- Index: int64(choice.Index),
- SafetyRatings: []dto.GeminiChatSafetyRating{},
- }
- // 设置结束原因
- var finishReason string
- switch choice.FinishReason {
- case "stop":
- finishReason = "STOP"
- case "length":
- finishReason = "MAX_TOKENS"
- case "content_filter":
- finishReason = "SAFETY"
- case "tool_calls":
- finishReason = "STOP"
- default:
- finishReason = "STOP"
- }
- candidate.FinishReason = &finishReason
- // 转换消息内容
- content := dto.GeminiChatContent{
- Role: "model",
- Parts: make([]dto.GeminiPart, 0),
- }
- // 处理工具调用
- toolCalls := choice.Message.ParseToolCalls()
- if len(toolCalls) > 0 {
- for _, toolCall := range toolCalls {
- // 解析参数
- var args map[string]interface{}
- if toolCall.Function.Arguments != "" {
- if err := json.Unmarshal([]byte(toolCall.Function.Arguments), &args); err != nil {
- args = map[string]interface{}{"arguments": toolCall.Function.Arguments}
- }
- } else {
- args = make(map[string]interface{})
- }
- part := dto.GeminiPart{
- FunctionCall: &dto.FunctionCall{
- FunctionName: toolCall.Function.Name,
- Arguments: args,
- },
- }
- content.Parts = append(content.Parts, part)
- }
- } else {
- // 处理文本内容
- textContent := choice.Message.StringContent()
- if textContent != "" {
- part := dto.GeminiPart{
- Text: textContent,
- }
- content.Parts = append(content.Parts, part)
- }
- }
- candidate.Content = content
- geminiResponse.Candidates = append(geminiResponse.Candidates, candidate)
- }
- return geminiResponse
- }
- // StreamResponseOpenAI2Gemini 将 OpenAI 流式响应转换为 Gemini 格式
- func StreamResponseOpenAI2Gemini(openAIResponse *dto.ChatCompletionsStreamResponse, info *relaycommon.RelayInfo) *dto.GeminiChatResponse {
- // 检查是否有实际内容或结束标志
- hasContent := false
- hasFinishReason := false
- for _, choice := range openAIResponse.Choices {
- if len(choice.Delta.GetContentString()) > 0 || (choice.Delta.ToolCalls != nil && len(choice.Delta.ToolCalls) > 0) {
- hasContent = true
- }
- if choice.FinishReason != nil {
- hasFinishReason = true
- }
- }
- // 如果没有实际内容且没有结束标志,跳过。主要针对 openai 流响应开头的空数据
- if !hasContent && !hasFinishReason {
- return nil
- }
- geminiResponse := &dto.GeminiChatResponse{
- Candidates: make([]dto.GeminiChatCandidate, 0, len(openAIResponse.Choices)),
- UsageMetadata: dto.GeminiUsageMetadata{
- PromptTokenCount: info.GetEstimatePromptTokens(),
- CandidatesTokenCount: 0, // 流式响应中可能没有完整的 usage 信息
- TotalTokenCount: info.GetEstimatePromptTokens(),
- },
- }
- if openAIResponse.Usage != nil {
- geminiResponse.UsageMetadata.PromptTokenCount = openAIResponse.Usage.PromptTokens
- geminiResponse.UsageMetadata.CandidatesTokenCount = openAIResponse.Usage.CompletionTokens
- geminiResponse.UsageMetadata.TotalTokenCount = openAIResponse.Usage.TotalTokens
- }
- for _, choice := range openAIResponse.Choices {
- candidate := dto.GeminiChatCandidate{
- Index: int64(choice.Index),
- SafetyRatings: []dto.GeminiChatSafetyRating{},
- }
- // 设置结束原因
- if choice.FinishReason != nil {
- var finishReason string
- switch *choice.FinishReason {
- case "stop":
- finishReason = "STOP"
- case "length":
- finishReason = "MAX_TOKENS"
- case "content_filter":
- finishReason = "SAFETY"
- case "tool_calls":
- finishReason = "STOP"
- default:
- finishReason = "STOP"
- }
- candidate.FinishReason = &finishReason
- }
- // 转换消息内容
- content := dto.GeminiChatContent{
- Role: "model",
- Parts: make([]dto.GeminiPart, 0),
- }
- // 处理工具调用
- if choice.Delta.ToolCalls != nil {
- for _, toolCall := range choice.Delta.ToolCalls {
- // 解析参数
- var args map[string]interface{}
- if toolCall.Function.Arguments != "" {
- if err := json.Unmarshal([]byte(toolCall.Function.Arguments), &args); err != nil {
- args = map[string]interface{}{"arguments": toolCall.Function.Arguments}
- }
- } else {
- args = make(map[string]interface{})
- }
- part := dto.GeminiPart{
- FunctionCall: &dto.FunctionCall{
- FunctionName: toolCall.Function.Name,
- Arguments: args,
- },
- }
- content.Parts = append(content.Parts, part)
- }
- } else {
- // 处理文本内容
- textContent := choice.Delta.GetContentString()
- if textContent != "" {
- part := dto.GeminiPart{
- Text: textContent,
- }
- content.Parts = append(content.Parts, part)
- }
- }
- candidate.Content = content
- geminiResponse.Candidates = append(geminiResponse.Candidates, candidate)
- }
- return geminiResponse
- }
|