claude_handler.go 5.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161
  1. package relay
  2. import (
  3. "bytes"
  4. "encoding/json"
  5. "errors"
  6. "fmt"
  7. "github.com/gin-gonic/gin"
  8. "io"
  9. "net/http"
  10. "one-api/common"
  11. "one-api/dto"
  12. relaycommon "one-api/relay/common"
  13. "one-api/relay/helper"
  14. "one-api/service"
  15. "one-api/setting/model_setting"
  16. "strings"
  17. )
  18. func getAndValidateClaudeRequest(c *gin.Context) (textRequest *dto.ClaudeRequest, err error) {
  19. textRequest = &dto.ClaudeRequest{}
  20. err = c.ShouldBindJSON(textRequest)
  21. if err != nil {
  22. return nil, err
  23. }
  24. if textRequest.Messages == nil || len(textRequest.Messages) == 0 {
  25. return nil, errors.New("field messages is required")
  26. }
  27. if textRequest.Model == "" {
  28. return nil, errors.New("field model is required")
  29. }
  30. return textRequest, nil
  31. }
  32. func ClaudeHelper(c *gin.Context) (claudeError *dto.ClaudeErrorWithStatusCode) {
  33. relayInfo := relaycommon.GenRelayInfoClaude(c)
  34. // get & validate textRequest 获取并验证文本请求
  35. textRequest, err := getAndValidateClaudeRequest(c)
  36. if err != nil {
  37. return service.ClaudeErrorWrapperLocal(err, "invalid_claude_request", http.StatusBadRequest)
  38. }
  39. if textRequest.Stream {
  40. relayInfo.IsStream = true
  41. }
  42. err = helper.ModelMappedHelper(c, relayInfo, textRequest)
  43. if err != nil {
  44. return service.ClaudeErrorWrapperLocal(err, "model_mapped_error", http.StatusInternalServerError)
  45. }
  46. promptTokens, err := getClaudePromptTokens(textRequest, relayInfo)
  47. // count messages token error 计算promptTokens错误
  48. if err != nil {
  49. return service.ClaudeErrorWrapperLocal(err, "count_token_messages_failed", http.StatusInternalServerError)
  50. }
  51. priceData, err := helper.ModelPriceHelper(c, relayInfo, promptTokens, int(textRequest.MaxTokens))
  52. if err != nil {
  53. return service.ClaudeErrorWrapperLocal(err, "model_price_error", http.StatusInternalServerError)
  54. }
  55. // pre-consume quota 预消耗配额
  56. preConsumedQuota, userQuota, openaiErr := preConsumeQuota(c, priceData.ShouldPreConsumedQuota, relayInfo)
  57. if openaiErr != nil {
  58. return service.OpenAIErrorToClaudeError(openaiErr)
  59. }
  60. defer func() {
  61. if openaiErr != nil {
  62. returnPreConsumedQuota(c, relayInfo, userQuota, preConsumedQuota)
  63. }
  64. }()
  65. adaptor := GetAdaptor(relayInfo.ApiType)
  66. if adaptor == nil {
  67. return service.ClaudeErrorWrapperLocal(fmt.Errorf("invalid api type: %d", relayInfo.ApiType), "invalid_api_type", http.StatusBadRequest)
  68. }
  69. adaptor.Init(relayInfo)
  70. var requestBody io.Reader
  71. if textRequest.MaxTokens == 0 {
  72. textRequest.MaxTokens = uint(model_setting.GetClaudeSettings().GetDefaultMaxTokens(textRequest.Model))
  73. }
  74. if model_setting.GetClaudeSettings().ThinkingAdapterEnabled &&
  75. strings.HasSuffix(textRequest.Model, "-thinking") {
  76. if textRequest.Thinking == nil {
  77. // 因为BudgetTokens 必须大于1024
  78. if textRequest.MaxTokens < 1280 {
  79. textRequest.MaxTokens = 1280
  80. }
  81. // BudgetTokens 为 max_tokens 的 80%
  82. textRequest.Thinking = &dto.Thinking{
  83. Type: "enabled",
  84. BudgetTokens: common.GetPointer[int](int(float64(textRequest.MaxTokens) * model_setting.GetClaudeSettings().ThinkingAdapterBudgetTokensPercentage)),
  85. }
  86. // TODO: 临时处理
  87. // https://docs.anthropic.com/en/docs/build-with-claude/extended-thinking#important-considerations-when-using-extended-thinking
  88. textRequest.TopP = 0
  89. textRequest.Temperature = common.GetPointer[float64](1.0)
  90. }
  91. textRequest.Model = strings.TrimSuffix(textRequest.Model, "-thinking")
  92. relayInfo.UpstreamModelName = textRequest.Model
  93. }
  94. convertedRequest, err := adaptor.ConvertClaudeRequest(c, relayInfo, textRequest)
  95. if err != nil {
  96. return service.ClaudeErrorWrapperLocal(err, "convert_request_failed", http.StatusInternalServerError)
  97. }
  98. jsonData, err := json.Marshal(convertedRequest)
  99. if common.DebugEnabled {
  100. println("requestBody: ", string(jsonData))
  101. }
  102. if err != nil {
  103. return service.ClaudeErrorWrapperLocal(err, "json_marshal_failed", http.StatusInternalServerError)
  104. }
  105. requestBody = bytes.NewBuffer(jsonData)
  106. statusCodeMappingStr := c.GetString("status_code_mapping")
  107. var httpResp *http.Response
  108. resp, err := adaptor.DoRequest(c, relayInfo, requestBody)
  109. if err != nil {
  110. return service.ClaudeErrorWrapper(err, "do_request_failed", http.StatusInternalServerError)
  111. }
  112. if resp != nil {
  113. httpResp = resp.(*http.Response)
  114. relayInfo.IsStream = relayInfo.IsStream || strings.HasPrefix(httpResp.Header.Get("Content-Type"), "text/event-stream")
  115. if httpResp.StatusCode != http.StatusOK {
  116. openaiErr = service.RelayErrorHandler(httpResp, false)
  117. // reset status code 重置状态码
  118. service.ResetStatusCode(openaiErr, statusCodeMappingStr)
  119. return service.OpenAIErrorToClaudeError(openaiErr)
  120. }
  121. }
  122. usage, openaiErr := adaptor.DoResponse(c, httpResp, relayInfo)
  123. //log.Printf("usage: %v", usage)
  124. if openaiErr != nil {
  125. // reset status code 重置状态码
  126. service.ResetStatusCode(openaiErr, statusCodeMappingStr)
  127. return service.OpenAIErrorToClaudeError(openaiErr)
  128. }
  129. service.PostClaudeConsumeQuota(c, relayInfo, usage.(*dto.Usage), preConsumedQuota, userQuota, priceData, "")
  130. return nil
  131. }
  132. func getClaudePromptTokens(textRequest *dto.ClaudeRequest, info *relaycommon.RelayInfo) (int, error) {
  133. var promptTokens int
  134. var err error
  135. switch info.RelayMode {
  136. default:
  137. promptTokens, err = service.CountTokenClaudeRequest(*textRequest, info.UpstreamModelName)
  138. }
  139. info.PromptTokens = promptTokens
  140. return promptTokens, err
  141. }