adaptor.go 7.9 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262
  1. package vertex
  2. import (
  3. "encoding/json"
  4. "errors"
  5. "fmt"
  6. "io"
  7. "net/http"
  8. "one-api/dto"
  9. "one-api/relay/channel"
  10. "one-api/relay/channel/claude"
  11. "one-api/relay/channel/gemini"
  12. "one-api/relay/channel/openai"
  13. relaycommon "one-api/relay/common"
  14. "one-api/relay/constant"
  15. "one-api/setting/model_setting"
  16. "one-api/types"
  17. "strings"
  18. "github.com/gin-gonic/gin"
  19. )
  20. const (
  21. RequestModeClaude = 1
  22. RequestModeGemini = 2
  23. RequestModeLlama = 3
  24. )
  25. var claudeModelMap = map[string]string{
  26. "claude-3-sonnet-20240229": "claude-3-sonnet@20240229",
  27. "claude-3-opus-20240229": "claude-3-opus@20240229",
  28. "claude-3-haiku-20240307": "claude-3-haiku@20240307",
  29. "claude-3-5-sonnet-20240620": "claude-3-5-sonnet@20240620",
  30. "claude-3-5-sonnet-20241022": "claude-3-5-sonnet-v2@20241022",
  31. "claude-3-7-sonnet-20250219": "claude-3-7-sonnet@20250219",
  32. "claude-sonnet-4-20250514": "claude-sonnet-4@20250514",
  33. "claude-opus-4-20250514": "claude-opus-4@20250514",
  34. }
  35. const anthropicVersion = "vertex-2023-10-16"
  36. type Adaptor struct {
  37. RequestMode int
  38. AccountCredentials Credentials
  39. }
  40. func (a *Adaptor) ConvertClaudeRequest(c *gin.Context, info *relaycommon.RelayInfo, request *dto.ClaudeRequest) (any, error) {
  41. if v, ok := claudeModelMap[info.UpstreamModelName]; ok {
  42. c.Set("request_model", v)
  43. } else {
  44. c.Set("request_model", request.Model)
  45. }
  46. vertexClaudeReq := copyRequest(request, anthropicVersion)
  47. return vertexClaudeReq, nil
  48. }
  49. func (a *Adaptor) ConvertAudioRequest(c *gin.Context, info *relaycommon.RelayInfo, request dto.AudioRequest) (io.Reader, error) {
  50. //TODO implement me
  51. return nil, errors.New("not implemented")
  52. }
  53. func (a *Adaptor) ConvertImageRequest(c *gin.Context, info *relaycommon.RelayInfo, request dto.ImageRequest) (any, error) {
  54. //TODO implement me
  55. return nil, errors.New("not implemented")
  56. }
  57. func (a *Adaptor) Init(info *relaycommon.RelayInfo) {
  58. if strings.HasPrefix(info.UpstreamModelName, "claude") {
  59. a.RequestMode = RequestModeClaude
  60. } else if strings.HasPrefix(info.UpstreamModelName, "gemini") {
  61. a.RequestMode = RequestModeGemini
  62. } else if strings.Contains(info.UpstreamModelName, "llama") {
  63. a.RequestMode = RequestModeLlama
  64. }
  65. }
  66. func (a *Adaptor) GetRequestURL(info *relaycommon.RelayInfo) (string, error) {
  67. adc := &Credentials{}
  68. if err := json.Unmarshal([]byte(info.ApiKey), adc); err != nil {
  69. return "", fmt.Errorf("failed to decode credentials file: %w", err)
  70. }
  71. region := GetModelRegion(info.ApiVersion, info.OriginModelName)
  72. a.AccountCredentials = *adc
  73. suffix := ""
  74. if a.RequestMode == RequestModeGemini {
  75. if model_setting.GetGeminiSettings().ThinkingAdapterEnabled {
  76. // 新增逻辑:处理 -thinking-<budget> 格式
  77. if strings.Contains(info.UpstreamModelName, "-thinking-") {
  78. parts := strings.Split(info.UpstreamModelName, "-thinking-")
  79. info.UpstreamModelName = parts[0]
  80. } else if strings.HasSuffix(info.UpstreamModelName, "-thinking") { // 旧的适配
  81. info.UpstreamModelName = strings.TrimSuffix(info.UpstreamModelName, "-thinking")
  82. } else if strings.HasSuffix(info.UpstreamModelName, "-nothinking") {
  83. info.UpstreamModelName = strings.TrimSuffix(info.UpstreamModelName, "-nothinking")
  84. }
  85. }
  86. if info.IsStream {
  87. suffix = "streamGenerateContent?alt=sse"
  88. } else {
  89. suffix = "generateContent"
  90. }
  91. if region == "global" {
  92. return fmt.Sprintf(
  93. "https://aiplatform.googleapis.com/v1/projects/%s/locations/global/publishers/google/models/%s:%s",
  94. adc.ProjectID,
  95. info.UpstreamModelName,
  96. suffix,
  97. ), nil
  98. } else {
  99. return fmt.Sprintf(
  100. "https://%s-aiplatform.googleapis.com/v1/projects/%s/locations/%s/publishers/google/models/%s:%s",
  101. region,
  102. adc.ProjectID,
  103. region,
  104. info.UpstreamModelName,
  105. suffix,
  106. ), nil
  107. }
  108. } else if a.RequestMode == RequestModeClaude {
  109. if info.IsStream {
  110. suffix = "streamRawPredict?alt=sse"
  111. } else {
  112. suffix = "rawPredict"
  113. }
  114. model := info.UpstreamModelName
  115. if v, ok := claudeModelMap[info.UpstreamModelName]; ok {
  116. model = v
  117. }
  118. if region == "global" {
  119. return fmt.Sprintf(
  120. "https://aiplatform.googleapis.com/v1/projects/%s/locations/global/publishers/anthropic/models/%s:%s",
  121. adc.ProjectID,
  122. model,
  123. suffix,
  124. ), nil
  125. } else {
  126. return fmt.Sprintf(
  127. "https://%s-aiplatform.googleapis.com/v1/projects/%s/locations/%s/publishers/anthropic/models/%s:%s",
  128. region,
  129. adc.ProjectID,
  130. region,
  131. model,
  132. suffix,
  133. ), nil
  134. }
  135. } else if a.RequestMode == RequestModeLlama {
  136. return fmt.Sprintf(
  137. "https://%s-aiplatform.googleapis.com/v1beta1/projects/%s/locations/%s/endpoints/openapi/chat/completions",
  138. region,
  139. adc.ProjectID,
  140. region,
  141. ), nil
  142. }
  143. return "", errors.New("unsupported request mode")
  144. }
  145. func (a *Adaptor) SetupRequestHeader(c *gin.Context, req *http.Header, info *relaycommon.RelayInfo) error {
  146. channel.SetupApiRequestHeader(info, c, req)
  147. accessToken, err := getAccessToken(a, info)
  148. if err != nil {
  149. return err
  150. }
  151. req.Set("Authorization", "Bearer "+accessToken)
  152. return nil
  153. }
  154. func (a *Adaptor) ConvertOpenAIRequest(c *gin.Context, info *relaycommon.RelayInfo, request *dto.GeneralOpenAIRequest) (any, error) {
  155. if request == nil {
  156. return nil, errors.New("request is nil")
  157. }
  158. if a.RequestMode == RequestModeClaude {
  159. claudeReq, err := claude.RequestOpenAI2ClaudeMessage(*request)
  160. if err != nil {
  161. return nil, err
  162. }
  163. vertexClaudeReq := copyRequest(claudeReq, anthropicVersion)
  164. c.Set("request_model", claudeReq.Model)
  165. info.UpstreamModelName = claudeReq.Model
  166. return vertexClaudeReq, nil
  167. } else if a.RequestMode == RequestModeGemini {
  168. geminiRequest, err := gemini.CovertGemini2OpenAI(*request, info)
  169. if err != nil {
  170. return nil, err
  171. }
  172. c.Set("request_model", request.Model)
  173. return geminiRequest, nil
  174. } else if a.RequestMode == RequestModeLlama {
  175. return request, nil
  176. }
  177. return nil, errors.New("unsupported request mode")
  178. }
  179. func (a *Adaptor) ConvertRerankRequest(c *gin.Context, relayMode int, request dto.RerankRequest) (any, error) {
  180. return nil, nil
  181. }
  182. func (a *Adaptor) ConvertEmbeddingRequest(c *gin.Context, info *relaycommon.RelayInfo, request dto.EmbeddingRequest) (any, error) {
  183. //TODO implement me
  184. return nil, errors.New("not implemented")
  185. }
  186. func (a *Adaptor) ConvertOpenAIResponsesRequest(c *gin.Context, info *relaycommon.RelayInfo, request dto.OpenAIResponsesRequest) (any, error) {
  187. // TODO implement me
  188. return nil, errors.New("not implemented")
  189. }
  190. func (a *Adaptor) DoRequest(c *gin.Context, info *relaycommon.RelayInfo, requestBody io.Reader) (any, error) {
  191. return channel.DoApiRequest(a, c, info, requestBody)
  192. }
  193. func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, info *relaycommon.RelayInfo) (usage any, err *types.NewAPIError) {
  194. if info.IsStream {
  195. switch a.RequestMode {
  196. case RequestModeClaude:
  197. err, usage = claude.ClaudeStreamHandler(c, resp, info, claude.RequestModeMessage)
  198. case RequestModeGemini:
  199. if info.RelayMode == constant.RelayModeGemini {
  200. usage, err = gemini.GeminiTextGenerationStreamHandler(c, info, resp)
  201. } else {
  202. usage, err = gemini.GeminiChatStreamHandler(c, info, resp)
  203. }
  204. case RequestModeLlama:
  205. usage, err = openai.OaiStreamHandler(c, info, resp)
  206. }
  207. } else {
  208. switch a.RequestMode {
  209. case RequestModeClaude:
  210. err, usage = claude.ClaudeHandler(c, resp, claude.RequestModeMessage, info)
  211. case RequestModeGemini:
  212. if info.RelayMode == constant.RelayModeGemini {
  213. usage, err = gemini.GeminiTextGenerationHandler(c, info, resp)
  214. } else {
  215. usage, err = gemini.GeminiChatHandler(c, info, resp)
  216. }
  217. case RequestModeLlama:
  218. usage, err = openai.OpenaiHandler(c, info, resp)
  219. }
  220. }
  221. return
  222. }
  223. func (a *Adaptor) GetModelList() []string {
  224. var modelList []string
  225. for i, s := range ModelList {
  226. modelList = append(modelList, s)
  227. ModelList[i] = s
  228. }
  229. for i, s := range claude.ModelList {
  230. modelList = append(modelList, s)
  231. claude.ModelList[i] = s
  232. }
  233. for i, s := range gemini.ModelList {
  234. modelList = append(modelList, s)
  235. gemini.ModelList[i] = s
  236. }
  237. return modelList
  238. }
  239. func (a *Adaptor) GetChannelName() string {
  240. return ChannelName
  241. }