| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363 |
- package middleware
- import (
- "errors"
- "fmt"
- "net/http"
- "strconv"
- "strings"
- "time"
- "github.com/QuantumNous/new-api/common"
- "github.com/QuantumNous/new-api/constant"
- "github.com/QuantumNous/new-api/dto"
- "github.com/QuantumNous/new-api/model"
- relayconstant "github.com/QuantumNous/new-api/relay/constant"
- "github.com/QuantumNous/new-api/service"
- "github.com/QuantumNous/new-api/setting"
- "github.com/QuantumNous/new-api/setting/ratio_setting"
- "github.com/QuantumNous/new-api/types"
- "github.com/gin-gonic/gin"
- )
- type ModelRequest struct {
- Model string `json:"model"`
- Group string `json:"group,omitempty"`
- }
- func Distribute() func(c *gin.Context) {
- return func(c *gin.Context) {
- var channel *model.Channel
- channelId, ok := common.GetContextKey(c, constant.ContextKeyTokenSpecificChannelId)
- modelRequest, shouldSelectChannel, err := getModelRequest(c)
- if err != nil {
- abortWithOpenAiMessage(c, http.StatusBadRequest, "Invalid request, "+err.Error())
- return
- }
- if ok {
- id, err := strconv.Atoi(channelId.(string))
- if err != nil {
- abortWithOpenAiMessage(c, http.StatusBadRequest, "无效的渠道 Id")
- return
- }
- channel, err = model.GetChannelById(id, true)
- if err != nil {
- abortWithOpenAiMessage(c, http.StatusBadRequest, "无效的渠道 Id")
- return
- }
- if channel.Status != common.ChannelStatusEnabled {
- abortWithOpenAiMessage(c, http.StatusForbidden, "该渠道已被禁用")
- return
- }
- } else {
- // Select a channel for the user
- // check token model mapping
- modelLimitEnable := common.GetContextKeyBool(c, constant.ContextKeyTokenModelLimitEnabled)
- if modelLimitEnable {
- s, ok := common.GetContextKey(c, constant.ContextKeyTokenModelLimit)
- if !ok {
- // token model limit is empty, all models are not allowed
- abortWithOpenAiMessage(c, http.StatusForbidden, "该令牌无权访问任何模型")
- return
- }
- var tokenModelLimit map[string]bool
- tokenModelLimit, ok = s.(map[string]bool)
- if !ok {
- tokenModelLimit = map[string]bool{}
- }
- matchName := ratio_setting.FormatMatchingModelName(modelRequest.Model) // match gpts & thinking-*
- if _, ok := tokenModelLimit[matchName]; !ok {
- abortWithOpenAiMessage(c, http.StatusForbidden, "该令牌无权访问模型 "+modelRequest.Model)
- return
- }
- }
- if shouldSelectChannel {
- if modelRequest.Model == "" {
- abortWithOpenAiMessage(c, http.StatusBadRequest, "未指定模型名称,模型名称不能为空")
- return
- }
- var selectGroup string
- userGroup := common.GetContextKeyString(c, constant.ContextKeyUsingGroup)
- // check path is /pg/chat/completions
- if strings.HasPrefix(c.Request.URL.Path, "/pg/chat/completions") {
- playgroundRequest := &dto.PlayGroundRequest{}
- err = common.UnmarshalBodyReusable(c, playgroundRequest)
- if err != nil {
- abortWithOpenAiMessage(c, http.StatusBadRequest, "无效的请求, "+err.Error())
- return
- }
- if playgroundRequest.Group != "" {
- if !setting.GroupInUserUsableGroups(playgroundRequest.Group) && playgroundRequest.Group != userGroup {
- abortWithOpenAiMessage(c, http.StatusForbidden, "无权访问该分组")
- return
- }
- userGroup = playgroundRequest.Group
- }
- }
- channel, selectGroup, err = model.CacheGetRandomSatisfiedChannel(c, userGroup, modelRequest.Model, 0)
- if err != nil {
- showGroup := userGroup
- if userGroup == "auto" {
- showGroup = fmt.Sprintf("auto(%s)", selectGroup)
- }
- message := fmt.Sprintf("获取分组 %s 下模型 %s 的可用渠道失败(数据库一致性已被破坏,distributor): %s", showGroup, modelRequest.Model, err.Error())
- // 如果错误,但是渠道不为空,说明是数据库一致性问题
- //if channel != nil {
- // common.SysError(fmt.Sprintf("渠道不存在:%d", channel.Id))
- // message = "数据库一致性已被破坏,请联系管理员"
- //}
- abortWithOpenAiMessage(c, http.StatusServiceUnavailable, message, string(types.ErrorCodeModelNotFound))
- return
- }
- if channel == nil {
- abortWithOpenAiMessage(c, http.StatusServiceUnavailable, fmt.Sprintf("分组 %s 下模型 %s 无可用渠道(distributor)", userGroup, modelRequest.Model), string(types.ErrorCodeModelNotFound))
- return
- }
- }
- }
- common.SetContextKey(c, constant.ContextKeyRequestStartTime, time.Now())
- SetupContextForSelectedChannel(c, channel, modelRequest.Model)
- c.Next()
- }
- }
- func getModelRequest(c *gin.Context) (*ModelRequest, bool, error) {
- var modelRequest ModelRequest
- shouldSelectChannel := true
- var err error
- if strings.Contains(c.Request.URL.Path, "/mj/") {
- relayMode := relayconstant.Path2RelayModeMidjourney(c.Request.URL.Path)
- if relayMode == relayconstant.RelayModeMidjourneyTaskFetch ||
- relayMode == relayconstant.RelayModeMidjourneyTaskFetchByCondition ||
- relayMode == relayconstant.RelayModeMidjourneyNotify ||
- relayMode == relayconstant.RelayModeMidjourneyTaskImageSeed {
- shouldSelectChannel = false
- } else {
- midjourneyRequest := dto.MidjourneyRequest{}
- err = common.UnmarshalBodyReusable(c, &midjourneyRequest)
- if err != nil {
- return nil, false, err
- }
- midjourneyModel, mjErr, success := service.GetMjRequestModel(relayMode, &midjourneyRequest)
- if mjErr != nil {
- return nil, false, fmt.Errorf(mjErr.Description)
- }
- if midjourneyModel == "" {
- if !success {
- return nil, false, fmt.Errorf("无效的请求, 无法解析模型")
- } else {
- // task fetch, task fetch by condition, notify
- shouldSelectChannel = false
- }
- }
- modelRequest.Model = midjourneyModel
- }
- c.Set("relay_mode", relayMode)
- } else if strings.Contains(c.Request.URL.Path, "/suno/") {
- relayMode := relayconstant.Path2RelaySuno(c.Request.Method, c.Request.URL.Path)
- if relayMode == relayconstant.RelayModeSunoFetch ||
- relayMode == relayconstant.RelayModeSunoFetchByID {
- shouldSelectChannel = false
- } else {
- modelName := service.CoverTaskActionToModelName(constant.TaskPlatformSuno, c.Param("action"))
- modelRequest.Model = modelName
- }
- c.Set("platform", string(constant.TaskPlatformSuno))
- c.Set("relay_mode", relayMode)
- } else if strings.Contains(c.Request.URL.Path, "/v1/videos") {
- //curl https://api.openai.com/v1/videos \
- // -H "Authorization: Bearer $OPENAI_API_KEY" \
- // -F "model=sora-2" \
- // -F "prompt=A calico cat playing a piano on stage"
- // -F input_reference="@image.jpg"
- relayMode := relayconstant.RelayModeUnknown
- if c.Request.Method == http.MethodPost {
- relayMode = relayconstant.RelayModeVideoSubmit
- contentType := c.Request.Header.Get("Content-Type")
- if strings.HasPrefix(contentType, "multipart/form-data") {
- form, err := common.ParseMultipartFormReusable(c)
- if err != nil {
- return nil, false, errors.New("无效的video请求, " + err.Error())
- }
- defer form.RemoveAll()
- if form != nil {
- if values, ok := form.Value["model"]; ok && len(values) > 0 {
- modelRequest.Model = values[0]
- }
- }
- } else if strings.HasPrefix(contentType, "application/json") {
- err = common.UnmarshalBodyReusable(c, &modelRequest)
- if err != nil {
- return nil, false, errors.New("无效的video请求, " + err.Error())
- }
- }
- } else if c.Request.Method == http.MethodGet {
- relayMode = relayconstant.RelayModeVideoFetchByID
- shouldSelectChannel = false
- }
- c.Set("relay_mode", relayMode)
- } else if strings.Contains(c.Request.URL.Path, "/v1/video/generations") {
- relayMode := relayconstant.RelayModeUnknown
- if c.Request.Method == http.MethodPost {
- err = common.UnmarshalBodyReusable(c, &modelRequest)
- if err != nil {
- return nil, false, errors.New("video无效的请求, " + err.Error())
- }
- relayMode = relayconstant.RelayModeVideoSubmit
- } else if c.Request.Method == http.MethodGet {
- relayMode = relayconstant.RelayModeVideoFetchByID
- shouldSelectChannel = false
- }
- if _, ok := c.Get("relay_mode"); !ok {
- c.Set("relay_mode", relayMode)
- }
- } else if strings.HasPrefix(c.Request.URL.Path, "/v1beta/models/") || strings.HasPrefix(c.Request.URL.Path, "/v1/models/") {
- // Gemini API 路径处理: /v1beta/models/gemini-2.0-flash:generateContent
- relayMode := relayconstant.RelayModeGemini
- modelName := extractModelNameFromGeminiPath(c.Request.URL.Path)
- if modelName != "" {
- modelRequest.Model = modelName
- }
- c.Set("relay_mode", relayMode)
- } else if !strings.HasPrefix(c.Request.URL.Path, "/v1/audio/transcriptions") && !strings.Contains(c.Request.Header.Get("Content-Type"), "multipart/form-data") {
- err = common.UnmarshalBodyReusable(c, &modelRequest)
- }
- if err != nil {
- return nil, false, errors.New("无效的请求, " + err.Error())
- }
- if strings.HasPrefix(c.Request.URL.Path, "/v1/realtime") {
- //wss://api.openai.com/v1/realtime?model=gpt-4o-realtime-preview-2024-10-01
- modelRequest.Model = c.Query("model")
- }
- if strings.HasPrefix(c.Request.URL.Path, "/v1/moderations") {
- if modelRequest.Model == "" {
- modelRequest.Model = "text-moderation-stable"
- }
- }
- if strings.HasSuffix(c.Request.URL.Path, "embeddings") {
- if modelRequest.Model == "" {
- modelRequest.Model = c.Param("model")
- }
- }
- if strings.HasPrefix(c.Request.URL.Path, "/v1/images/generations") {
- modelRequest.Model = common.GetStringIfEmpty(modelRequest.Model, "dall-e")
- } else if strings.HasPrefix(c.Request.URL.Path, "/v1/images/edits") {
- //modelRequest.Model = common.GetStringIfEmpty(c.PostForm("model"), "gpt-image-1")
- if strings.Contains(c.Request.Header.Get("Content-Type"), "multipart/form-data") {
- modelRequest.Model = c.PostForm("model")
- }
- }
- if strings.HasPrefix(c.Request.URL.Path, "/v1/audio") {
- relayMode := relayconstant.RelayModeAudioSpeech
- if strings.HasPrefix(c.Request.URL.Path, "/v1/audio/speech") {
- modelRequest.Model = common.GetStringIfEmpty(modelRequest.Model, "tts-1")
- } else if strings.HasPrefix(c.Request.URL.Path, "/v1/audio/translations") {
- modelRequest.Model = common.GetStringIfEmpty(modelRequest.Model, c.PostForm("model"))
- modelRequest.Model = common.GetStringIfEmpty(modelRequest.Model, "whisper-1")
- relayMode = relayconstant.RelayModeAudioTranslation
- } else if strings.HasPrefix(c.Request.URL.Path, "/v1/audio/transcriptions") {
- modelRequest.Model = common.GetStringIfEmpty(modelRequest.Model, c.PostForm("model"))
- modelRequest.Model = common.GetStringIfEmpty(modelRequest.Model, "whisper-1")
- relayMode = relayconstant.RelayModeAudioTranscription
- }
- c.Set("relay_mode", relayMode)
- }
- if strings.HasPrefix(c.Request.URL.Path, "/pg/chat/completions") {
- // playground chat completions
- err = common.UnmarshalBodyReusable(c, &modelRequest)
- if err != nil {
- return nil, false, errors.New("无效的请求, " + err.Error())
- }
- common.SetContextKey(c, constant.ContextKeyTokenGroup, modelRequest.Group)
- }
- return &modelRequest, shouldSelectChannel, nil
- }
- func SetupContextForSelectedChannel(c *gin.Context, channel *model.Channel, modelName string) *types.NewAPIError {
- c.Set("original_model", modelName) // for retry
- if channel == nil {
- return types.NewError(errors.New("channel is nil"), types.ErrorCodeGetChannelFailed, types.ErrOptionWithSkipRetry())
- }
- common.SetContextKey(c, constant.ContextKeyChannelId, channel.Id)
- common.SetContextKey(c, constant.ContextKeyChannelName, channel.Name)
- common.SetContextKey(c, constant.ContextKeyChannelType, channel.Type)
- common.SetContextKey(c, constant.ContextKeyChannelCreateTime, channel.CreatedTime)
- common.SetContextKey(c, constant.ContextKeyChannelSetting, channel.GetSetting())
- common.SetContextKey(c, constant.ContextKeyChannelOtherSetting, channel.GetOtherSettings())
- common.SetContextKey(c, constant.ContextKeyChannelParamOverride, channel.GetParamOverride())
- common.SetContextKey(c, constant.ContextKeyChannelHeaderOverride, channel.GetHeaderOverride())
- if nil != channel.OpenAIOrganization && *channel.OpenAIOrganization != "" {
- common.SetContextKey(c, constant.ContextKeyChannelOrganization, *channel.OpenAIOrganization)
- }
- common.SetContextKey(c, constant.ContextKeyChannelAutoBan, channel.GetAutoBan())
- common.SetContextKey(c, constant.ContextKeyChannelModelMapping, channel.GetModelMapping())
- common.SetContextKey(c, constant.ContextKeyChannelStatusCodeMapping, channel.GetStatusCodeMapping())
- key, index, newAPIError := channel.GetNextEnabledKey()
- if newAPIError != nil {
- return newAPIError
- }
- if channel.ChannelInfo.IsMultiKey {
- common.SetContextKey(c, constant.ContextKeyChannelIsMultiKey, true)
- common.SetContextKey(c, constant.ContextKeyChannelMultiKeyIndex, index)
- } else {
- // 必须设置为 false,否则在重试到单个 key 的时候会导致日志显示错误
- common.SetContextKey(c, constant.ContextKeyChannelIsMultiKey, false)
- }
- // c.Request.Header.Set("Authorization", fmt.Sprintf("Bearer %s", key))
- common.SetContextKey(c, constant.ContextKeyChannelKey, key)
- common.SetContextKey(c, constant.ContextKeyChannelBaseUrl, channel.GetBaseURL())
- common.SetContextKey(c, constant.ContextKeySystemPromptOverride, false)
- // TODO: api_version统一
- switch channel.Type {
- case constant.ChannelTypeAzure:
- c.Set("api_version", channel.Other)
- case constant.ChannelTypeVertexAi:
- c.Set("region", channel.Other)
- case constant.ChannelTypeXunfei:
- c.Set("api_version", channel.Other)
- case constant.ChannelTypeGemini:
- c.Set("api_version", channel.Other)
- case constant.ChannelTypeAli:
- c.Set("plugin", channel.Other)
- case constant.ChannelCloudflare:
- c.Set("api_version", channel.Other)
- case constant.ChannelTypeMokaAI:
- c.Set("api_version", channel.Other)
- case constant.ChannelTypeCoze:
- c.Set("bot_id", channel.Other)
- }
- return nil
- }
- // extractModelNameFromGeminiPath 从 Gemini API URL 路径中提取模型名
- // 输入格式: /v1beta/models/gemini-2.0-flash:generateContent
- // 输出: gemini-2.0-flash
- func extractModelNameFromGeminiPath(path string) string {
- // 查找 "/models/" 的位置
- modelsPrefix := "/models/"
- modelsIndex := strings.Index(path, modelsPrefix)
- if modelsIndex == -1 {
- return ""
- }
- // 从 "/models/" 之后开始提取
- startIndex := modelsIndex + len(modelsPrefix)
- if startIndex >= len(path) {
- return ""
- }
- // 查找 ":" 的位置,模型名在 ":" 之前
- colonIndex := strings.Index(path[startIndex:], ":")
- if colonIndex == -1 {
- // 如果没有找到 ":",返回从 "/models/" 到路径结尾的部分
- return path[startIndex:]
- }
- // 返回模型名部分
- return path[startIndex : startIndex+colonIndex]
- }
|