| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888 |
- package controller
- import (
- "bytes"
- "encoding/json"
- "errors"
- "fmt"
- "io"
- "math"
- "net/http"
- "net/http/httptest"
- "net/url"
- "strconv"
- "strings"
- "sync"
- "time"
- "github.com/QuantumNous/new-api/common"
- "github.com/QuantumNous/new-api/constant"
- "github.com/QuantumNous/new-api/dto"
- "github.com/QuantumNous/new-api/middleware"
- "github.com/QuantumNous/new-api/model"
- "github.com/QuantumNous/new-api/relay"
- relaycommon "github.com/QuantumNous/new-api/relay/common"
- relayconstant "github.com/QuantumNous/new-api/relay/constant"
- "github.com/QuantumNous/new-api/relay/helper"
- "github.com/QuantumNous/new-api/service"
- "github.com/QuantumNous/new-api/setting/operation_setting"
- "github.com/QuantumNous/new-api/setting/ratio_setting"
- "github.com/QuantumNous/new-api/types"
- "github.com/bytedance/gopkg/util/gopool"
- "github.com/samber/lo"
- "github.com/tidwall/gjson"
- "github.com/gin-gonic/gin"
- )
- type testResult struct {
- context *gin.Context
- localErr error
- newAPIError *types.NewAPIError
- }
- func normalizeChannelTestEndpoint(channel *model.Channel, modelName, endpointType string) string {
- normalized := strings.TrimSpace(endpointType)
- if normalized != "" {
- return normalized
- }
- if strings.HasSuffix(modelName, ratio_setting.CompactModelSuffix) {
- return string(constant.EndpointTypeOpenAIResponseCompact)
- }
- if channel != nil && channel.Type == constant.ChannelTypeCodex {
- return string(constant.EndpointTypeOpenAIResponse)
- }
- return normalized
- }
- func testChannel(channel *model.Channel, testModel string, endpointType string, isStream bool) testResult {
- tik := time.Now()
- var unsupportedTestChannelTypes = []int{
- constant.ChannelTypeMidjourney,
- constant.ChannelTypeMidjourneyPlus,
- constant.ChannelTypeSunoAPI,
- constant.ChannelTypeKling,
- constant.ChannelTypeJimeng,
- constant.ChannelTypeDoubaoVideo,
- constant.ChannelTypeVidu,
- }
- if lo.Contains(unsupportedTestChannelTypes, channel.Type) {
- channelTypeName := constant.GetChannelTypeName(channel.Type)
- return testResult{
- localErr: fmt.Errorf("%s channel test is not supported", channelTypeName),
- }
- }
- w := httptest.NewRecorder()
- c, _ := gin.CreateTestContext(w)
- testModel = strings.TrimSpace(testModel)
- if testModel == "" {
- if channel.TestModel != nil && *channel.TestModel != "" {
- testModel = strings.TrimSpace(*channel.TestModel)
- } else {
- models := channel.GetModels()
- if len(models) > 0 {
- testModel = strings.TrimSpace(models[0])
- }
- if testModel == "" {
- testModel = "gpt-4o-mini"
- }
- }
- }
- endpointType = normalizeChannelTestEndpoint(channel, testModel, endpointType)
- requestPath := "/v1/chat/completions"
- // 如果指定了端点类型,使用指定的端点类型
- if endpointType != "" {
- if endpointInfo, ok := common.GetDefaultEndpointInfo(constant.EndpointType(endpointType)); ok {
- requestPath = endpointInfo.Path
- }
- } else {
- // 如果没有指定端点类型,使用原有的自动检测逻辑
- if strings.Contains(strings.ToLower(testModel), "rerank") {
- requestPath = "/v1/rerank"
- }
- // 先判断是否为 Embedding 模型
- if strings.Contains(strings.ToLower(testModel), "embedding") ||
- strings.HasPrefix(testModel, "m3e") || // m3e 系列模型
- strings.Contains(testModel, "bge-") || // bge 系列模型
- strings.Contains(testModel, "embed") ||
- channel.Type == constant.ChannelTypeMokaAI { // 其他 embedding 模型
- requestPath = "/v1/embeddings" // 修改请求路径
- }
- // VolcEngine 图像生成模型
- if channel.Type == constant.ChannelTypeVolcEngine && strings.Contains(testModel, "seedream") {
- requestPath = "/v1/images/generations"
- }
- // responses-only models
- if strings.Contains(strings.ToLower(testModel), "codex") {
- requestPath = "/v1/responses"
- }
- // responses compaction models (must use /v1/responses/compact)
- if strings.HasSuffix(testModel, ratio_setting.CompactModelSuffix) {
- requestPath = "/v1/responses/compact"
- }
- }
- if strings.HasPrefix(requestPath, "/v1/responses/compact") {
- testModel = ratio_setting.WithCompactModelSuffix(testModel)
- }
- c.Request = &http.Request{
- Method: "POST",
- URL: &url.URL{Path: requestPath}, // 使用动态路径
- Body: nil,
- Header: make(http.Header),
- }
- cache, err := model.GetUserCache(1)
- if err != nil {
- return testResult{
- localErr: err,
- newAPIError: nil,
- }
- }
- cache.WriteContext(c)
- //c.Request.Header.Set("Authorization", "Bearer "+channel.Key)
- c.Request.Header.Set("Content-Type", "application/json")
- c.Set("channel", channel.Type)
- c.Set("base_url", channel.GetBaseURL())
- group, _ := model.GetUserGroup(1, false)
- c.Set("group", group)
- newAPIError := middleware.SetupContextForSelectedChannel(c, channel, testModel)
- if newAPIError != nil {
- return testResult{
- context: c,
- localErr: newAPIError,
- newAPIError: newAPIError,
- }
- }
- // Determine relay format based on endpoint type or request path
- var relayFormat types.RelayFormat
- if endpointType != "" {
- // 根据指定的端点类型设置 relayFormat
- switch constant.EndpointType(endpointType) {
- case constant.EndpointTypeOpenAI:
- relayFormat = types.RelayFormatOpenAI
- case constant.EndpointTypeOpenAIResponse:
- relayFormat = types.RelayFormatOpenAIResponses
- case constant.EndpointTypeOpenAIResponseCompact:
- relayFormat = types.RelayFormatOpenAIResponsesCompaction
- case constant.EndpointTypeAnthropic:
- relayFormat = types.RelayFormatClaude
- case constant.EndpointTypeGemini:
- relayFormat = types.RelayFormatGemini
- case constant.EndpointTypeJinaRerank:
- relayFormat = types.RelayFormatRerank
- case constant.EndpointTypeImageGeneration:
- relayFormat = types.RelayFormatOpenAIImage
- case constant.EndpointTypeEmbeddings:
- relayFormat = types.RelayFormatEmbedding
- default:
- relayFormat = types.RelayFormatOpenAI
- }
- } else {
- // 根据请求路径自动检测
- relayFormat = types.RelayFormatOpenAI
- if c.Request.URL.Path == "/v1/embeddings" {
- relayFormat = types.RelayFormatEmbedding
- }
- if c.Request.URL.Path == "/v1/images/generations" {
- relayFormat = types.RelayFormatOpenAIImage
- }
- if c.Request.URL.Path == "/v1/messages" {
- relayFormat = types.RelayFormatClaude
- }
- if strings.Contains(c.Request.URL.Path, "/v1beta/models") {
- relayFormat = types.RelayFormatGemini
- }
- if c.Request.URL.Path == "/v1/rerank" || c.Request.URL.Path == "/rerank" {
- relayFormat = types.RelayFormatRerank
- }
- if c.Request.URL.Path == "/v1/responses" {
- relayFormat = types.RelayFormatOpenAIResponses
- }
- if strings.HasPrefix(c.Request.URL.Path, "/v1/responses/compact") {
- relayFormat = types.RelayFormatOpenAIResponsesCompaction
- }
- }
- request := buildTestRequest(testModel, endpointType, channel, isStream)
- info, err := relaycommon.GenRelayInfo(c, relayFormat, request, nil)
- if err != nil {
- return testResult{
- context: c,
- localErr: err,
- newAPIError: types.NewError(err, types.ErrorCodeGenRelayInfoFailed),
- }
- }
- info.IsChannelTest = true
- info.InitChannelMeta(c)
- err = helper.ModelMappedHelper(c, info, request)
- if err != nil {
- return testResult{
- context: c,
- localErr: err,
- newAPIError: types.NewError(err, types.ErrorCodeChannelModelMappedError),
- }
- }
- testModel = info.UpstreamModelName
- // 更新请求中的模型名称
- request.SetModelName(testModel)
- apiType, _ := common.ChannelType2APIType(channel.Type)
- if info.RelayMode == relayconstant.RelayModeResponsesCompact &&
- apiType != constant.APITypeOpenAI &&
- apiType != constant.APITypeCodex {
- return testResult{
- context: c,
- localErr: fmt.Errorf("responses compaction test only supports openai/codex channels, got api type %d", apiType),
- newAPIError: types.NewError(fmt.Errorf("unsupported api type: %d", apiType), types.ErrorCodeInvalidApiType),
- }
- }
- adaptor := relay.GetAdaptor(apiType)
- if adaptor == nil {
- return testResult{
- context: c,
- localErr: fmt.Errorf("invalid api type: %d, adaptor is nil", apiType),
- newAPIError: types.NewError(fmt.Errorf("invalid api type: %d, adaptor is nil", apiType), types.ErrorCodeInvalidApiType),
- }
- }
- //// 创建一个用于日志的 info 副本,移除 ApiKey
- //logInfo := info
- //logInfo.ApiKey = ""
- common.SysLog(fmt.Sprintf("testing channel %d with model %s , info %+v ", channel.Id, testModel, info.ToString()))
- priceData, err := helper.ModelPriceHelper(c, info, 0, request.GetTokenCountMeta())
- if err != nil {
- return testResult{
- context: c,
- localErr: err,
- newAPIError: types.NewError(err, types.ErrorCodeModelPriceError),
- }
- }
- adaptor.Init(info)
- var convertedRequest any
- // 根据 RelayMode 选择正确的转换函数
- switch info.RelayMode {
- case relayconstant.RelayModeEmbeddings:
- // Embedding 请求 - request 已经是正确的类型
- if embeddingReq, ok := request.(*dto.EmbeddingRequest); ok {
- convertedRequest, err = adaptor.ConvertEmbeddingRequest(c, info, *embeddingReq)
- } else {
- return testResult{
- context: c,
- localErr: errors.New("invalid embedding request type"),
- newAPIError: types.NewError(errors.New("invalid embedding request type"), types.ErrorCodeConvertRequestFailed),
- }
- }
- case relayconstant.RelayModeImagesGenerations:
- // 图像生成请求 - request 已经是正确的类型
- if imageReq, ok := request.(*dto.ImageRequest); ok {
- convertedRequest, err = adaptor.ConvertImageRequest(c, info, *imageReq)
- } else {
- return testResult{
- context: c,
- localErr: errors.New("invalid image request type"),
- newAPIError: types.NewError(errors.New("invalid image request type"), types.ErrorCodeConvertRequestFailed),
- }
- }
- case relayconstant.RelayModeRerank:
- // Rerank 请求 - request 已经是正确的类型
- if rerankReq, ok := request.(*dto.RerankRequest); ok {
- convertedRequest, err = adaptor.ConvertRerankRequest(c, info.RelayMode, *rerankReq)
- } else {
- return testResult{
- context: c,
- localErr: errors.New("invalid rerank request type"),
- newAPIError: types.NewError(errors.New("invalid rerank request type"), types.ErrorCodeConvertRequestFailed),
- }
- }
- case relayconstant.RelayModeResponses:
- // Response 请求 - request 已经是正确的类型
- if responseReq, ok := request.(*dto.OpenAIResponsesRequest); ok {
- convertedRequest, err = adaptor.ConvertOpenAIResponsesRequest(c, info, *responseReq)
- } else {
- return testResult{
- context: c,
- localErr: errors.New("invalid response request type"),
- newAPIError: types.NewError(errors.New("invalid response request type"), types.ErrorCodeConvertRequestFailed),
- }
- }
- case relayconstant.RelayModeResponsesCompact:
- // Response compaction request - convert to OpenAIResponsesRequest before adapting
- switch req := request.(type) {
- case *dto.OpenAIResponsesCompactionRequest:
- convertedRequest, err = adaptor.ConvertOpenAIResponsesRequest(c, info, dto.OpenAIResponsesRequest{
- Model: req.Model,
- Input: req.Input,
- Instructions: req.Instructions,
- PreviousResponseID: req.PreviousResponseID,
- })
- case *dto.OpenAIResponsesRequest:
- convertedRequest, err = adaptor.ConvertOpenAIResponsesRequest(c, info, *req)
- default:
- return testResult{
- context: c,
- localErr: errors.New("invalid response compaction request type"),
- newAPIError: types.NewError(errors.New("invalid response compaction request type"), types.ErrorCodeConvertRequestFailed),
- }
- }
- default:
- // Chat/Completion 等其他请求类型
- if generalReq, ok := request.(*dto.GeneralOpenAIRequest); ok {
- convertedRequest, err = adaptor.ConvertOpenAIRequest(c, info, generalReq)
- } else {
- return testResult{
- context: c,
- localErr: errors.New("invalid general request type"),
- newAPIError: types.NewError(errors.New("invalid general request type"), types.ErrorCodeConvertRequestFailed),
- }
- }
- }
- if err != nil {
- return testResult{
- context: c,
- localErr: err,
- newAPIError: types.NewError(err, types.ErrorCodeConvertRequestFailed),
- }
- }
- jsonData, err := json.Marshal(convertedRequest)
- if err != nil {
- return testResult{
- context: c,
- localErr: err,
- newAPIError: types.NewError(err, types.ErrorCodeJsonMarshalFailed),
- }
- }
- //jsonData, err = relaycommon.RemoveDisabledFields(jsonData, info.ChannelOtherSettings)
- //if err != nil {
- // return testResult{
- // context: c,
- // localErr: err,
- // newAPIError: types.NewError(err, types.ErrorCodeConvertRequestFailed),
- // }
- //}
- if len(info.ParamOverride) > 0 {
- jsonData, err = relaycommon.ApplyParamOverride(jsonData, info.ParamOverride, relaycommon.BuildParamOverrideContext(info))
- if err != nil {
- return testResult{
- context: c,
- localErr: err,
- newAPIError: types.NewError(err, types.ErrorCodeChannelParamOverrideInvalid),
- }
- }
- }
- requestBody := bytes.NewBuffer(jsonData)
- c.Request.Body = io.NopCloser(bytes.NewBuffer(jsonData))
- resp, err := adaptor.DoRequest(c, info, requestBody)
- if err != nil {
- return testResult{
- context: c,
- localErr: err,
- newAPIError: types.NewOpenAIError(err, types.ErrorCodeDoRequestFailed, http.StatusInternalServerError),
- }
- }
- var httpResp *http.Response
- if resp != nil {
- httpResp = resp.(*http.Response)
- if httpResp.StatusCode != http.StatusOK {
- err := service.RelayErrorHandler(c.Request.Context(), httpResp, true)
- common.SysError(fmt.Sprintf(
- "channel test bad response: channel_id=%d name=%s type=%d model=%s endpoint_type=%s status=%d err=%v",
- channel.Id,
- channel.Name,
- channel.Type,
- testModel,
- endpointType,
- httpResp.StatusCode,
- err,
- ))
- return testResult{
- context: c,
- localErr: err,
- newAPIError: types.NewOpenAIError(err, types.ErrorCodeBadResponse, http.StatusInternalServerError),
- }
- }
- }
- usageA, respErr := adaptor.DoResponse(c, httpResp, info)
- if respErr != nil {
- return testResult{
- context: c,
- localErr: respErr,
- newAPIError: respErr,
- }
- }
- usage, usageErr := coerceTestUsage(usageA, isStream, info.GetEstimatePromptTokens())
- if usageErr != nil {
- return testResult{
- context: c,
- localErr: usageErr,
- newAPIError: types.NewOpenAIError(usageErr, types.ErrorCodeBadResponseBody, http.StatusInternalServerError),
- }
- }
- result := w.Result()
- respBody, err := readTestResponseBody(result.Body, isStream)
- if err != nil {
- return testResult{
- context: c,
- localErr: err,
- newAPIError: types.NewOpenAIError(err, types.ErrorCodeReadResponseBodyFailed, http.StatusInternalServerError),
- }
- }
- if bodyErr := detectErrorFromTestResponseBody(respBody); bodyErr != nil {
- return testResult{
- context: c,
- localErr: bodyErr,
- newAPIError: types.NewOpenAIError(bodyErr, types.ErrorCodeBadResponseBody, http.StatusInternalServerError),
- }
- }
- info.SetEstimatePromptTokens(usage.PromptTokens)
- quota := 0
- if !priceData.UsePrice {
- quota = usage.PromptTokens + int(math.Round(float64(usage.CompletionTokens)*priceData.CompletionRatio))
- quota = int(math.Round(float64(quota) * priceData.ModelRatio))
- if priceData.ModelRatio != 0 && quota <= 0 {
- quota = 1
- }
- } else {
- quota = int(priceData.ModelPrice * common.QuotaPerUnit)
- }
- tok := time.Now()
- milliseconds := tok.Sub(tik).Milliseconds()
- consumedTime := float64(milliseconds) / 1000.0
- other := service.GenerateTextOtherInfo(c, info, priceData.ModelRatio, priceData.GroupRatioInfo.GroupRatio, priceData.CompletionRatio,
- usage.PromptTokensDetails.CachedTokens, priceData.CacheRatio, priceData.ModelPrice, priceData.GroupRatioInfo.GroupSpecialRatio)
- model.RecordConsumeLog(c, 1, model.RecordConsumeLogParams{
- ChannelId: channel.Id,
- PromptTokens: usage.PromptTokens,
- CompletionTokens: usage.CompletionTokens,
- ModelName: info.OriginModelName,
- TokenName: "模型测试",
- Quota: quota,
- Content: "模型测试",
- UseTimeSeconds: int(consumedTime),
- IsStream: info.IsStream,
- Group: info.UsingGroup,
- Other: other,
- })
- common.SysLog(fmt.Sprintf("testing channel #%d, response: \n%s", channel.Id, string(respBody)))
- return testResult{
- context: c,
- localErr: nil,
- newAPIError: nil,
- }
- }
- func coerceTestUsage(usageAny any, isStream bool, estimatePromptTokens int) (*dto.Usage, error) {
- switch u := usageAny.(type) {
- case *dto.Usage:
- return u, nil
- case dto.Usage:
- return &u, nil
- case nil:
- if !isStream {
- return nil, errors.New("usage is nil")
- }
- usage := &dto.Usage{
- PromptTokens: estimatePromptTokens,
- }
- usage.TotalTokens = usage.PromptTokens
- return usage, nil
- default:
- if !isStream {
- return nil, fmt.Errorf("invalid usage type: %T", usageAny)
- }
- usage := &dto.Usage{
- PromptTokens: estimatePromptTokens,
- }
- usage.TotalTokens = usage.PromptTokens
- return usage, nil
- }
- }
- func readTestResponseBody(body io.ReadCloser, isStream bool) ([]byte, error) {
- defer func() { _ = body.Close() }()
- const maxStreamLogBytes = 8 << 10
- if isStream {
- return io.ReadAll(io.LimitReader(body, maxStreamLogBytes))
- }
- return io.ReadAll(body)
- }
- func detectErrorFromTestResponseBody(respBody []byte) error {
- b := bytes.TrimSpace(respBody)
- if len(b) == 0 {
- return nil
- }
- if message := detectErrorMessageFromJSONBytes(b); message != "" {
- return fmt.Errorf("upstream error: %s", message)
- }
- for _, line := range bytes.Split(b, []byte{'\n'}) {
- line = bytes.TrimSpace(line)
- if len(line) == 0 {
- continue
- }
- if !bytes.HasPrefix(line, []byte("data:")) {
- continue
- }
- payload := bytes.TrimSpace(bytes.TrimPrefix(line, []byte("data:")))
- if len(payload) == 0 || bytes.Equal(payload, []byte("[DONE]")) {
- continue
- }
- if message := detectErrorMessageFromJSONBytes(payload); message != "" {
- return fmt.Errorf("upstream error: %s", message)
- }
- }
- return nil
- }
- func detectErrorMessageFromJSONBytes(jsonBytes []byte) string {
- if len(jsonBytes) == 0 {
- return ""
- }
- if jsonBytes[0] != '{' && jsonBytes[0] != '[' {
- return ""
- }
- errVal := gjson.GetBytes(jsonBytes, "error")
- if !errVal.Exists() || errVal.Type == gjson.Null {
- return ""
- }
- message := gjson.GetBytes(jsonBytes, "error.message").String()
- if message == "" {
- message = gjson.GetBytes(jsonBytes, "error.error.message").String()
- }
- if message == "" && errVal.Type == gjson.String {
- message = errVal.String()
- }
- if message == "" {
- message = errVal.Raw
- }
- message = strings.TrimSpace(message)
- if message == "" {
- return "upstream returned error payload"
- }
- return message
- }
- func buildTestRequest(model string, endpointType string, channel *model.Channel, isStream bool) dto.Request {
- testResponsesInput := json.RawMessage(`[{"role":"user","content":"hi"}]`)
- // 根据端点类型构建不同的测试请求
- if endpointType != "" {
- switch constant.EndpointType(endpointType) {
- case constant.EndpointTypeEmbeddings:
- // 返回 EmbeddingRequest
- return &dto.EmbeddingRequest{
- Model: model,
- Input: []any{"hello world"},
- }
- case constant.EndpointTypeImageGeneration:
- // 返回 ImageRequest
- return &dto.ImageRequest{
- Model: model,
- Prompt: "a cute cat",
- N: 1,
- Size: "1024x1024",
- }
- case constant.EndpointTypeJinaRerank:
- // 返回 RerankRequest
- return &dto.RerankRequest{
- Model: model,
- Query: "What is Deep Learning?",
- Documents: []any{"Deep Learning is a subset of machine learning.", "Machine learning is a field of artificial intelligence."},
- TopN: 2,
- }
- case constant.EndpointTypeOpenAIResponse:
- // 返回 OpenAIResponsesRequest
- return &dto.OpenAIResponsesRequest{
- Model: model,
- Input: json.RawMessage(`[{"role":"user","content":"hi"}]`),
- Stream: isStream,
- }
- case constant.EndpointTypeOpenAIResponseCompact:
- // 返回 OpenAIResponsesCompactionRequest
- return &dto.OpenAIResponsesCompactionRequest{
- Model: model,
- Input: testResponsesInput,
- }
- case constant.EndpointTypeAnthropic, constant.EndpointTypeGemini, constant.EndpointTypeOpenAI:
- // 返回 GeneralOpenAIRequest
- maxTokens := uint(16)
- if constant.EndpointType(endpointType) == constant.EndpointTypeGemini {
- maxTokens = 3000
- }
- req := &dto.GeneralOpenAIRequest{
- Model: model,
- Stream: isStream,
- Messages: []dto.Message{
- {
- Role: "user",
- Content: "hi",
- },
- },
- MaxTokens: maxTokens,
- }
- if isStream {
- req.StreamOptions = &dto.StreamOptions{IncludeUsage: true}
- }
- return req
- }
- }
- // 自动检测逻辑(保持原有行为)
- if strings.Contains(strings.ToLower(model), "rerank") {
- return &dto.RerankRequest{
- Model: model,
- Query: "What is Deep Learning?",
- Documents: []any{"Deep Learning is a subset of machine learning.", "Machine learning is a field of artificial intelligence."},
- TopN: 2,
- }
- }
- // 先判断是否为 Embedding 模型
- if strings.Contains(strings.ToLower(model), "embedding") ||
- strings.HasPrefix(model, "m3e") ||
- strings.Contains(model, "bge-") {
- // 返回 EmbeddingRequest
- return &dto.EmbeddingRequest{
- Model: model,
- Input: []any{"hello world"},
- }
- }
- // Responses compaction models (must use /v1/responses/compact)
- if strings.HasSuffix(model, ratio_setting.CompactModelSuffix) {
- return &dto.OpenAIResponsesCompactionRequest{
- Model: model,
- Input: testResponsesInput,
- }
- }
- // Responses-only models (e.g. codex series)
- if strings.Contains(strings.ToLower(model), "codex") {
- return &dto.OpenAIResponsesRequest{
- Model: model,
- Input: json.RawMessage(`[{"role":"user","content":"hi"}]`),
- Stream: isStream,
- }
- }
- // Chat/Completion 请求 - 返回 GeneralOpenAIRequest
- testRequest := &dto.GeneralOpenAIRequest{
- Model: model,
- Stream: isStream,
- Messages: []dto.Message{
- {
- Role: "user",
- Content: "hi",
- },
- },
- }
- if isStream {
- testRequest.StreamOptions = &dto.StreamOptions{IncludeUsage: true}
- }
- if strings.HasPrefix(model, "o") {
- testRequest.MaxCompletionTokens = 16
- } else if strings.Contains(model, "thinking") {
- if !strings.Contains(model, "claude") {
- testRequest.MaxTokens = 50
- }
- } else if strings.Contains(model, "gemini") {
- testRequest.MaxTokens = 3000
- } else {
- testRequest.MaxTokens = 16
- }
- return testRequest
- }
- func TestChannel(c *gin.Context) {
- channelId, err := strconv.Atoi(c.Param("id"))
- if err != nil {
- common.ApiError(c, err)
- return
- }
- channel, err := model.CacheGetChannel(channelId)
- if err != nil {
- channel, err = model.GetChannelById(channelId, true)
- if err != nil {
- common.ApiError(c, err)
- return
- }
- }
- //defer func() {
- // if channel.ChannelInfo.IsMultiKey {
- // go func() { _ = channel.SaveChannelInfo() }()
- // }
- //}()
- testModel := c.Query("model")
- endpointType := c.Query("endpoint_type")
- isStream, _ := strconv.ParseBool(c.Query("stream"))
- tik := time.Now()
- result := testChannel(channel, testModel, endpointType, isStream)
- if result.localErr != nil {
- c.JSON(http.StatusOK, gin.H{
- "success": false,
- "message": result.localErr.Error(),
- "time": 0.0,
- })
- return
- }
- tok := time.Now()
- milliseconds := tok.Sub(tik).Milliseconds()
- go channel.UpdateResponseTime(milliseconds)
- consumedTime := float64(milliseconds) / 1000.0
- if result.newAPIError != nil {
- c.JSON(http.StatusOK, gin.H{
- "success": false,
- "message": result.newAPIError.Error(),
- "time": consumedTime,
- })
- return
- }
- c.JSON(http.StatusOK, gin.H{
- "success": true,
- "message": "",
- "time": consumedTime,
- })
- }
- var testAllChannelsLock sync.Mutex
- var testAllChannelsRunning bool = false
- func testAllChannels(notify bool) error {
- testAllChannelsLock.Lock()
- if testAllChannelsRunning {
- testAllChannelsLock.Unlock()
- return errors.New("测试已在运行中")
- }
- testAllChannelsRunning = true
- testAllChannelsLock.Unlock()
- channels, getChannelErr := model.GetAllChannels(0, 0, true, false)
- if getChannelErr != nil {
- return getChannelErr
- }
- var disableThreshold = int64(common.ChannelDisableThreshold * 1000)
- if disableThreshold == 0 {
- disableThreshold = 10000000 // a impossible value
- }
- gopool.Go(func() {
- // 使用 defer 确保无论如何都会重置运行状态,防止死锁
- defer func() {
- testAllChannelsLock.Lock()
- testAllChannelsRunning = false
- testAllChannelsLock.Unlock()
- }()
- for _, channel := range channels {
- isChannelEnabled := channel.Status == common.ChannelStatusEnabled
- tik := time.Now()
- result := testChannel(channel, "", "", false)
- tok := time.Now()
- milliseconds := tok.Sub(tik).Milliseconds()
- shouldBanChannel := false
- newAPIError := result.newAPIError
- // request error disables the channel
- if newAPIError != nil {
- shouldBanChannel = service.ShouldDisableChannel(channel.Type, result.newAPIError)
- }
- // 当错误检查通过,才检查响应时间
- if common.AutomaticDisableChannelEnabled && !shouldBanChannel {
- if milliseconds > disableThreshold {
- err := fmt.Errorf("响应时间 %.2fs 超过阈值 %.2fs", float64(milliseconds)/1000.0, float64(disableThreshold)/1000.0)
- newAPIError = types.NewOpenAIError(err, types.ErrorCodeChannelResponseTimeExceeded, http.StatusRequestTimeout)
- shouldBanChannel = true
- }
- }
- // disable channel
- if isChannelEnabled && shouldBanChannel && channel.GetAutoBan() {
- processChannelError(result.context, *types.NewChannelError(channel.Id, channel.Type, channel.Name, channel.ChannelInfo.IsMultiKey, common.GetContextKeyString(result.context, constant.ContextKeyChannelKey), channel.GetAutoBan()), newAPIError)
- }
- // enable channel
- if !isChannelEnabled && service.ShouldEnableChannel(newAPIError, channel.Status) {
- service.EnableChannel(channel.Id, common.GetContextKeyString(result.context, constant.ContextKeyChannelKey), channel.Name)
- }
- channel.UpdateResponseTime(milliseconds)
- time.Sleep(common.RequestInterval)
- }
- if notify {
- service.NotifyRootUser(dto.NotifyTypeChannelTest, "通道测试完成", "所有通道测试已完成")
- }
- })
- return nil
- }
- func TestAllChannels(c *gin.Context) {
- err := testAllChannels(true)
- if err != nil {
- common.ApiError(c, err)
- return
- }
- c.JSON(http.StatusOK, gin.H{
- "success": true,
- "message": "",
- })
- }
- var autoTestChannelsOnce sync.Once
- func AutomaticallyTestChannels() {
- // 只在Master节点定时测试渠道
- if !common.IsMasterNode {
- return
- }
- autoTestChannelsOnce.Do(func() {
- for {
- if !operation_setting.GetMonitorSetting().AutoTestChannelEnabled {
- time.Sleep(1 * time.Minute)
- continue
- }
- for {
- frequency := operation_setting.GetMonitorSetting().AutoTestChannelMinutes
- time.Sleep(time.Duration(int(math.Round(frequency))) * time.Minute)
- common.SysLog(fmt.Sprintf("automatically test channels with interval %f minutes", frequency))
- common.SysLog("automatically testing all channels")
- _ = testAllChannels(false)
- common.SysLog("automatically channel test finished")
- if !operation_setting.GetMonitorSetting().AutoTestChannelEnabled {
- break
- }
- }
- }
- })
- }
|