| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030 |
- import { beforeEach, describe, expect, it, vi } from "vitest";
- import type { ModelPrice, ModelPriceData } from "@/types/model-price";
- import type { SystemSettings } from "@/types/system-config";
- const asyncTasks: Promise<void>[] = [];
- const cloudPriceSyncRequests: Array<{ reason: string }> = [];
- vi.mock("@/lib/async-task-manager", () => ({
- AsyncTaskManager: {
- register: (_taskId: string, promise: Promise<void>) => {
- asyncTasks.push(promise);
- return new AbortController();
- },
- cleanup: () => {},
- cancel: () => {},
- },
- }));
- vi.mock("@/lib/logger", () => ({
- logger: {
- debug: () => {},
- info: () => {},
- warn: () => {},
- error: () => {},
- trace: () => {},
- },
- }));
- vi.mock("@/lib/price-sync/cloud-price-updater", () => ({
- requestCloudPriceTableSync: (payload: { reason: string }) => {
- cloudPriceSyncRequests.push(payload);
- },
- }));
- vi.mock("@/repository/model-price", () => ({
- findLatestPriceByModel: vi.fn(),
- }));
- vi.mock("@/repository/system-config", () => ({
- getSystemSettings: vi.fn(),
- }));
- vi.mock("@/repository/message", () => ({
- updateMessageRequestCost: vi.fn(),
- updateMessageRequestDetails: vi.fn(),
- updateMessageRequestDuration: vi.fn(),
- }));
- vi.mock("@/lib/session-manager", () => ({
- SessionManager: {
- updateSessionUsage: vi.fn(),
- storeSessionResponse: vi.fn(),
- extractCodexPromptCacheKey: vi.fn(),
- updateSessionWithCodexCacheKey: vi.fn(),
- },
- }));
- vi.mock("@/lib/rate-limit", () => ({
- RateLimitService: {
- trackCost: vi.fn(),
- trackUserDailyCost: vi.fn(),
- },
- }));
- vi.mock("@/lib/session-tracker", () => ({
- SessionTracker: {
- refreshSession: vi.fn(),
- },
- }));
- vi.mock("@/lib/proxy-status-tracker", () => ({
- ProxyStatusTracker: {
- getInstance: () => ({
- endRequest: () => {},
- }),
- },
- }));
- import { ProxyResponseHandler } from "@/app/v1/_lib/proxy/response-handler";
- import { ProxySession } from "@/app/v1/_lib/proxy/session";
- import { getCachedSystemSettings, invalidateSystemSettingsCache } from "@/lib/config";
- import { SessionManager } from "@/lib/session-manager";
- import { RateLimitService } from "@/lib/rate-limit";
- import { SessionTracker } from "@/lib/session-tracker";
- import {
- updateMessageRequestCost,
- updateMessageRequestDetails,
- updateMessageRequestDuration,
- } from "@/repository/message";
- import { findLatestPriceByModel } from "@/repository/model-price";
- import { getSystemSettings } from "@/repository/system-config";
- beforeEach(() => {
- cloudPriceSyncRequests.splice(0, cloudPriceSyncRequests.length);
- invalidateSystemSettingsCache();
- });
- function makeSystemSettings(
- billingModelSource: SystemSettings["billingModelSource"],
- codexPriorityBillingSource: SystemSettings["codexPriorityBillingSource"] = "requested"
- ): SystemSettings {
- const now = new Date();
- return {
- id: 1,
- siteTitle: "test",
- allowGlobalUsageView: false,
- currencyDisplay: "USD",
- billingModelSource,
- codexPriorityBillingSource,
- timezone: null,
- enableAutoCleanup: false,
- cleanupRetentionDays: 30,
- cleanupSchedule: "0 2 * * *",
- cleanupBatchSize: 10000,
- enableClientVersionCheck: false,
- verboseProviderError: false,
- enableHttp2: false,
- interceptAnthropicWarmupRequests: false,
- enableThinkingSignatureRectifier: true,
- enableThinkingBudgetRectifier: true,
- enableBillingHeaderRectifier: true,
- enableResponseInputRectifier: true,
- enableCodexSessionIdCompletion: true,
- enableClaudeMetadataUserIdInjection: true,
- enableResponseFixer: true,
- responseFixerConfig: {
- fixTruncatedJson: true,
- fixSseFormat: true,
- fixEncoding: true,
- maxJsonDepth: 200,
- maxFixSize: 1024 * 1024,
- },
- createdAt: now,
- updatedAt: now,
- };
- }
- function makePriceRecord(modelName: string, priceData: ModelPriceData): ModelPrice {
- const now = new Date();
- return {
- id: 1,
- modelName,
- priceData,
- createdAt: now,
- updatedAt: now,
- };
- }
- function createSession({
- originalModel,
- redirectedModel,
- sessionId,
- messageId,
- providerOverrides,
- requestMessage,
- }: {
- originalModel: string;
- redirectedModel: string;
- sessionId: string;
- messageId: number;
- providerOverrides?: Record<string, unknown>;
- requestMessage?: Record<string, unknown>;
- }): ProxySession {
- const session = new (
- ProxySession as unknown as {
- new (init: {
- startTime: number;
- method: string;
- requestUrl: URL;
- headers: Headers;
- headerLog: string;
- request: { message: Record<string, unknown>; log: string; model: string | null };
- userAgent: string | null;
- context: unknown;
- clientAbortSignal: AbortSignal | null;
- }): ProxySession;
- }
- )({
- startTime: Date.now(),
- method: "POST",
- requestUrl: new URL("http://localhost/v1/messages"),
- headers: new Headers(),
- headerLog: "",
- request: { message: requestMessage ?? {}, log: "(test)", model: redirectedModel },
- userAgent: null,
- context: {},
- clientAbortSignal: null,
- });
- session.setOriginalModel(originalModel);
- session.setSessionId(sessionId);
- const provider = {
- id: 99,
- name: "test-provider",
- url: "https://api.anthropic.com",
- providerType: "claude",
- costMultiplier: 1.0,
- streamingIdleTimeoutMs: 0,
- ...providerOverrides,
- } as any;
- const user = {
- id: 123,
- name: "test-user",
- dailyResetTime: "00:00",
- dailyResetMode: "fixed",
- } as any;
- const key = {
- id: 456,
- name: "test-key",
- dailyResetTime: "00:00",
- dailyResetMode: "fixed",
- } as any;
- session.setProvider(provider);
- session.setAuthState({
- user,
- key,
- apiKey: "sk-test",
- success: true,
- });
- session.setMessageContext({
- id: messageId,
- createdAt: new Date(),
- user,
- key,
- apiKey: "sk-test",
- });
- return session;
- }
- function createNonStreamResponse(
- usage: { input_tokens: number; output_tokens: number },
- extras?: Record<string, unknown>
- ): Response {
- return new Response(
- JSON.stringify({
- type: "message",
- usage,
- ...(extras ?? {}),
- }),
- {
- status: 200,
- headers: { "content-type": "application/json" },
- }
- );
- }
- function createStreamResponse(usage: { input_tokens: number; output_tokens: number }): Response {
- const sseText = `event: message_delta\ndata: ${JSON.stringify({ usage })}\n\n`;
- const encoder = new TextEncoder();
- const stream = new ReadableStream<Uint8Array>({
- start(controller) {
- controller.enqueue(encoder.encode(sseText));
- controller.close();
- },
- });
- return new Response(stream, {
- status: 200,
- headers: { "content-type": "text/event-stream" },
- });
- }
- async function drainAsyncTasks(): Promise<void> {
- const tasks = asyncTasks.splice(0, asyncTasks.length);
- await Promise.all(tasks);
- }
- function captureRateLimitCosts(): number[] {
- const rateLimitCosts: number[] = [];
- vi.mocked(RateLimitService.trackCost).mockImplementation(
- async (_keyId: number, _providerId: number, _sessionId: string, costUsd: number) => {
- rateLimitCosts.push(costUsd);
- }
- );
- return rateLimitCosts;
- }
- async function runScenario({
- billingModelSource,
- isStream,
- }: {
- billingModelSource: SystemSettings["billingModelSource"];
- isStream: boolean;
- }): Promise<{ dbCostUsd: string; sessionCostUsd: string; rateLimitCost: number }> {
- invalidateSystemSettingsCache();
- const usage = { input_tokens: 2, output_tokens: 3 };
- const originalModel = "original-model";
- const redirectedModel = "redirected-model";
- const originalPriceData: ModelPriceData = { input_cost_per_token: 1, output_cost_per_token: 1 };
- const redirectedPriceData: ModelPriceData = {
- input_cost_per_token: 10,
- output_cost_per_token: 10,
- };
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings(billingModelSource));
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === originalModel) {
- return makePriceRecord(modelName, originalPriceData);
- }
- if (modelName === redirectedModel) {
- return makePriceRecord(modelName, redirectedPriceData);
- }
- return null;
- });
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- });
- const sessionCosts: string[] = [];
- vi.mocked(SessionManager.updateSessionUsage).mockImplementation(
- async (_sessionId: string, payload: Record<string, unknown>) => {
- if (typeof payload.costUsd === "string") {
- sessionCosts.push(payload.costUsd);
- }
- }
- );
- const rateLimitCosts: number[] = [];
- vi.mocked(RateLimitService.trackCost).mockImplementation(
- async (_keyId: number, _providerId: number, _sessionId: string, costUsd: number) => {
- rateLimitCosts.push(costUsd);
- }
- );
- const session = createSession({
- originalModel,
- redirectedModel,
- sessionId: `sess-${billingModelSource}-${isStream ? "s" : "n"}`,
- messageId: isStream ? 2001 : 2000,
- });
- const response = isStream ? createStreamResponse(usage) : createNonStreamResponse(usage);
- const clientResponse = await ProxyResponseHandler.dispatch(session, response);
- if (isStream) {
- await clientResponse.text();
- }
- await drainAsyncTasks();
- const dbCostUsd = dbCosts[0] ?? "";
- const sessionCostUsd = sessionCosts[0] ?? "";
- const rateLimitCost = rateLimitCosts[0] ?? Number.NaN;
- return { dbCostUsd, sessionCostUsd, rateLimitCost };
- }
- describe("Billing model source - Redis session cost vs DB cost", () => {
- it("非流式响应:配置 = original 时 Session 成本与数据库一致", async () => {
- const result = await runScenario({ billingModelSource: "original", isStream: false });
- expect(result.dbCostUsd).toBe("5");
- expect(result.sessionCostUsd).toBe("5");
- expect(result.rateLimitCost).toBe(5);
- });
- it("非流式响应:配置 = redirected 时 Session 成本与数据库一致", async () => {
- const result = await runScenario({ billingModelSource: "redirected", isStream: false });
- expect(result.dbCostUsd).toBe("50");
- expect(result.sessionCostUsd).toBe("50");
- expect(result.rateLimitCost).toBe(50);
- });
- it("流式响应:配置 = original 时 Session 成本与数据库一致", async () => {
- const result = await runScenario({ billingModelSource: "original", isStream: true });
- expect(result.dbCostUsd).toBe("5");
- expect(result.sessionCostUsd).toBe("5");
- expect(result.rateLimitCost).toBe(5);
- });
- it("流式响应:配置 = redirected 时 Session 成本与数据库一致", async () => {
- const result = await runScenario({ billingModelSource: "redirected", isStream: true });
- expect(result.dbCostUsd).toBe("50");
- expect(result.sessionCostUsd).toBe("50");
- expect(result.rateLimitCost).toBe(50);
- });
- it("从 original 切换到 redirected 后应生效", async () => {
- const original = await runScenario({ billingModelSource: "original", isStream: false });
- const redirected = await runScenario({ billingModelSource: "redirected", isStream: false });
- expect(original.sessionCostUsd).toBe("5");
- expect(redirected.sessionCostUsd).toBe("50");
- expect(original.sessionCostUsd).not.toBe(redirected.sessionCostUsd);
- });
- it("nested pricing: gpt-5.4 alias model should bill from pricing.openai when provider is chatgpt", async () => {
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings("redirected"));
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === "gpt-5.4") {
- return makePriceRecord(modelName, {
- mode: "responses",
- model_family: "gpt",
- litellm_provider: "chatgpt",
- pricing: {
- openai: {
- input_cost_per_token: 2.5,
- output_cost_per_token: 15,
- },
- },
- });
- }
- return null;
- });
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(
- async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- }
- );
- const rateLimitCosts = captureRateLimitCosts();
- const sessionCosts: string[] = [];
- vi.mocked(SessionManager.updateSessionUsage).mockImplementation(
- async (_sessionId: string, payload: Record<string, unknown>) => {
- if (typeof payload.costUsd === "string") {
- sessionCosts.push(payload.costUsd);
- }
- }
- );
- const session = createSession({
- originalModel: "gpt-5.4",
- redirectedModel: "gpt-5.4",
- sessionId: "sess-gpt54-chatgpt",
- messageId: 3100,
- providerOverrides: {
- name: "ChatGPT",
- url: "https://chatgpt.com/backend-api/codex",
- providerType: "codex",
- },
- });
- const response = createNonStreamResponse({ input_tokens: 2, output_tokens: 3 });
- await ProxyResponseHandler.dispatch(session, response);
- await drainAsyncTasks();
- expect(dbCosts[0]).toBe("50");
- expect(sessionCosts[0]).toBe("50");
- });
- it("codex fast: requested mode ignores actual priority when request tier is default", async () => {
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings("redirected"));
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === "gpt-5.4") {
- return makePriceRecord(modelName, {
- mode: "responses",
- model_family: "gpt",
- litellm_provider: "chatgpt",
- pricing: {
- openai: {
- input_cost_per_token: 1,
- output_cost_per_token: 10,
- input_cost_per_token_priority: 2,
- output_cost_per_token_priority: 20,
- },
- },
- });
- }
- return null;
- });
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(
- async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- }
- );
- const rateLimitCosts = captureRateLimitCosts();
- const sessionCosts: string[] = [];
- vi.mocked(SessionManager.updateSessionUsage).mockImplementation(
- async (_sessionId: string, payload: Record<string, unknown>) => {
- if (typeof payload.costUsd === "string") {
- sessionCosts.push(payload.costUsd);
- }
- }
- );
- const session = createSession({
- originalModel: "gpt-5.4",
- redirectedModel: "gpt-5.4",
- sessionId: "sess-gpt54-priority-actual",
- messageId: 3200,
- providerOverrides: {
- name: "ChatGPT",
- url: "https://chatgpt.com/backend-api/codex",
- providerType: "codex",
- },
- requestMessage: { service_tier: "default" },
- });
- const response = createNonStreamResponse(
- { input_tokens: 2, output_tokens: 3 },
- { service_tier: "priority" }
- );
- await ProxyResponseHandler.dispatch(session, response);
- await drainAsyncTasks();
- expect(dbCosts[0]).toBe("32");
- expect(sessionCosts[0]).toBe("32");
- expect(rateLimitCosts[0]).toBe(32);
- });
- it("codex fast: falls back to requested priority pricing when response omits service_tier", async () => {
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings("redirected"));
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === "gpt-5.4") {
- return makePriceRecord(modelName, {
- mode: "responses",
- model_family: "gpt",
- litellm_provider: "chatgpt",
- pricing: {
- openai: {
- input_cost_per_token: 1,
- output_cost_per_token: 10,
- input_cost_per_token_priority: 2,
- output_cost_per_token_priority: 20,
- },
- },
- });
- }
- return null;
- });
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(
- async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- }
- );
- const rateLimitCosts = captureRateLimitCosts();
- const session = createSession({
- originalModel: "gpt-5.4",
- redirectedModel: "gpt-5.4",
- sessionId: "sess-gpt54-priority-requested",
- messageId: 3201,
- providerOverrides: {
- name: "ChatGPT",
- url: "https://chatgpt.com/backend-api/codex",
- providerType: "codex",
- },
- requestMessage: { service_tier: "priority" },
- });
- const response = createNonStreamResponse({ input_tokens: 2, output_tokens: 3 });
- await ProxyResponseHandler.dispatch(session, response);
- await drainAsyncTasks();
- expect(dbCosts[0]).toBe("64");
- expect(rateLimitCosts[0]).toBe(64);
- });
- it("codex fast: uses long-context priority pricing when request is priority and response omits service_tier", async () => {
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings("redirected"));
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === "gpt-5.4") {
- return makePriceRecord(modelName, {
- mode: "responses",
- model_family: "gpt",
- litellm_provider: "chatgpt",
- pricing: {
- openai: {
- input_cost_per_token: 1,
- output_cost_per_token: 10,
- input_cost_per_token_priority: 2,
- output_cost_per_token_priority: 20,
- input_cost_per_token_above_272k_tokens: 5,
- output_cost_per_token_above_272k_tokens: 50,
- input_cost_per_token_above_272k_tokens_priority: 7,
- output_cost_per_token_above_272k_tokens_priority: 70,
- },
- },
- });
- }
- return null;
- });
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(
- async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- }
- );
- const rateLimitCosts = captureRateLimitCosts();
- const sessionCosts: string[] = [];
- vi.mocked(SessionManager.updateSessionUsage).mockImplementation(
- async (_sessionId: string, payload: Record<string, unknown>) => {
- if (typeof payload.costUsd === "string") {
- sessionCosts.push(payload.costUsd);
- }
- }
- );
- const session = createSession({
- originalModel: "gpt-5.4",
- redirectedModel: "gpt-5.4",
- sessionId: "sess-gpt54-priority-requested-long-context",
- messageId: 3203,
- providerOverrides: {
- name: "ChatGPT",
- url: "https://chatgpt.com/backend-api/codex",
- providerType: "codex",
- },
- requestMessage: { service_tier: "priority" },
- });
- const response = createNonStreamResponse({ input_tokens: 272001, output_tokens: 2 });
- await ProxyResponseHandler.dispatch(session, response);
- await drainAsyncTasks();
- expect(dbCosts[0]).toBe("1904147");
- expect(sessionCosts[0]).toBe("1904147");
- expect(rateLimitCosts[0]).toBe(1904147);
- });
- it("codex fast: requested mode keeps priority pricing even when actual tier is downgraded", async () => {
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings("redirected"));
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === "gpt-5.4") {
- return makePriceRecord(modelName, {
- mode: "responses",
- model_family: "gpt",
- litellm_provider: "chatgpt",
- pricing: {
- openai: {
- input_cost_per_token: 1,
- output_cost_per_token: 10,
- input_cost_per_token_priority: 2,
- output_cost_per_token_priority: 20,
- },
- },
- });
- }
- return null;
- });
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(
- async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- }
- );
- const rateLimitCosts = captureRateLimitCosts();
- const session = createSession({
- originalModel: "gpt-5.4",
- redirectedModel: "gpt-5.4",
- sessionId: "sess-gpt54-priority-downgraded",
- messageId: 3202,
- providerOverrides: {
- name: "ChatGPT",
- url: "https://chatgpt.com/backend-api/codex",
- providerType: "codex",
- },
- requestMessage: { service_tier: "priority" },
- });
- const response = createNonStreamResponse(
- { input_tokens: 2, output_tokens: 3 },
- { service_tier: "default" }
- );
- await ProxyResponseHandler.dispatch(session, response);
- await drainAsyncTasks();
- expect(dbCosts[0]).toBe("64");
- expect(rateLimitCosts[0]).toBe(64);
- });
- it("codex fast: actual mode uses priority pricing when response reports service_tier=priority", async () => {
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings("redirected", "actual"));
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === "gpt-5.4") {
- return makePriceRecord(modelName, {
- mode: "responses",
- model_family: "gpt",
- litellm_provider: "chatgpt",
- pricing: {
- openai: {
- input_cost_per_token: 1,
- output_cost_per_token: 10,
- input_cost_per_token_priority: 2,
- output_cost_per_token_priority: 20,
- },
- },
- });
- }
- return null;
- });
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(
- async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- }
- );
- const rateLimitCosts = captureRateLimitCosts();
- const session = createSession({
- originalModel: "gpt-5.4",
- redirectedModel: "gpt-5.4",
- sessionId: "sess-gpt54-priority-actual-mode-upgrade",
- messageId: 3204,
- providerOverrides: {
- name: "ChatGPT",
- url: "https://chatgpt.com/backend-api/codex",
- providerType: "codex",
- },
- requestMessage: { service_tier: "default" },
- });
- const response = createNonStreamResponse(
- { input_tokens: 2, output_tokens: 3 },
- { service_tier: "priority" }
- );
- await ProxyResponseHandler.dispatch(session, response);
- await drainAsyncTasks();
- expect(dbCosts[0]).toBe("64");
- expect(rateLimitCosts[0]).toBe(64);
- });
- it("codex fast: actual mode does not use priority pricing when response explicitly reports non-priority tier", async () => {
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings("redirected", "actual"));
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === "gpt-5.4") {
- return makePriceRecord(modelName, {
- mode: "responses",
- model_family: "gpt",
- litellm_provider: "chatgpt",
- pricing: {
- openai: {
- input_cost_per_token: 1,
- output_cost_per_token: 10,
- input_cost_per_token_priority: 2,
- output_cost_per_token_priority: 20,
- },
- },
- });
- }
- return null;
- });
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(
- async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- }
- );
- const rateLimitCosts = captureRateLimitCosts();
- const session = createSession({
- originalModel: "gpt-5.4",
- redirectedModel: "gpt-5.4",
- sessionId: "sess-gpt54-priority-actual-mode-downgrade",
- messageId: 3205,
- providerOverrides: {
- name: "ChatGPT",
- url: "https://chatgpt.com/backend-api/codex",
- providerType: "codex",
- },
- requestMessage: { service_tier: "priority" },
- });
- const response = createNonStreamResponse(
- { input_tokens: 2, output_tokens: 3 },
- { service_tier: "default" }
- );
- await ProxyResponseHandler.dispatch(session, response);
- await drainAsyncTasks();
- expect(dbCosts[0]).toBe("32");
- expect(rateLimitCosts[0]).toBe(32);
- });
- it("codex fast: actual mode falls back to requested priority pricing when response omits service_tier", async () => {
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings("redirected", "actual"));
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === "gpt-5.4") {
- return makePriceRecord(modelName, {
- mode: "responses",
- model_family: "gpt",
- litellm_provider: "chatgpt",
- pricing: {
- openai: {
- input_cost_per_token: 1,
- output_cost_per_token: 10,
- input_cost_per_token_priority: 2,
- output_cost_per_token_priority: 20,
- },
- },
- });
- }
- return null;
- });
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(
- async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- }
- );
- const rateLimitCosts = captureRateLimitCosts();
- const session = createSession({
- originalModel: "gpt-5.4",
- redirectedModel: "gpt-5.4",
- sessionId: "sess-gpt54-priority-actual-mode-fallback",
- messageId: 3206,
- providerOverrides: {
- name: "ChatGPT",
- url: "https://chatgpt.com/backend-api/codex",
- providerType: "codex",
- },
- requestMessage: { service_tier: "priority" },
- });
- const response = createNonStreamResponse({ input_tokens: 2, output_tokens: 3 });
- await ProxyResponseHandler.dispatch(session, response);
- await drainAsyncTasks();
- expect(dbCosts[0]).toBe("64");
- expect(rateLimitCosts[0]).toBe(64);
- });
- it("codex fast: actual mode reuses cached system setting when direct settings read fails", async () => {
- vi.mocked(getSystemSettings).mockResolvedValueOnce(makeSystemSettings("redirected", "actual"));
- await getCachedSystemSettings();
- vi.mocked(getSystemSettings).mockRejectedValueOnce(new Error("db down"));
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(findLatestPriceByModel).mockImplementation(async (modelName: string) => {
- if (modelName === "gpt-5.4") {
- return makePriceRecord(modelName, {
- mode: "responses",
- model_family: "gpt",
- litellm_provider: "chatgpt",
- pricing: {
- openai: {
- input_cost_per_token: 1,
- output_cost_per_token: 10,
- input_cost_per_token_priority: 2,
- output_cost_per_token_priority: 20,
- },
- },
- });
- }
- return null;
- });
- const dbCosts: string[] = [];
- vi.mocked(updateMessageRequestCost).mockImplementation(
- async (_id: number, costUsd: unknown) => {
- dbCosts.push(String(costUsd));
- }
- );
- const rateLimitCosts = captureRateLimitCosts();
- const session = createSession({
- originalModel: "gpt-5.4",
- redirectedModel: "gpt-5.4",
- sessionId: "sess-gpt54-priority-actual-mode-cached-settings",
- messageId: 3207,
- providerOverrides: {
- name: "ChatGPT",
- url: "https://chatgpt.com/backend-api/codex",
- providerType: "codex",
- },
- requestMessage: { service_tier: "priority" },
- });
- const response = createNonStreamResponse(
- { input_tokens: 2, output_tokens: 3 },
- { service_tier: "default" }
- );
- await ProxyResponseHandler.dispatch(session, response);
- await drainAsyncTasks();
- expect(dbCosts[0]).toBe("32");
- expect(rateLimitCosts[0]).toBe(32);
- });
- });
- describe("价格表缺失/查询失败:不计费放行", () => {
- async function runNoPriceScenario(options: {
- billingModelSource: SystemSettings["billingModelSource"];
- isStream: boolean;
- priceLookup: "none" | "throws";
- }): Promise<{ dbCostCalls: number; rateLimitCalls: number }> {
- const usage = { input_tokens: 2, output_tokens: 3 };
- const originalModel = "original-model";
- const redirectedModel = "redirected-model";
- vi.mocked(getSystemSettings).mockResolvedValue(makeSystemSettings(options.billingModelSource));
- if (options.priceLookup === "none") {
- vi.mocked(findLatestPriceByModel).mockResolvedValue(null);
- } else {
- vi.mocked(findLatestPriceByModel).mockImplementation(async () => {
- throw new Error("db query failed");
- });
- }
- vi.mocked(updateMessageRequestDetails).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestDuration).mockResolvedValue(undefined);
- vi.mocked(SessionManager.storeSessionResponse).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackUserDailyCost).mockResolvedValue(undefined);
- vi.mocked(SessionTracker.refreshSession).mockResolvedValue(undefined);
- vi.mocked(updateMessageRequestCost).mockResolvedValue(undefined);
- vi.mocked(RateLimitService.trackCost).mockResolvedValue(undefined);
- vi.mocked(SessionManager.updateSessionUsage).mockResolvedValue(undefined);
- const session = createSession({
- originalModel,
- redirectedModel,
- sessionId: `sess-no-price-${options.billingModelSource}-${options.isStream ? "s" : "n"}`,
- messageId: options.isStream ? 3001 : 3000,
- });
- const response = options.isStream
- ? createStreamResponse(usage)
- : createNonStreamResponse(usage);
- const clientResponse = await ProxyResponseHandler.dispatch(session, response);
- await clientResponse.text();
- await drainAsyncTasks();
- return {
- dbCostCalls: vi.mocked(updateMessageRequestCost).mock.calls.length,
- rateLimitCalls: vi.mocked(RateLimitService.trackCost).mock.calls.length,
- };
- }
- it("无价格:不写入 DB cost,不追踪限流 cost,并触发一次异步同步", async () => {
- const result = await runNoPriceScenario({
- billingModelSource: "redirected",
- isStream: false,
- priceLookup: "none",
- });
- expect(result.dbCostCalls).toBe(0);
- expect(result.rateLimitCalls).toBe(0);
- expect(cloudPriceSyncRequests).toEqual([{ reason: "missing-model" }]);
- });
- it("价格查询抛错:不应影响响应,不写入 DB cost,不追踪限流 cost", async () => {
- const result = await runNoPriceScenario({
- billingModelSource: "original",
- isStream: true,
- priceLookup: "throws",
- });
- expect(result.dbCostCalls).toBe(0);
- expect(result.rateLimitCalls).toBe(0);
- });
- });
|