|
@@ -599,7 +599,7 @@ describe("OpenAiHandler", () => {
|
|
|
stream: true,
|
|
stream: true,
|
|
|
stream_options: { include_usage: true },
|
|
stream_options: { include_usage: true },
|
|
|
reasoning_effort: "medium",
|
|
reasoning_effort: "medium",
|
|
|
- temperature: 0.5,
|
|
|
|
|
|
|
+ temperature: undefined,
|
|
|
// O3 models do not support deprecated max_tokens but do support max_completion_tokens
|
|
// O3 models do not support deprecated max_tokens but do support max_completion_tokens
|
|
|
max_completion_tokens: 32000,
|
|
max_completion_tokens: 32000,
|
|
|
}),
|
|
}),
|
|
@@ -640,7 +640,7 @@ describe("OpenAiHandler", () => {
|
|
|
stream: true,
|
|
stream: true,
|
|
|
stream_options: { include_usage: true },
|
|
stream_options: { include_usage: true },
|
|
|
reasoning_effort: "medium",
|
|
reasoning_effort: "medium",
|
|
|
- temperature: 0.7,
|
|
|
|
|
|
|
+ temperature: undefined,
|
|
|
}),
|
|
}),
|
|
|
{},
|
|
{},
|
|
|
)
|
|
)
|
|
@@ -682,7 +682,7 @@ describe("OpenAiHandler", () => {
|
|
|
{ role: "user", content: "Hello!" },
|
|
{ role: "user", content: "Hello!" },
|
|
|
],
|
|
],
|
|
|
reasoning_effort: "medium",
|
|
reasoning_effort: "medium",
|
|
|
- temperature: 0.3,
|
|
|
|
|
|
|
+ temperature: undefined,
|
|
|
// O3 models do not support deprecated max_tokens but do support max_completion_tokens
|
|
// O3 models do not support deprecated max_tokens but do support max_completion_tokens
|
|
|
max_completion_tokens: 65536, // Using default maxTokens from o3Options
|
|
max_completion_tokens: 65536, // Using default maxTokens from o3Options
|
|
|
}),
|
|
}),
|
|
@@ -712,7 +712,7 @@ describe("OpenAiHandler", () => {
|
|
|
|
|
|
|
|
expect(mockCreate).toHaveBeenCalledWith(
|
|
expect(mockCreate).toHaveBeenCalledWith(
|
|
|
expect.objectContaining({
|
|
expect.objectContaining({
|
|
|
- temperature: 0, // Default temperature
|
|
|
|
|
|
|
+ temperature: undefined, // Temperature is not supported for O3 models
|
|
|
}),
|
|
}),
|
|
|
{},
|
|
{},
|
|
|
)
|
|
)
|