|
@@ -1,6 +1,5 @@
|
|
|
import { Anthropic } from "@anthropic-ai/sdk"
|
|
import { Anthropic } from "@anthropic-ai/sdk"
|
|
|
import { ModelInfo } from "../../shared/api"
|
|
import { ModelInfo } from "../../shared/api"
|
|
|
-import { MessageParam } from "@anthropic-ai/sdk/resources/messages.mjs"
|
|
|
|
|
|
|
|
|
|
export function truncateConversation(
|
|
export function truncateConversation(
|
|
|
messages: Anthropic.Messages.MessageParam[],
|
|
messages: Anthropic.Messages.MessageParam[],
|
|
@@ -16,10 +15,10 @@ export function truncateConversation(
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
export function truncateConversationIfNeeded(
|
|
export function truncateConversationIfNeeded(
|
|
|
- messages: MessageParam[],
|
|
|
|
|
|
|
+ messages: Anthropic.Messages.MessageParam[],
|
|
|
totalTokens: number,
|
|
totalTokens: number,
|
|
|
modelInfo: ModelInfo,
|
|
modelInfo: ModelInfo,
|
|
|
-): MessageParam[] {
|
|
|
|
|
|
|
+): Anthropic.Messages.MessageParam[] {
|
|
|
if (modelInfo.supportsPromptCache) {
|
|
if (modelInfo.supportsPromptCache) {
|
|
|
return totalTokens < getMaxTokensForPromptCachingModels(modelInfo)
|
|
return totalTokens < getMaxTokensForPromptCachingModels(modelInfo)
|
|
|
? messages
|
|
? messages
|