diff --git a/packages/openai-adapters/src/apis/Anthropic.ts b/packages/openai-adapters/src/apis/Anthropic.ts index 0336bab9f3..8beec707c5 100644 --- a/packages/openai-adapters/src/apis/Anthropic.ts +++ b/packages/openai-adapters/src/apis/Anthropic.ts @@ -436,7 +436,7 @@ export class AnthropicApi implements BaseLlmApi { system: systemText, messages: nonSystemMessages as any, temperature: body.temperature ?? undefined, - maxTokens: body.max_tokens ?? undefined, + maxOutputTokens: body.max_tokens ?? undefined, topP: body.top_p ?? undefined, stopSequences: body.stop ? Array.isArray(body.stop) @@ -454,7 +454,7 @@ export class AnthropicApi implements BaseLlmApi { type: "function" as const, function: { name: tc.toolName, - arguments: JSON.stringify(tc.args), + arguments: JSON.stringify(tc.input), }, })); @@ -478,9 +478,9 @@ export class AnthropicApi implements BaseLlmApi { }, ], usage: { - prompt_tokens: result.usage.promptTokens, - completion_tokens: result.usage.completionTokens, - total_tokens: result.usage.totalTokens, + prompt_tokens: result.usage.inputTokens ?? 0, + completion_tokens: result.usage.outputTokens ?? 0, + total_tokens: result.usage.totalTokens ?? 0, prompt_tokens_details: { cached_tokens: (result.usage as any).promptTokensDetails?.cachedTokens ?? 0, @@ -652,7 +652,7 @@ export class AnthropicApi implements BaseLlmApi { system: systemText, messages: nonSystemMessages as any, temperature: body.temperature ?? undefined, - maxTokens: body.max_tokens ?? undefined, + maxOutputTokens: body.max_tokens ?? undefined, topP: body.top_p ?? undefined, stopSequences: body.stop ? Array.isArray(body.stop) diff --git a/packages/openai-adapters/src/apis/OpenAI.ts b/packages/openai-adapters/src/apis/OpenAI.ts index c865c00b7d..f67364a291 100644 --- a/packages/openai-adapters/src/apis/OpenAI.ts +++ b/packages/openai-adapters/src/apis/OpenAI.ts @@ -196,7 +196,7 @@ export class OpenAIApi implements BaseLlmApi { model, messages: modifiedBody.messages as any, temperature: modifiedBody.temperature ?? undefined, - maxTokens: + maxOutputTokens: modifiedBody.max_completion_tokens ?? modifiedBody.max_tokens ?? undefined, @@ -219,7 +219,7 @@ export class OpenAIApi implements BaseLlmApi { type: "function" as const, function: { name: tc.toolName, - arguments: JSON.stringify(tc.args), + arguments: JSON.stringify(tc.input), }, })); @@ -243,9 +243,9 @@ export class OpenAIApi implements BaseLlmApi { }, ], usage: { - prompt_tokens: result.usage.promptTokens, - completion_tokens: result.usage.completionTokens, - total_tokens: result.usage.totalTokens, + prompt_tokens: result.usage.inputTokens ?? 0, + completion_tokens: result.usage.outputTokens ?? 0, + total_tokens: result.usage.totalTokens ?? 0, }, }; } @@ -316,7 +316,7 @@ export class OpenAIApi implements BaseLlmApi { model, messages: modifiedBody.messages as any, temperature: modifiedBody.temperature ?? undefined, - maxTokens: + maxOutputTokens: modifiedBody.max_completion_tokens ?? modifiedBody.max_tokens ?? undefined,