From 2329d59c83496b69749c370bc2cb3013156ea99f Mon Sep 17 00:00:00 2001 From: Davidlasky Date: Tue, 1 Apr 2025 13:33:38 -0500 Subject: [PATCH] defaults o3-mini to high --- app/client/platforms/openai.ts | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/app/client/platforms/openai.ts b/app/client/platforms/openai.ts index a4ac638c9..b6c2393e6 100644 --- a/app/client/platforms/openai.ts +++ b/app/client/platforms/openai.ts @@ -67,6 +67,8 @@ export interface RequestPayload { top_p: number; max_tokens?: number; max_completion_tokens?: number; + reasoning_effort?: string; + // O3 only } export interface DalleRequestPayload { @@ -239,7 +241,11 @@ export class ChatGPTApi implements LLMApi { // O1 使用 max_completion_tokens 控制token数 (https://platform.openai.com/docs/guides/reasoning#controlling-costs) if (isO1OrO3) { - requestPayload["max_completion_tokens"] = modelConfig.max_tokens; + requestPayload["max_completion_tokens"] = 25000; + } + + if (isO3) { + requestPayload["reasoning_effort"] = "high"; } // add max_tokens to vision model