Skip to content

Commit 108069a

Browse files
authored
Merge pull request ChatGPTNextWeb#5788 from ConnectAI-E/fix-o1-maxtokens
chore: o1模型使用max_completion_tokens
2 parents fbb9385 + d5bda29 commit 108069a

File tree

1 file changed

+6
-0
lines changed

1 file changed

+6
-0
lines changed

app/client/platforms/openai.ts

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,7 @@ export interface RequestPayload {
6565
frequency_penalty: number;
6666
top_p: number;
6767
max_tokens?: number;
68+
max_completion_tokens?: number;
6869
}
6970

7071
export interface DalleRequestPayload {
@@ -233,6 +234,11 @@ export class ChatGPTApi implements LLMApi {
233234
// Please do not ask me why not send max_tokens, no reason, this param is just shit, I dont want to explain anymore.
234235
};
235236

237+
// O1 使用 max_completion_tokens 控制token数 (https://platform.openai.com/docs/guides/reasoning#controlling-costs)
238+
if (isO1) {
239+
requestPayload["max_completion_tokens"] = modelConfig.max_tokens;
240+
}
241+
236242
// add max_tokens to vision model
237243
if (visionModel) {
238244
requestPayload["max_tokens"] = Math.max(modelConfig.max_tokens, 4000);

0 commit comments

Comments
 (0)