Lower o3/o4 mini max output tokens to maximize input tokens (#424)
This commit is contained in:
@@ -44,7 +44,8 @@ export const MODEL_OPTIONS: Record<string, ModelOption[]> = {
|
|||||||
name: "o3-mini",
|
name: "o3-mini",
|
||||||
displayName: "o3 mini",
|
displayName: "o3 mini",
|
||||||
description: "Reasoning model",
|
description: "Reasoning model",
|
||||||
maxOutputTokens: 100_000,
|
// See o4-mini comment below for why we set this to 32k
|
||||||
|
maxOutputTokens: 32_000,
|
||||||
contextWindow: 200_000,
|
contextWindow: 200_000,
|
||||||
},
|
},
|
||||||
// https://platform.openai.com/docs/models/o4-mini
|
// https://platform.openai.com/docs/models/o4-mini
|
||||||
@@ -52,7 +53,10 @@ export const MODEL_OPTIONS: Record<string, ModelOption[]> = {
|
|||||||
name: "o4-mini",
|
name: "o4-mini",
|
||||||
displayName: "o4 mini",
|
displayName: "o4 mini",
|
||||||
description: "Reasoning model",
|
description: "Reasoning model",
|
||||||
maxOutputTokens: 100_000,
|
// Technically the max output tokens is 100k, *however* if the user has a lot of input tokens,
|
||||||
|
// then setting a high max output token will cause the request to fail because
|
||||||
|
// the max output tokens is *included* in the context window limit.
|
||||||
|
maxOutputTokens: 32_000,
|
||||||
contextWindow: 200_000,
|
contextWindow: 200_000,
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
|
|||||||
Reference in New Issue
Block a user