diff --git a/src/ipc/shared/language_model_helpers.ts b/src/ipc/shared/language_model_helpers.ts index 1a1097d..e891bc3 100644 --- a/src/ipc/shared/language_model_helpers.ts +++ b/src/ipc/shared/language_model_helpers.ts @@ -102,9 +102,9 @@ export const MODEL_OPTIONS: Record = { google: [ // https://ai.google.dev/gemini-api/docs/models#gemini-2.5-pro-preview-03-25 { - name: "gemini-2.5-pro", + name: "gemini-2.5-pro-preview-05-06", displayName: "Gemini 2.5 Pro", - description: "Google's Gemini 2.5 Pro model", + description: "Preview version of Google's Gemini 2.5 Pro model", // See Flash 2.5 comment below (go 1 below just to be safe, even though it seems OK now). maxOutputTokens: 65_536 - 1, // Gemini context window = input token + output token @@ -113,9 +113,10 @@ export const MODEL_OPTIONS: Record = { }, // https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview { - name: "gemini-2.5-flash", + name: "gemini-2.5-flash-preview-05-20", displayName: "Gemini 2.5 Flash", - description: "Google's Gemini 2.5 Flash model (free tier available)", + description: + "Preview version of Google's Gemini 2.5 Flash model (free tier available)", // Weirdly for Vertex AI, the output token limit is *exclusive* of the stated limit. maxOutputTokens: 65_536 - 1, // Gemini context window = input token + output token