diff --git a/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context---auto-includes-only-1.txt b/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context---auto-includes-only-1.txt index 53ee547..e144756 100644 --- a/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context---auto-includes-only-1.txt +++ b/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context---auto-includes-only-1.txt @@ -1,6 +1,6 @@ { "body": { - "model": "gemini/gemini-2.5-pro-preview-05-06", + "model": "gemini/gemini-2.5-pro", "max_tokens": 65535, "temperature": 0, "messages": [ diff --git a/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-1.txt b/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-1.txt index 9a4be9b..fd5650c 100644 --- a/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-1.txt +++ b/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-1.txt @@ -1,6 +1,6 @@ { "body": { - "model": "gemini/gemini-2.5-pro-preview-05-06", + "model": "gemini/gemini-2.5-pro", "max_tokens": 65535, "temperature": 0, "messages": [ diff --git a/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-3.txt b/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-3.txt index 174f140..74ab41b 100644 --- a/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-3.txt +++ b/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-3.txt @@ -1,6 +1,6 @@ { "body": { - "model": "gemini/gemini-2.5-pro-preview-05-06", + "model": "gemini/gemini-2.5-pro", "max_tokens": 65535, "temperature": 0, "messages": [ diff --git a/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-4.txt b/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-4.txt index 0e8d295..fba0955 100644 --- a/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-4.txt +++ b/e2e-tests/snapshots/context_manage.spec.ts_manage-context---smart-context-4.txt @@ -1,6 +1,6 @@ { "body": { - "model": "gemini/gemini-2.5-pro-preview-05-06", + "model": "gemini/gemini-2.5-pro", "max_tokens": 65535, "temperature": 0, "messages": [ diff --git a/e2e-tests/snapshots/engine.spec.ts_auto-should-send-message-to-gateway-1.txt b/e2e-tests/snapshots/engine.spec.ts_auto-should-send-message-to-gateway-1.txt index cd3f45e..d619875 100644 --- a/e2e-tests/snapshots/engine.spec.ts_auto-should-send-message-to-gateway-1.txt +++ b/e2e-tests/snapshots/engine.spec.ts_auto-should-send-message-to-gateway-1.txt @@ -1,6 +1,6 @@ { "body": { - "model": "gemini/gemini-2.5-flash-preview-05-20", + "model": "gemini/gemini-2.5-flash", "max_tokens": 8000, "temperature": 0, "thinking": { diff --git a/e2e-tests/snapshots/engine.spec.ts_send-message-to-engine-1.txt b/e2e-tests/snapshots/engine.spec.ts_send-message-to-engine-1.txt index 40ab6a9..ac2eec7 100644 --- a/e2e-tests/snapshots/engine.spec.ts_send-message-to-engine-1.txt +++ b/e2e-tests/snapshots/engine.spec.ts_send-message-to-engine-1.txt @@ -1,6 +1,6 @@ { "body": { - "model": "gemini/gemini-2.5-pro-preview-05-06", + "model": "gemini/gemini-2.5-pro", "max_tokens": 65535, "temperature": 0, "messages": [ diff --git a/e2e-tests/snapshots/gateway.spec.ts_gemini-2-5-flash-1.txt b/e2e-tests/snapshots/gateway.spec.ts_gemini-2-5-flash-1.txt index 5bcff44..5030498 100644 --- a/e2e-tests/snapshots/gateway.spec.ts_gemini-2-5-flash-1.txt +++ b/e2e-tests/snapshots/gateway.spec.ts_gemini-2-5-flash-1.txt @@ -1,6 +1,6 @@ { "body": { - "model": "gemini/gemini-2.5-flash-preview-05-20", + "model": "gemini/gemini-2.5-flash", "max_tokens": 65535, "temperature": 0, "thinking": { diff --git a/src/ipc/shared/language_model_helpers.ts b/src/ipc/shared/language_model_helpers.ts index 3ecc743..1365dfd 100644 --- a/src/ipc/shared/language_model_helpers.ts +++ b/src/ipc/shared/language_model_helpers.ts @@ -102,9 +102,9 @@ export const MODEL_OPTIONS: Record = { google: [ // https://ai.google.dev/gemini-api/docs/models#gemini-2.5-pro-preview-03-25 { - name: "gemini-2.5-pro-preview-05-06", + name: "gemini-2.5-pro", displayName: "Gemini 2.5 Pro", - description: "Preview version of Google's Gemini 2.5 Pro model", + description: "Google's Gemini 2.5 Pro model", // See Flash 2.5 comment below (go 1 below just to be safe, even though it seems OK now). maxOutputTokens: 65_536 - 1, // Gemini context window = input token + output token @@ -113,10 +113,9 @@ export const MODEL_OPTIONS: Record = { }, // https://ai.google.dev/gemini-api/docs/models#gemini-2.5-flash-preview { - name: "gemini-2.5-flash-preview-05-20", + name: "gemini-2.5-flash", displayName: "Gemini 2.5 Flash", - description: - "Preview version of Google's Gemini 2.5 Flash model (free tier available)", + description: "Google's Gemini 2.5 Flash model (free tier available)", // Weirdly for Vertex AI, the output token limit is *exclusive* of the stated limit. maxOutputTokens: 65_536 - 1, // Gemini context window = input token + output token diff --git a/src/ipc/utils/get_model_client.ts b/src/ipc/utils/get_model_client.ts index 717c08c..a179cff 100644 --- a/src/ipc/utils/get_model_client.ts +++ b/src/ipc/utils/get_model_client.ts @@ -20,11 +20,11 @@ const dyadGatewayUrl = process.env.DYAD_GATEWAY_URL; const AUTO_MODELS = [ { provider: "google", - name: "gemini-2.5-flash-preview-05-20", + name: "gemini-2.5-flash", }, { provider: "anthropic", - name: "claude-3-7-sonnet-latest", + name: "claude-sonnet-4-20250514", }, { provider: "openai",