diff --git a/src/proxy/middleware/request/preprocessors/validate-context-size.ts b/src/proxy/middleware/request/preprocessors/validate-context-size.ts index 9d7cab5..887b039 100644 --- a/src/proxy/middleware/request/preprocessors/validate-context-size.ts +++ b/src/proxy/middleware/request/preprocessors/validate-context-size.ts @@ -7,7 +7,7 @@ import { RequestPreprocessor } from "../index"; const CLAUDE_MAX_CONTEXT = config.maxContextTokensAnthropic; const OPENAI_MAX_CONTEXT = config.maxContextTokensOpenAI; // todo: make configurable -const GOOGLE_AI_MAX_CONTEXT = 1024000; +const GOOGLE_AI_MAX_CONTEXT = 2048000; const MISTRAL_AI_MAX_CONTENT = 131072; /** @@ -68,6 +68,12 @@ export const validateContextSize: RequestPreprocessor = async (req) => { modelMax = 131072; } else if (model.match(/^gpt-4(-\d{4})?-vision(-preview)?$/)) { modelMax = 131072; + } else if (model.match(/^o1(-\d{4}-\d{2}-\d{2})?$/)) { + modelMax = 200000; + } else if (model.match(/^o1-mini(-\d{4}-\d{2}-\d{2})?$/)) { + modelMax = 128000; + } else if (model.match(/^o1-preview(-\d{4}-\d{2}-\d{2})?$/)) { + modelMax = 128000; } else if (model.match(/gpt-3.5-turbo/)) { modelMax = 16384; } else if (model.match(/gpt-4-32k/)) {