From 2da3942ce29efcee27afca5823d49b28890bfa9a Mon Sep 17 00:00:00 2001 From: Enrico Ros Date: Tue, 24 Feb 2026 23:24:32 -0800 Subject: [PATCH] LLMs: OpenAI: Update models --- .../server/openai/models/openai.models.ts | 40 +++++++++++++++++-- 1 file changed, 37 insertions(+), 3 deletions(-) diff --git a/src/modules/llms/server/openai/models/openai.models.ts b/src/modules/llms/server/openai/models/openai.models.ts index af7f0b6ff..399a81c2c 100644 --- a/src/modules/llms/server/openai/models/openai.models.ts +++ b/src/modules/llms/server/openai/models/openai.models.ts @@ -384,9 +384,8 @@ export const _knownOpenAIChatModels: ManualMappings = [ symLink: 'gpt-5-nano-2025-08-07', }, - /// GPT-5.3 Codex - Released February 5, 2026 (API access coming soon) + /// GPT-5.3 Codex - Released February 5, 2026 { - hidden: true, // API access not yet available - rolling out soon idPrefix: 'gpt-5.3-codex', label: 'GPT-5.3 Codex', description: 'Most capable agentic coding model. Combines frontier coding performance of GPT-5.2-Codex with reasoning and professional knowledge of GPT-5.2. ~25% faster.', @@ -398,7 +397,24 @@ export const _knownOpenAIChatModels: ManualMappings = [ { paramId: 'llmVndOaiWebSearchContext' }, { paramId: 'llmForceNoStream' }, ], - // chatPrice: { input: 1.75, cache: { cType: 'oai-ac', read: 0.175 }, output: 14 }, // estimated from gpt-5.2-codex pricing + chatPrice: { input: 1.75, cache: { cType: 'oai-ac', read: 0.175 }, output: 14 }, + // benchmark: TBD + }, + + // GPT-5.3 Codex Spark - Research preview, text-only, optimized for real-time coding iteration + { + hidden: true, // Research preview, ChatGPT Pro only - API access limited to design partners + idPrefix: 'gpt-5.3-codex-spark', + label: 'GPT-5.3 Codex Spark', + description: 'Text-only research preview optimized for real-time coding iteration. Delivers 1000+ tokens/sec on low-latency hardware.', + contextWindow: 128000, + maxCompletionTokens: 16384, + interfaces: [LLM_IF_OAI_Responses, LLM_IF_OAI_Chat, LLM_IF_OAI_Fn, LLM_IF_OAI_Json, LLM_IF_OAI_PromptCaching, LLM_IF_OAI_Reasoning, LLM_IF_HOTFIX_NoTemperature, LLM_IF_HOTFIX_StripImages], + parameterSpecs: [ + { paramId: 'llmVndOaiEffort', enumValues: ['low', 'medium', 'high'] }, + { paramId: 'llmForceNoStream' }, + ], + // chatPrice: TBD - separate usage limits, not standard token pricing // benchmark: TBD }, @@ -655,6 +671,18 @@ export const _knownOpenAIChatModels: ManualMappings = [ /// GPT-Audio series - General availability audio models + // gpt-audio-1.5 + { + idPrefix: 'gpt-audio-1.5', + label: 'GPT Audio 1.5', + description: 'Best voice model for audio in, audio out with Chat Completions. Accepts audio inputs and outputs.', + contextWindow: 128000, + maxCompletionTokens: 16384, + interfaces: IFS_GPT_AUDIO, + chatPrice: { input: 2.5, output: 10 }, + // benchmark: TBD + }, + // gpt-audio { idPrefix: 'gpt-audio-2025-08-28', @@ -1005,6 +1033,7 @@ const openAIModelsDenyList: string[] = [ '4o-mini-realtime', 'gpt-realtime', 'gpt-realtime-mini', + 'gpt-realtime-1.5', // [OpenAI, 2025-03-11] FIXME: NOT YET SUPPORTED - "RESPONSES API" 'computer-use-preview', 'computer-use-preview-2025-03-11', // FIXME: support these @@ -1041,6 +1070,9 @@ const openAIModelsDenyList: string[] = [ // Video models: /v1/videos 'sora-2-pro', 'sora-2', + // Safety/moderation models + 'gpt-oss-safeguard', + // Moderation models 'omni-moderation-latest', 'omni-moderation-2024-09-26', 'text-moderation-latest', ]; @@ -1060,6 +1092,7 @@ export function openAIInjectVariants(acc: ModelDescriptionSchema[], model: Model const _manualOrderingIdPrefixes = [ // GPT-5.3 + 'gpt-5.3-codex-spark', 'gpt-5.3-codex', // GPT-5.2 'gpt-5.2-20', @@ -1121,6 +1154,7 @@ const _manualOrderingIdPrefixes = [ 'gpt-4.1-nano', 'gpt-4.1', // 4o-derived? + 'gpt-audio-1.5', 'gpt-audio-2', 'gpt-4o-audio-preview', 'gpt-audio-mini-',