LLMs: OpenAI: Update models

This commit is contained in:
Enrico Ros
2026-02-24 23:24:32 -08:00
parent 26547dec0d
commit 2da3942ce2
@@ -384,9 +384,8 @@ export const _knownOpenAIChatModels: ManualMappings = [
symLink: 'gpt-5-nano-2025-08-07',
},
/// GPT-5.3 Codex - Released February 5, 2026 (API access coming soon)
/// GPT-5.3 Codex - Released February 5, 2026
{
hidden: true, // API access not yet available - rolling out soon
idPrefix: 'gpt-5.3-codex',
label: 'GPT-5.3 Codex',
description: 'Most capable agentic coding model. Combines frontier coding performance of GPT-5.2-Codex with reasoning and professional knowledge of GPT-5.2. ~25% faster.',
@@ -398,7 +397,24 @@ export const _knownOpenAIChatModels: ManualMappings = [
{ paramId: 'llmVndOaiWebSearchContext' },
{ paramId: 'llmForceNoStream' },
],
// chatPrice: { input: 1.75, cache: { cType: 'oai-ac', read: 0.175 }, output: 14 }, // estimated from gpt-5.2-codex pricing
chatPrice: { input: 1.75, cache: { cType: 'oai-ac', read: 0.175 }, output: 14 },
// benchmark: TBD
},
// GPT-5.3 Codex Spark - Research preview, text-only, optimized for real-time coding iteration
{
hidden: true, // Research preview, ChatGPT Pro only - API access limited to design partners
idPrefix: 'gpt-5.3-codex-spark',
label: 'GPT-5.3 Codex Spark',
description: 'Text-only research preview optimized for real-time coding iteration. Delivers 1000+ tokens/sec on low-latency hardware.',
contextWindow: 128000,
maxCompletionTokens: 16384,
interfaces: [LLM_IF_OAI_Responses, LLM_IF_OAI_Chat, LLM_IF_OAI_Fn, LLM_IF_OAI_Json, LLM_IF_OAI_PromptCaching, LLM_IF_OAI_Reasoning, LLM_IF_HOTFIX_NoTemperature, LLM_IF_HOTFIX_StripImages],
parameterSpecs: [
{ paramId: 'llmVndOaiEffort', enumValues: ['low', 'medium', 'high'] },
{ paramId: 'llmForceNoStream' },
],
// chatPrice: TBD - separate usage limits, not standard token pricing
// benchmark: TBD
},
@@ -655,6 +671,18 @@ export const _knownOpenAIChatModels: ManualMappings = [
/// GPT-Audio series - General availability audio models
// gpt-audio-1.5
{
idPrefix: 'gpt-audio-1.5',
label: 'GPT Audio 1.5',
description: 'Best voice model for audio in, audio out with Chat Completions. Accepts audio inputs and outputs.',
contextWindow: 128000,
maxCompletionTokens: 16384,
interfaces: IFS_GPT_AUDIO,
chatPrice: { input: 2.5, output: 10 },
// benchmark: TBD
},
// gpt-audio
{
idPrefix: 'gpt-audio-2025-08-28',
@@ -1005,6 +1033,7 @@ const openAIModelsDenyList: string[] = [
'4o-mini-realtime',
'gpt-realtime',
'gpt-realtime-mini',
'gpt-realtime-1.5',
// [OpenAI, 2025-03-11] FIXME: NOT YET SUPPORTED - "RESPONSES API"
'computer-use-preview', 'computer-use-preview-2025-03-11', // FIXME: support these
@@ -1041,6 +1070,9 @@ const openAIModelsDenyList: string[] = [
// Video models: /v1/videos
'sora-2-pro', 'sora-2',
// Safety/moderation models
'gpt-oss-safeguard',
// Moderation models
'omni-moderation-latest', 'omni-moderation-2024-09-26', 'text-moderation-latest',
];
@@ -1060,6 +1092,7 @@ export function openAIInjectVariants(acc: ModelDescriptionSchema[], model: Model
const _manualOrderingIdPrefixes = [
// GPT-5.3
'gpt-5.3-codex-spark',
'gpt-5.3-codex',
// GPT-5.2
'gpt-5.2-20',
@@ -1121,6 +1154,7 @@ const _manualOrderingIdPrefixes = [
'gpt-4.1-nano',
'gpt-4.1',
// 4o-derived?
'gpt-audio-1.5',
'gpt-audio-2',
'gpt-4o-audio-preview',
'gpt-audio-mini-',