mirror of
https://github.com/enricoros/big-AGI.git
synced 2026-05-10 21:50:14 -07:00
LLMs: OpenAI: Update models
This commit is contained in:
@@ -384,9 +384,8 @@ export const _knownOpenAIChatModels: ManualMappings = [
|
||||
symLink: 'gpt-5-nano-2025-08-07',
|
||||
},
|
||||
|
||||
/// GPT-5.3 Codex - Released February 5, 2026 (API access coming soon)
|
||||
/// GPT-5.3 Codex - Released February 5, 2026
|
||||
{
|
||||
hidden: true, // API access not yet available - rolling out soon
|
||||
idPrefix: 'gpt-5.3-codex',
|
||||
label: 'GPT-5.3 Codex',
|
||||
description: 'Most capable agentic coding model. Combines frontier coding performance of GPT-5.2-Codex with reasoning and professional knowledge of GPT-5.2. ~25% faster.',
|
||||
@@ -398,7 +397,24 @@ export const _knownOpenAIChatModels: ManualMappings = [
|
||||
{ paramId: 'llmVndOaiWebSearchContext' },
|
||||
{ paramId: 'llmForceNoStream' },
|
||||
],
|
||||
// chatPrice: { input: 1.75, cache: { cType: 'oai-ac', read: 0.175 }, output: 14 }, // estimated from gpt-5.2-codex pricing
|
||||
chatPrice: { input: 1.75, cache: { cType: 'oai-ac', read: 0.175 }, output: 14 },
|
||||
// benchmark: TBD
|
||||
},
|
||||
|
||||
// GPT-5.3 Codex Spark - Research preview, text-only, optimized for real-time coding iteration
|
||||
{
|
||||
hidden: true, // Research preview, ChatGPT Pro only - API access limited to design partners
|
||||
idPrefix: 'gpt-5.3-codex-spark',
|
||||
label: 'GPT-5.3 Codex Spark',
|
||||
description: 'Text-only research preview optimized for real-time coding iteration. Delivers 1000+ tokens/sec on low-latency hardware.',
|
||||
contextWindow: 128000,
|
||||
maxCompletionTokens: 16384,
|
||||
interfaces: [LLM_IF_OAI_Responses, LLM_IF_OAI_Chat, LLM_IF_OAI_Fn, LLM_IF_OAI_Json, LLM_IF_OAI_PromptCaching, LLM_IF_OAI_Reasoning, LLM_IF_HOTFIX_NoTemperature, LLM_IF_HOTFIX_StripImages],
|
||||
parameterSpecs: [
|
||||
{ paramId: 'llmVndOaiEffort', enumValues: ['low', 'medium', 'high'] },
|
||||
{ paramId: 'llmForceNoStream' },
|
||||
],
|
||||
// chatPrice: TBD - separate usage limits, not standard token pricing
|
||||
// benchmark: TBD
|
||||
},
|
||||
|
||||
@@ -655,6 +671,18 @@ export const _knownOpenAIChatModels: ManualMappings = [
|
||||
|
||||
/// GPT-Audio series - General availability audio models
|
||||
|
||||
// gpt-audio-1.5
|
||||
{
|
||||
idPrefix: 'gpt-audio-1.5',
|
||||
label: 'GPT Audio 1.5',
|
||||
description: 'Best voice model for audio in, audio out with Chat Completions. Accepts audio inputs and outputs.',
|
||||
contextWindow: 128000,
|
||||
maxCompletionTokens: 16384,
|
||||
interfaces: IFS_GPT_AUDIO,
|
||||
chatPrice: { input: 2.5, output: 10 },
|
||||
// benchmark: TBD
|
||||
},
|
||||
|
||||
// gpt-audio
|
||||
{
|
||||
idPrefix: 'gpt-audio-2025-08-28',
|
||||
@@ -1005,6 +1033,7 @@ const openAIModelsDenyList: string[] = [
|
||||
'4o-mini-realtime',
|
||||
'gpt-realtime',
|
||||
'gpt-realtime-mini',
|
||||
'gpt-realtime-1.5',
|
||||
|
||||
// [OpenAI, 2025-03-11] FIXME: NOT YET SUPPORTED - "RESPONSES API"
|
||||
'computer-use-preview', 'computer-use-preview-2025-03-11', // FIXME: support these
|
||||
@@ -1041,6 +1070,9 @@ const openAIModelsDenyList: string[] = [
|
||||
// Video models: /v1/videos
|
||||
'sora-2-pro', 'sora-2',
|
||||
|
||||
// Safety/moderation models
|
||||
'gpt-oss-safeguard',
|
||||
|
||||
// Moderation models
|
||||
'omni-moderation-latest', 'omni-moderation-2024-09-26', 'text-moderation-latest',
|
||||
];
|
||||
@@ -1060,6 +1092,7 @@ export function openAIInjectVariants(acc: ModelDescriptionSchema[], model: Model
|
||||
|
||||
const _manualOrderingIdPrefixes = [
|
||||
// GPT-5.3
|
||||
'gpt-5.3-codex-spark',
|
||||
'gpt-5.3-codex',
|
||||
// GPT-5.2
|
||||
'gpt-5.2-20',
|
||||
@@ -1121,6 +1154,7 @@ const _manualOrderingIdPrefixes = [
|
||||
'gpt-4.1-nano',
|
||||
'gpt-4.1',
|
||||
// 4o-derived?
|
||||
'gpt-audio-1.5',
|
||||
'gpt-audio-2',
|
||||
'gpt-4o-audio-preview',
|
||||
'gpt-audio-mini-',
|
||||
|
||||
Reference in New Issue
Block a user