AIX: remove llms's legacy VChatMessage and llmStreamingChatGenerate

This commit is contained in:
Enrico Ros
2024-12-10 19:53:28 -08:00
parent 2506d60058
commit 4183e1e2c5
+3 -46
View File
@@ -2,42 +2,17 @@ import { sendGAEvent } from '@next/third-parties/google';
import { hasGoogleAnalytics } from '~/common/components/GoogleAnalytics';
import type { OpenAIWire_Tools } from '~/modules/aix/server/dispatch/wiretypes/openai.wiretypes';
import type { DModelsService, DModelsServiceId } from '~/common/stores/llms/modelsservice.types';
import { DLLM, DLLMId, LLM_IF_OAI_Chat } from '~/common/stores/llms/llms.types';
import { llmsStoreActions } from '~/common/stores/llms/store-llms';
import { DLLM, LLM_IF_OAI_Chat } from '~/common/stores/llms/llms.types';
import { isModelPricingFree } from '~/common/stores/llms/llms.pricing';
import { llmsStoreActions } from '~/common/stores/llms/store-llms';
import type { ModelDescriptionSchema } from './server/llm.server.types';
import { DOpenAILLMOptions, FALLBACK_LLM_TEMPERATURE } from './vendors/openai/openai.vendor';
import { findServiceAccessOrThrow } from './vendors/vendor.helpers';
// LLM Client Types
// NOTE: Model List types in '../server/llm.server.types';
export interface VChatMessageIn {
role: 'assistant' | 'system' | 'user'; // | 'function';
content: string;
//name?: string; // when role: 'function'
}
export type VChatFunctionIn = OpenAIWire_Tools.FunctionDefinition;
export interface VChatMessageOut {
role: 'assistant' | 'system' | 'user';
content: string;
finish_reason: 'stop' | 'length' | null;
}
export interface VChatMessageOrFunctionCallOut extends VChatMessageOut {
function_name: string;
function_arguments: object | null;
}
// LLM Client Functions
// LLM Model Updates Client Functions
export async function llmsUpdateModelsForServiceOrThrow(serviceId: DModelsServiceId, keepUserEdits: boolean): Promise<{ models: ModelDescriptionSchema[] }> {
@@ -123,21 +98,3 @@ function _createDLLMFromModelDescription(d: ModelDescriptionSchema, service: DMo
return dllm;
}
export async function llmStreamingChatGenerate<
TServiceSettings extends object = {},
TAccess = undefined,
TLLMOptions = unknown
>(
llmId: DLLMId,
messages: VChatMessageIn[],
contextName: string,
contextRef: string | null,
functions: VChatFunctionIn[] | null,
forceFunctionName: string | null,
abortSignal: AbortSignal,
onUpdate: (update: any, done: boolean) => void,
): Promise<void> {
throw new Error(`llmStreamingChatGenerate: ${contextName} not migrated to AIX yet.`);
}