refactors api transformers and adds oai->anthropic chat api translation
This commit is contained in:
+100
-34
@@ -83,17 +83,19 @@ const anthropicResponseHandler: ProxyResHandlerWithBody = async (
|
||||
body.proxy_note = `Prompts are logged on this proxy instance. See ${host} for more information.`;
|
||||
}
|
||||
|
||||
if (req.inboundApi === "openai") {
|
||||
req.log.info("Transforming Anthropic text to OpenAI format");
|
||||
body = transformAnthropicTextResponseToOpenAI(body, req);
|
||||
}
|
||||
|
||||
if (
|
||||
req.inboundApi === "anthropic-text" &&
|
||||
req.outboundApi === "anthropic-chat"
|
||||
) {
|
||||
req.log.info("Transforming Anthropic text to Anthropic chat format");
|
||||
body = transformAnthropicChatResponseToAnthropicText(body);
|
||||
switch (`${req.inboundApi}<-${req.outboundApi}`) {
|
||||
case "openai<-anthropic-text":
|
||||
req.log.info("Transforming Anthropic Text back to OpenAI format");
|
||||
body = transformAnthropicTextResponseToOpenAI(body, req);
|
||||
break;
|
||||
case "openai<-anthropic-chat":
|
||||
req.log.info("Transforming Anthropic Chat back to OpenAI format");
|
||||
body = transformAnthropicChatResponseToOpenAI(body);
|
||||
break;
|
||||
case "anthropic-text<-anthropic-chat":
|
||||
req.log.info("Transforming Anthropic Chat back to Anthropic chat format");
|
||||
body = transformAnthropicChatResponseToAnthropicText(body);
|
||||
break;
|
||||
}
|
||||
|
||||
if (req.tokenizerInfo) {
|
||||
@@ -103,17 +105,23 @@ const anthropicResponseHandler: ProxyResHandlerWithBody = async (
|
||||
res.status(200).json(body);
|
||||
};
|
||||
|
||||
function flattenChatResponse(
|
||||
content: { type: string; text: string }[]
|
||||
): string {
|
||||
return content
|
||||
.map((part: { type: string; text: string }) =>
|
||||
part.type === "text" ? part.text : ""
|
||||
)
|
||||
.join("\n");
|
||||
}
|
||||
|
||||
export function transformAnthropicChatResponseToAnthropicText(
|
||||
anthropicBody: Record<string, any>
|
||||
): Record<string, any> {
|
||||
return {
|
||||
type: "completion",
|
||||
id: "trans-" + anthropicBody.id,
|
||||
completion: anthropicBody.content
|
||||
.map((part: { type: string; text: string }) =>
|
||||
part.type === "text" ? part.text : ""
|
||||
)
|
||||
.join(""),
|
||||
id: "ant-" + anthropicBody.id,
|
||||
completion: flattenChatResponse(anthropicBody.content),
|
||||
stop_reason: anthropicBody.stop_reason,
|
||||
stop: anthropicBody.stop_sequence,
|
||||
model: anthropicBody.model,
|
||||
@@ -155,6 +163,28 @@ function transformAnthropicTextResponseToOpenAI(
|
||||
};
|
||||
}
|
||||
|
||||
function transformAnthropicChatResponseToOpenAI(
|
||||
anthropicBody: Record<string, any>
|
||||
): Record<string, any> {
|
||||
return {
|
||||
id: "ant-" + anthropicBody.id,
|
||||
object: "chat.completion",
|
||||
created: Date.now(),
|
||||
model: anthropicBody.model,
|
||||
usage: anthropicBody.usage,
|
||||
choices: [
|
||||
{
|
||||
message: {
|
||||
role: "assistant",
|
||||
content: flattenChatResponse(anthropicBody.content),
|
||||
},
|
||||
finish_reason: anthropicBody.stop_reason,
|
||||
index: 0,
|
||||
},
|
||||
],
|
||||
};
|
||||
}
|
||||
|
||||
const anthropicProxy = createQueueMiddleware({
|
||||
proxyMiddleware: createProxyMiddleware({
|
||||
target: "https://api.anthropic.com",
|
||||
@@ -178,6 +208,9 @@ const anthropicProxy = createQueueMiddleware({
|
||||
if (isText && pathname === "/v1/chat/completions") {
|
||||
req.url = "/v1/complete";
|
||||
}
|
||||
if (isChat && pathname === "/v1/chat/completions") {
|
||||
req.url = "/v1/messages";
|
||||
}
|
||||
if (isChat && ["sonnet", "opus"].includes(req.params.type)) {
|
||||
req.url = "/v1/messages";
|
||||
}
|
||||
@@ -202,7 +235,7 @@ const textToChatPreprocessor = createPreprocessorMiddleware({
|
||||
* Routes text completion prompts to anthropic-chat if they need translation
|
||||
* (claude-3 based models do not support the old text completion endpoint).
|
||||
*/
|
||||
const claudeTextCompletionRouter: RequestHandler = (req, res, next) => {
|
||||
const preprocessAnthropicTextRequest: RequestHandler = (req, res, next) => {
|
||||
if (req.body.model?.startsWith("claude-3")) {
|
||||
textToChatPreprocessor(req, res, next);
|
||||
} else {
|
||||
@@ -210,15 +243,33 @@ const claudeTextCompletionRouter: RequestHandler = (req, res, next) => {
|
||||
}
|
||||
};
|
||||
|
||||
const oaiToTextPreprocessor = createPreprocessorMiddleware({
|
||||
inApi: "openai",
|
||||
outApi: "anthropic-text",
|
||||
service: "anthropic",
|
||||
});
|
||||
|
||||
const oaiToChatPreprocessor = createPreprocessorMiddleware({
|
||||
inApi: "openai",
|
||||
outApi: "anthropic-chat",
|
||||
service: "anthropic",
|
||||
});
|
||||
|
||||
/**
|
||||
* Routes an OpenAI prompt to either the legacy Claude text completion endpoint
|
||||
* or the new Claude chat completion endpoint, based on the requested model.
|
||||
*/
|
||||
const preprocessOpenAICompatRequest: RequestHandler = (req, res, next) => {
|
||||
maybeReassignModel(req);
|
||||
if (req.body.model?.includes("claude-3")) {
|
||||
oaiToChatPreprocessor(req, res, next);
|
||||
} else {
|
||||
oaiToTextPreprocessor(req, res, next);
|
||||
}
|
||||
};
|
||||
|
||||
const anthropicRouter = Router();
|
||||
anthropicRouter.get("/v1/models", handleModelRequest);
|
||||
// Anthropic text completion endpoint. Dynamic routing based on model.
|
||||
anthropicRouter.post(
|
||||
"/v1/complete",
|
||||
ipLimiter,
|
||||
claudeTextCompletionRouter,
|
||||
anthropicProxy
|
||||
);
|
||||
// Native Anthropic chat completion endpoint.
|
||||
anthropicRouter.post(
|
||||
"/v1/messages",
|
||||
@@ -230,23 +281,30 @@ anthropicRouter.post(
|
||||
}),
|
||||
anthropicProxy
|
||||
);
|
||||
// OpenAI-to-Anthropic Text compatibility endpoint.
|
||||
// Anthropic text completion endpoint. Translates to Anthropic chat completion
|
||||
// if the requested model is a Claude 3 model.
|
||||
anthropicRouter.post(
|
||||
"/v1/complete",
|
||||
ipLimiter,
|
||||
preprocessAnthropicTextRequest,
|
||||
anthropicProxy
|
||||
);
|
||||
// OpenAI-to-Anthropic compatibility endpoint. Accepts an OpenAI chat completion
|
||||
// request and transforms/routes it to the appropriate Anthropic format and
|
||||
// endpoint based on the requested model.
|
||||
anthropicRouter.post(
|
||||
"/v1/chat/completions",
|
||||
ipLimiter,
|
||||
createPreprocessorMiddleware(
|
||||
{ inApi: "openai", outApi: "anthropic-text", service: "anthropic" },
|
||||
{ afterTransform: [maybeReassignModel] }
|
||||
),
|
||||
preprocessOpenAICompatRequest,
|
||||
anthropicProxy
|
||||
);
|
||||
// Temporary force Anthropic Text to Anthropic Chat for frontends which do not
|
||||
// Temporarily force Anthropic Text to Anthropic Chat for frontends which do not
|
||||
// yet support the new model. Forces claude-3. Will be removed once common
|
||||
// frontends have been updated.
|
||||
anthropicRouter.post(
|
||||
"/v1/:type(sonnet|opus)/:action(complete|messages)",
|
||||
ipLimiter,
|
||||
handleCompatibilityRequest,
|
||||
handleAnthropicTextCompatRequest,
|
||||
createPreprocessorMiddleware({
|
||||
inApi: "anthropic-text",
|
||||
outApi: "anthropic-chat",
|
||||
@@ -255,7 +313,11 @@ anthropicRouter.post(
|
||||
anthropicProxy
|
||||
);
|
||||
|
||||
function handleCompatibilityRequest(req: Request, res: Response, next: any) {
|
||||
function handleAnthropicTextCompatRequest(
|
||||
req: Request,
|
||||
res: Response,
|
||||
next: any
|
||||
) {
|
||||
const type = req.params.type;
|
||||
const action = req.params.action;
|
||||
const alreadyInChatFormat = Boolean(req.body.messages);
|
||||
@@ -287,10 +349,14 @@ function handleCompatibilityRequest(req: Request, res: Response, next: any) {
|
||||
next();
|
||||
}
|
||||
|
||||
/**
|
||||
* If a client using the OpenAI compatibility endpoint requests an actual OpenAI
|
||||
* model, reassigns it to Claude 3 Sonnet.
|
||||
*/
|
||||
function maybeReassignModel(req: Request) {
|
||||
const model = req.body.model;
|
||||
if (!model.startsWith("gpt-")) return;
|
||||
req.body.model = "claude-2.1";
|
||||
req.body.model = "claude-3-sonnet-20240229";
|
||||
}
|
||||
|
||||
export const anthropic = anthropicRouter;
|
||||
|
||||
@@ -5,7 +5,7 @@ import { HttpRequest } from "@smithy/protocol-http";
|
||||
import {
|
||||
AnthropicV1TextSchema,
|
||||
AnthropicV1MessagesSchema,
|
||||
} from "../../../../shared/api-schemas/anthropic";
|
||||
} from "../../../../shared/api-schemas";
|
||||
import { keyPool } from "../../../../shared/key-management";
|
||||
import { RequestPreprocessor } from "../index";
|
||||
|
||||
|
||||
@@ -1,12 +1,9 @@
|
||||
import {
|
||||
anthropicTextToAnthropicChat,
|
||||
openAIToAnthropicText,
|
||||
} from "../../../../shared/api-schemas/anthropic";
|
||||
import { openAIToOpenAIText } from "../../../../shared/api-schemas/openai-text";
|
||||
import { openAIToOpenAIImage } from "../../../../shared/api-schemas/openai-image";
|
||||
import { openAIToGoogleAI } from "../../../../shared/api-schemas/google-ai";
|
||||
API_REQUEST_VALIDATORS,
|
||||
API_REQUEST_TRANSFORMERS,
|
||||
} from "../../../../shared/api-schemas";
|
||||
import { BadRequestError } from "../../../../shared/errors";
|
||||
import { fixMistralPrompt } from "../../../../shared/api-schemas/mistral-ai";
|
||||
import { API_SCHEMA_VALIDATORS } from "../../../../shared/api-schemas";
|
||||
import {
|
||||
isImageGenerationRequest,
|
||||
isTextGenerationRequest,
|
||||
@@ -22,6 +19,7 @@ export const transformOutboundPayload: RequestPreprocessor = async (req) => {
|
||||
|
||||
if (alreadyTransformed || notTransformable) return;
|
||||
|
||||
// TODO: this should be an APIFormatTransformer
|
||||
if (req.inboundApi === "mistral-ai") {
|
||||
const messages = req.body.messages;
|
||||
req.body.messages = fixMistralPrompt(messages);
|
||||
@@ -32,9 +30,9 @@ export const transformOutboundPayload: RequestPreprocessor = async (req) => {
|
||||
}
|
||||
|
||||
if (sameService) {
|
||||
const result = API_SCHEMA_VALIDATORS[req.inboundApi].safeParse(req.body);
|
||||
const result = API_REQUEST_VALIDATORS[req.inboundApi].safeParse(req.body);
|
||||
if (!result.success) {
|
||||
req.log.error(
|
||||
req.log.warn(
|
||||
{ issues: result.error.issues, body: req.body },
|
||||
"Request validation failed"
|
||||
);
|
||||
@@ -44,35 +42,16 @@ export const transformOutboundPayload: RequestPreprocessor = async (req) => {
|
||||
return;
|
||||
}
|
||||
|
||||
if (
|
||||
req.inboundApi === "anthropic-text" &&
|
||||
req.outboundApi === "anthropic-chat"
|
||||
) {
|
||||
req.body = anthropicTextToAnthropicChat(req);
|
||||
const transformation = `${req.inboundApi}->${req.outboundApi}` as const;
|
||||
const transFn = API_REQUEST_TRANSFORMERS[transformation];
|
||||
|
||||
if (transFn) {
|
||||
req.log.info({ transformation }, "Transforming request");
|
||||
req.body = await transFn(req);
|
||||
return;
|
||||
}
|
||||
|
||||
if (req.inboundApi === "openai" && req.outboundApi === "anthropic-text") {
|
||||
req.body = openAIToAnthropicText(req);
|
||||
return;
|
||||
}
|
||||
|
||||
if (req.inboundApi === "openai" && req.outboundApi === "google-ai") {
|
||||
req.body = openAIToGoogleAI(req);
|
||||
return;
|
||||
}
|
||||
|
||||
if (req.inboundApi === "openai" && req.outboundApi === "openai-text") {
|
||||
req.body = openAIToOpenAIText(req);
|
||||
return;
|
||||
}
|
||||
|
||||
if (req.inboundApi === "openai" && req.outboundApi === "openai-image") {
|
||||
req.body = openAIToOpenAIImage(req);
|
||||
return;
|
||||
}
|
||||
|
||||
throw new Error(
|
||||
`'${req.inboundApi}' -> '${req.outboundApi}' request proxying is not supported. Make sure your client is configured to use the correct API.`
|
||||
throw new BadRequestError(
|
||||
`${transformation} proxying is not supported. Make sure your client is configured to send requests in the correct format and to the correct endpoint.`
|
||||
);
|
||||
};
|
||||
|
||||
@@ -39,6 +39,7 @@ export { openAITextToOpenAIChat } from "./transformers/openai-text-to-openai";
|
||||
export { anthropicV1ToOpenAI } from "./transformers/anthropic-v1-to-openai";
|
||||
export { anthropicV2ToOpenAI } from "./transformers/anthropic-v2-to-openai";
|
||||
export { anthropicChatToAnthropicV2 } from "./transformers/anthropic-chat-to-anthropic-v2";
|
||||
export { anthropicChatToOpenAI } from "./transformers/anthropic-chat-to-openai";
|
||||
export { googleAIToOpenAI } from "./transformers/google-ai-to-openai";
|
||||
export { passthroughToOpenAI } from "./transformers/passthrough-to-openai";
|
||||
export { mergeEventsForOpenAIChat } from "./aggregators/openai-chat";
|
||||
|
||||
@@ -3,6 +3,7 @@ import { logger } from "../../../../logger";
|
||||
import { APIFormat } from "../../../../shared/key-management";
|
||||
import { assertNever } from "../../../../shared/utils";
|
||||
import {
|
||||
anthropicChatToOpenAI,
|
||||
anthropicChatToAnthropicV2,
|
||||
anthropicV1ToOpenAI,
|
||||
AnthropicV2StreamEvent,
|
||||
@@ -117,7 +118,11 @@ function eventIsOpenAIEvent(
|
||||
|
||||
function getTransformer(
|
||||
responseApi: APIFormat,
|
||||
version?: string
|
||||
version?: string,
|
||||
// There's only one case where we're not transforming back to OpenAI, which is
|
||||
// Anthropic Chat response -> Anthropic Text request. This parameter is only
|
||||
// used for that case.
|
||||
requestApi: APIFormat = "openai"
|
||||
): StreamingCompletionTransformer<
|
||||
OpenAIChatCompletionStreamEvent | AnthropicV2StreamEvent
|
||||
> {
|
||||
@@ -132,7 +137,9 @@ function getTransformer(
|
||||
? anthropicV1ToOpenAI
|
||||
: anthropicV2ToOpenAI;
|
||||
case "anthropic-chat":
|
||||
return anthropicChatToAnthropicV2;
|
||||
return requestApi === "anthropic-text"
|
||||
? anthropicChatToAnthropicV2
|
||||
: anthropicChatToOpenAI;
|
||||
case "google-ai":
|
||||
return googleAIToOpenAI;
|
||||
case "openai-image":
|
||||
|
||||
Reference in New Issue
Block a user