|
@@ -2,7 +2,7 @@ import { NextRequest, NextResponse } from "next/server";
|
|
|
import { getServerSideConfig } from "../config/server";
|
|
import { getServerSideConfig } from "../config/server";
|
|
|
import { OPENAI_BASE_URL, ServiceProvider } from "../constant";
|
|
import { OPENAI_BASE_URL, ServiceProvider } from "../constant";
|
|
|
import { cloudflareAIGatewayUrl } from "../utils/cloudflare";
|
|
import { cloudflareAIGatewayUrl } from "../utils/cloudflare";
|
|
|
-import { getModelProvider, isModelAvailableInServer } from "../utils/model";
|
|
|
|
|
|
|
+import { getModelProvider, isModelNotavailableInServer } from "../utils/model";
|
|
|
|
|
|
|
|
const serverConfig = getServerSideConfig();
|
|
const serverConfig = getServerSideConfig();
|
|
|
|
|
|
|
@@ -118,15 +118,14 @@ export async function requestOpenai(req: NextRequest) {
|
|
|
|
|
|
|
|
// not undefined and is false
|
|
// not undefined and is false
|
|
|
if (
|
|
if (
|
|
|
- isModelAvailableInServer(
|
|
|
|
|
|
|
+ isModelNotavailableInServer(
|
|
|
serverConfig.customModels,
|
|
serverConfig.customModels,
|
|
|
jsonBody?.model as string,
|
|
jsonBody?.model as string,
|
|
|
- ServiceProvider.OpenAI as string,
|
|
|
|
|
- ) ||
|
|
|
|
|
- isModelAvailableInServer(
|
|
|
|
|
- serverConfig.customModels,
|
|
|
|
|
- jsonBody?.model as string,
|
|
|
|
|
- ServiceProvider.Azure as string,
|
|
|
|
|
|
|
+ [
|
|
|
|
|
+ ServiceProvider.OpenAI,
|
|
|
|
|
+ ServiceProvider.Azure,
|
|
|
|
|
+ jsonBody?.model as string, // support provider-unspecified model
|
|
|
|
|
+ ],
|
|
|
)
|
|
)
|
|
|
) {
|
|
) {
|
|
|
return NextResponse.json(
|
|
return NextResponse.json(
|