Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 6 additions & 0 deletions sample.config.toml
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,12 @@ API_KEY = ""
API_URL = ""
MODEL_NAME = ""

[MODELS.AZURE_OPENAI]
API_KEY = ""
ENDPOINT = ""
MODEL_NAME = ""
API_VERSION = ""

[MODELS.OLLAMA]
API_URL = "" # Ollama API URL - http://host.docker.internal:11434

Expand Down
22 changes: 20 additions & 2 deletions src/app/api/chat/route.ts
Original file line number Diff line number Diff line change
Expand Up @@ -14,12 +14,18 @@ import { chats, messages as messagesSchema } from '@/lib/db/schema';
import { and, eq, gt } from 'drizzle-orm';
import { getFileDetails } from '@/lib/utils/files';
import { BaseChatModel } from '@langchain/core/language_models/chat_models';
import { ChatOpenAI } from '@langchain/openai';
import { ChatOpenAI, AzureChatOpenAI } from '@langchain/openai';
import {
getCustomOpenaiApiKey,
getCustomOpenaiApiUrl,
getCustomOpenaiModelName,
} from '@/lib/config';
import {
getAzureOpenaiApiKey,
getAzureOpenaiEndpoint,
getAzureOpenaiModelName,
getAzureOpenaiApiVersion,
} from '@/lib/config';
import { searchHandlers } from '@/lib/search';

export const runtime = 'nodejs';
Expand Down Expand Up @@ -186,6 +192,8 @@ export const POST = async (req: Request) => {
const body = (await req.json()) as Body;
const { message } = body;

console.error('An error occurred while processing chat request:', "here");

if (message.content === '') {
return Response.json(
{
Expand Down Expand Up @@ -222,6 +230,7 @@ export const POST = async (req: Request) => {
let embedding = embeddingModel.model;

if (body.chatModel?.provider === 'custom_openai') {
console.error('An error occurred while processing chat request:', "custom_openai");
llm = new ChatOpenAI({
openAIApiKey: getCustomOpenaiApiKey(),
modelName: getCustomOpenaiModelName(),
Expand All @@ -230,6 +239,15 @@ export const POST = async (req: Request) => {
baseURL: getCustomOpenaiApiUrl(),
},
}) as unknown as BaseChatModel;
} else if (body.chatModel?.provider == 'azure_openai') {
console.error('An error occurred while processing chat request:', "azure_openai");
llm = new AzureChatOpenAI({
openAIApiKey: getAzureOpenaiApiKey(),
deploymentName: getAzureOpenaiModelName(),
openAIBasePath: getAzureOpenaiEndpoint(),
openAIApiVersion: getAzureOpenaiApiVersion(),
temperature: 0.7
}) as unknown as BaseChatModel
} else if (chatModelProvider && chatModel) {
llm = chatModel.model;
}
Expand Down Expand Up @@ -297,7 +315,7 @@ export const POST = async (req: Request) => {
},
});
} catch (err) {
console.error('An error occurred while processing chat request:', err);
console.error('An error occurred while processing chat request 123:', err);
return Response.json(
{ message: 'An error occurred while processing chat request' },
{ status: 500 },
Expand Down
14 changes: 14 additions & 0 deletions src/app/api/config/route.ts
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,10 @@ import {
getCustomOpenaiApiKey,
getCustomOpenaiApiUrl,
getCustomOpenaiModelName,
getAzureOpenaiApiKey,
getAzureOpenaiApiVersion,
getAzureOpenaiModelName,
getAzureOpenaiEndpoint,
getGeminiApiKey,
getGroqApiKey,
getOllamaApiEndpoint,
Expand Down Expand Up @@ -58,6 +62,10 @@ export const GET = async (req: Request) => {
config['customOpenaiApiUrl'] = getCustomOpenaiApiUrl();
config['customOpenaiApiKey'] = getCustomOpenaiApiKey();
config['customOpenaiModelName'] = getCustomOpenaiModelName();
config['azureOpenaiApiKey'] = getAzureOpenaiApiKey();
config['azureOpenaiApiVersion'] = getAzureOpenaiApiVersion();
config['azureOpenaiModelName'] = getAzureOpenaiModelName();
config['azureOpenaiEndpoint'] = getAzureOpenaiEndpoint();

return Response.json({ ...config }, { status: 200 });
} catch (err) {
Expand Down Expand Up @@ -98,6 +106,12 @@ export const POST = async (req: Request) => {
API_KEY: config.customOpenaiApiKey,
MODEL_NAME: config.customOpenaiModelName,
},
AZURE_OPENAI: {
API_KEY: config.azureOpenaiApiKey,
MODEL_NAME: config.azureOpenaiModelName,
ENDPOINT: config.azureOpenaiEndpoint,
API_VERSION: config.azureOpenaiApiVersion,
}
},
};

Expand Down
16 changes: 15 additions & 1 deletion src/app/api/images/route.ts
Original file line number Diff line number Diff line change
Expand Up @@ -4,10 +4,16 @@ import {
getCustomOpenaiApiUrl,
getCustomOpenaiModelName,
} from '@/lib/config';
import {
getAzureOpenaiApiKey,
getAzureOpenaiEndpoint,
getAzureOpenaiModelName,
getAzureOpenaiApiVersion,
} from '@/lib/config';
import { getAvailableChatModelProviders } from '@/lib/providers';
import { BaseChatModel } from '@langchain/core/language_models/chat_models';
import { AIMessage, BaseMessage, HumanMessage } from '@langchain/core/messages';
import { ChatOpenAI } from '@langchain/openai';
import { ChatOpenAI, AzureChatOpenAI } from '@langchain/openai';

interface ChatModel {
provider: string;
Expand Down Expand Up @@ -56,6 +62,14 @@ export const POST = async (req: Request) => {
baseURL: getCustomOpenaiApiUrl(),
},
}) as unknown as BaseChatModel;
} else if (body.chatModel?.provider == 'azure_openai') {
llm = new AzureChatOpenAI({
openAIApiKey: getAzureOpenaiApiKey(),
deploymentName: getAzureOpenaiModelName(),
openAIBasePath: getAzureOpenaiEndpoint(),
openAIApiVersion: getAzureOpenaiApiVersion(),
temperature: 0.7
}) as unknown as BaseChatModel
} else if (chatModelProvider && chatModel) {
llm = chatModel.model;
}
Expand Down
3 changes: 3 additions & 0 deletions src/app/api/models/route.ts
Original file line number Diff line number Diff line change
Expand Up @@ -10,13 +10,16 @@ export const GET = async (req: Request) => {
getAvailableEmbeddingModelProviders(),
]);



Object.keys(chatModelProviders).forEach((provider) => {
Object.keys(chatModelProviders[provider]).forEach((model) => {
delete (chatModelProviders[provider][model] as { model?: unknown })
.model;
});
});


Object.keys(embeddingModelProviders).forEach((provider) => {
Object.keys(embeddingModelProviders[provider]).forEach((model) => {
delete (embeddingModelProviders[provider][model] as { model?: unknown })
Expand Down
20 changes: 19 additions & 1 deletion src/app/api/search/route.ts
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
import type { BaseChatModel } from '@langchain/core/language_models/chat_models';
import type { Embeddings } from '@langchain/core/embeddings';
import { ChatOpenAI } from '@langchain/openai';
import { ChatOpenAI, AzureChatOpenAI } from '@langchain/openai';
import {
getAvailableChatModelProviders,
getAvailableEmbeddingModelProviders,
Expand All @@ -12,13 +12,23 @@ import {
getCustomOpenaiApiUrl,
getCustomOpenaiModelName,
} from '@/lib/config';
import {
getAzureOpenaiApiKey,
getAzureOpenaiEndpoint,
getAzureOpenaiModelName,
getAzureOpenaiApiVersion,
} from '@/lib/config';
import { searchHandlers } from '@/lib/search';

interface chatModel {
provider: string;
name: string;
customOpenAIKey?: string;
customOpenAIBaseURL?: string;
azureOpenAIApiVersion?: string;
azureOpenAIApiKey?: string;
azureOpenAIApiDeploymentName?: string;
azureOpenAIEndpoint?: string;
}

interface embeddingModel {
Expand Down Expand Up @@ -89,6 +99,14 @@ export const POST = async (req: Request) => {
body.chatModel?.customOpenAIBaseURL || getCustomOpenaiApiUrl(),
},
}) as unknown as BaseChatModel;
} else if (body.chatModel?.provider == 'azure_openai') {
llm = new AzureChatOpenAI({
openAIApiKey: body.chatModel?.azureOpenAIApiKey || getAzureOpenaiApiKey(),
deploymentName: body.chatModel?.azureOpenAIApiDeploymentName || getAzureOpenaiModelName(),
openAIBasePath: body.chatModel?.azureOpenAIEndpoint || getAzureOpenaiEndpoint(),
openAIApiVersion: body.chatModel?.azureOpenAIApiVersion || getAzureOpenaiApiVersion(),
temperature: 0.7
}) as unknown as BaseChatModel
} else if (
chatModelProviders[chatModelProvider] &&
chatModelProviders[chatModelProvider][chatModel]
Expand Down
16 changes: 15 additions & 1 deletion src/app/api/suggestions/route.ts
Original file line number Diff line number Diff line change
Expand Up @@ -4,10 +4,16 @@ import {
getCustomOpenaiApiUrl,
getCustomOpenaiModelName,
} from '@/lib/config';
import {
getAzureOpenaiApiKey,
getAzureOpenaiEndpoint,
getAzureOpenaiModelName,
getAzureOpenaiApiVersion,
} from '@/lib/config';
import { getAvailableChatModelProviders } from '@/lib/providers';
import { BaseChatModel } from '@langchain/core/language_models/chat_models';
import { AIMessage, BaseMessage, HumanMessage } from '@langchain/core/messages';
import { ChatOpenAI } from '@langchain/openai';
import { ChatOpenAI, AzureChatOpenAI } from '@langchain/openai';

interface ChatModel {
provider: string;
Expand Down Expand Up @@ -55,6 +61,14 @@ export const POST = async (req: Request) => {
baseURL: getCustomOpenaiApiUrl(),
},
}) as unknown as BaseChatModel;
} else if (body.chatModel?.provider == 'azure_openai') {
llm = new AzureChatOpenAI({
openAIApiKey: getAzureOpenaiApiKey(),
deploymentName: getAzureOpenaiModelName(),
openAIBasePath: getAzureOpenaiEndpoint(),
openAIApiVersion: getAzureOpenaiApiVersion(),
temperature: 0.7
}) as unknown as BaseChatModel
} else if (chatModelProvider && chatModel) {
llm = chatModel.model;
}
Expand Down
16 changes: 15 additions & 1 deletion src/app/api/videos/route.ts
Original file line number Diff line number Diff line change
Expand Up @@ -4,10 +4,16 @@ import {
getCustomOpenaiApiUrl,
getCustomOpenaiModelName,
} from '@/lib/config';
import {
getAzureOpenaiApiKey,
getAzureOpenaiEndpoint,
getAzureOpenaiModelName,
getAzureOpenaiApiVersion,
} from '@/lib/config';
import { getAvailableChatModelProviders } from '@/lib/providers';
import { BaseChatModel } from '@langchain/core/language_models/chat_models';
import { AIMessage, BaseMessage, HumanMessage } from '@langchain/core/messages';
import { ChatOpenAI } from '@langchain/openai';
import { ChatOpenAI, AzureChatOpenAI } from '@langchain/openai';

interface ChatModel {
provider: string;
Expand Down Expand Up @@ -56,6 +62,14 @@ export const POST = async (req: Request) => {
baseURL: getCustomOpenaiApiUrl(),
},
}) as unknown as BaseChatModel;
} else if (body.chatModel?.provider == 'azure_openai') {
llm = new AzureChatOpenAI({
openAIApiKey: getAzureOpenaiApiKey(),
deploymentName: getAzureOpenaiModelName(),
openAIBasePath: getAzureOpenaiEndpoint(),
openAIApiVersion: getAzureOpenaiApiVersion(),
temperature: 0.7
}) as unknown as BaseChatModel
} else if (chatModelProvider && chatModel) {
llm = chatModel.model;
}
Expand Down
94 changes: 93 additions & 1 deletion src/app/settings/page.tsx
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,10 @@ interface SettingsType {
customOpenaiApiKey: string;
customOpenaiApiUrl: string;
customOpenaiModelName: string;
azureOpenaiModelName: string;
azureOpenaiEndpoint: string;
azureOpenaiApiKey: string;
azureOpenaiApiVersion: string;
}

interface InputProps extends React.InputHTMLAttributes<HTMLInputElement> {
Expand Down Expand Up @@ -556,7 +560,8 @@ const Page = () => {
</div>

{selectedChatModelProvider &&
selectedChatModelProvider != 'custom_openai' && (
selectedChatModelProvider != 'custom_openai' &&
selectedChatModelProvider != 'azure_openai' && (
<div className="flex flex-col space-y-1">
<p className="text-black/70 dark:text-white/70 text-sm">
Chat Model
Expand Down Expand Up @@ -666,6 +671,93 @@ const Page = () => {
</div>
</div>
)}

{selectedChatModelProvider &&
selectedChatModelProvider === 'azure_openai' && (
<div className="flex flex-col space-y-4">
<div className="flex flex-col space-y-1">
<p className="text-black/70 dark:text-white/70 text-sm">
Model Name
</p>
<Input
type="text"
placeholder="Model name"
value={config.azureOpenaiModelName}
isSaving={savingStates['azureOpenaiModelName']}
onChange={(e: React.ChangeEvent<HTMLInputElement>) => {
setConfig((prev) => ({
...prev!,
azureOpenaiModelName: e.target.value,
}));
}}
onSave={(value) =>
saveConfig('azureOpenaiModelName', value)
}
/>
</div>
<div className="flex flex-col space-y-1">
<p className="text-black/70 dark:text-white/70 text-sm">
Azure OpenAI API Key
</p>
<Input
type="text"
placeholder="Azure OpenAI API Key"
value={config.azureOpenaiApiKey}
isSaving={savingStates['azureOpenaiApiKey']}
onChange={(e: React.ChangeEvent<HTMLInputElement>) => {
setConfig((prev) => ({
...prev!,
azureOpenaiApiKey: e.target.value,
}));
}}
onSave={(value) =>
saveConfig('azureOpenaiApiKey', value)
}
/>
</div>
<div className="flex flex-col space-y-1">
<p className="text-black/70 dark:text-white/70 text-sm">
Azure OpenAI Base URL
</p>
<Input
type="text"
placeholder="Azure OpenAI Base URL"
value={config.azureOpenaiEndpoint}
isSaving={savingStates['azureOpenaiEndpoint']}
onChange={(e: React.ChangeEvent<HTMLInputElement>) => {
setConfig((prev) => ({
...prev!,
azureOpenaiEndpoint: e.target.value,
}));
}}
onSave={(value) =>
saveConfig('azureOpenaiEndpoint', value)
}
/>
</div>
<div className="flex flex-col space-y-1">
<p className="text-black/70 dark:text-white/70 text-sm">
Azure OpenAI Api Version
</p>
<Input
type="text"
placeholder="Azure OpenAI Api Version"
value={config.azureOpenaiApiVersion}
isSaving={savingStates['azureOpenaiApiVersion']}
onChange={(e: React.ChangeEvent<HTMLInputElement>) => {
setConfig((prev) => ({
...prev!,
azureOpenaiApiVersion: e.target.value,
}));
}}
onSave={(value) =>
saveConfig('azureOpenaiApiVersion', value)
}
/>
</div>
</div>
)}


{config.embeddingModelProviders && (
<div className="flex flex-col space-y-4 mt-4 pt-4 border-t border-light-200 dark:border-dark-200">
Expand Down
Loading