|
|
|
@ -1,10 +1,20 @@
|
|
|
|
|
import * as plugins from './plugins.js';
|
|
|
|
|
import * as paths from './paths.js';
|
|
|
|
|
|
|
|
|
|
// Custom type definition for chat completion messages
|
|
|
|
|
export type TChatCompletionRequestMessage = {
|
|
|
|
|
role: "system" | "user" | "assistant";
|
|
|
|
|
content: string;
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
import { MultiModalModel } from './abstract.classes.multimodal.js';
|
|
|
|
|
|
|
|
|
|
export interface IOpenaiProviderOptions {
|
|
|
|
|
openaiToken: string;
|
|
|
|
|
chatModel?: string;
|
|
|
|
|
audioModel?: string;
|
|
|
|
|
visionModel?: string;
|
|
|
|
|
// Optionally add more model options (e.g., documentModel) if needed.
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
export class OpenAiProvider extends MultiModalModel {
|
|
|
|
@ -31,11 +41,14 @@ export class OpenAiProvider extends MultiModalModel {
|
|
|
|
|
// Create a TextDecoder to handle incoming chunks
|
|
|
|
|
const decoder = new TextDecoder();
|
|
|
|
|
let buffer = '';
|
|
|
|
|
let currentMessage: { role: string; content: string; } | null = null;
|
|
|
|
|
let currentMessage: {
|
|
|
|
|
role: "function" | "user" | "system" | "assistant" | "tool" | "developer";
|
|
|
|
|
content: string;
|
|
|
|
|
} | null = null;
|
|
|
|
|
|
|
|
|
|
// Create a TransformStream to process the input
|
|
|
|
|
const transform = new TransformStream<Uint8Array, string>({
|
|
|
|
|
async transform(chunk, controller) {
|
|
|
|
|
transform: async (chunk, controller) => {
|
|
|
|
|
buffer += decoder.decode(chunk, { stream: true });
|
|
|
|
|
|
|
|
|
|
// Try to parse complete JSON messages from the buffer
|
|
|
|
@ -50,7 +63,7 @@ export class OpenAiProvider extends MultiModalModel {
|
|
|
|
|
try {
|
|
|
|
|
const message = JSON.parse(line);
|
|
|
|
|
currentMessage = {
|
|
|
|
|
role: message.role || 'user',
|
|
|
|
|
role: (message.role || 'user') as "function" | "user" | "system" | "assistant" | "tool" | "developer",
|
|
|
|
|
content: message.content || '',
|
|
|
|
|
};
|
|
|
|
|
} catch (e) {
|
|
|
|
@ -61,9 +74,11 @@ export class OpenAiProvider extends MultiModalModel {
|
|
|
|
|
|
|
|
|
|
// If we have a complete message, send it to OpenAI
|
|
|
|
|
if (currentMessage) {
|
|
|
|
|
const messageToSend = { role: "user" as const, content: currentMessage.content };
|
|
|
|
|
const stream = await this.openAiApiClient.chat.completions.create({
|
|
|
|
|
model: 'gpt-4',
|
|
|
|
|
messages: [{ role: currentMessage.role, content: currentMessage.content }],
|
|
|
|
|
model: this.options.chatModel ?? 'o3-mini',
|
|
|
|
|
temperature: 0,
|
|
|
|
|
messages: [messageToSend],
|
|
|
|
|
stream: true,
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
@ -105,8 +120,8 @@ export class OpenAiProvider extends MultiModalModel {
|
|
|
|
|
}[];
|
|
|
|
|
}) {
|
|
|
|
|
const result = await this.openAiApiClient.chat.completions.create({
|
|
|
|
|
model: 'gpt-4o',
|
|
|
|
|
|
|
|
|
|
model: this.options.chatModel ?? 'o3-mini',
|
|
|
|
|
temperature: 0,
|
|
|
|
|
messages: [
|
|
|
|
|
{ role: 'system', content: optionsArg.systemMessage },
|
|
|
|
|
...optionsArg.messageHistory,
|
|
|
|
@ -122,7 +137,7 @@ export class OpenAiProvider extends MultiModalModel {
|
|
|
|
|
public async audio(optionsArg: { message: string }): Promise<NodeJS.ReadableStream> {
|
|
|
|
|
const done = plugins.smartpromise.defer<NodeJS.ReadableStream>();
|
|
|
|
|
const result = await this.openAiApiClient.audio.speech.create({
|
|
|
|
|
model: 'tts-1-hd',
|
|
|
|
|
model: this.options.audioModel ?? 'o3-mini',
|
|
|
|
|
input: optionsArg.message,
|
|
|
|
|
voice: 'nova',
|
|
|
|
|
response_format: 'mp3',
|
|
|
|
@ -163,8 +178,8 @@ export class OpenAiProvider extends MultiModalModel {
|
|
|
|
|
);
|
|
|
|
|
|
|
|
|
|
const result = await this.openAiApiClient.chat.completions.create({
|
|
|
|
|
model: 'gpt-4o',
|
|
|
|
|
// response_format: { type: "json_object" }, // not supported for now
|
|
|
|
|
model: this.options.chatModel ?? 'o3-mini',
|
|
|
|
|
temperature: 0,
|
|
|
|
|
messages: [
|
|
|
|
|
{ role: 'system', content: optionsArg.systemMessage },
|
|
|
|
|
...optionsArg.messageHistory,
|
|
|
|
@ -195,7 +210,8 @@ export class OpenAiProvider extends MultiModalModel {
|
|
|
|
|
|
|
|
|
|
public async vision(optionsArg: { image: Buffer; prompt: string }): Promise<string> {
|
|
|
|
|
const result = await this.openAiApiClient.chat.completions.create({
|
|
|
|
|
model: 'gpt-4-vision-preview',
|
|
|
|
|
model: this.options.visionModel ?? 'o3-mini',
|
|
|
|
|
temperature: 0,
|
|
|
|
|
messages: [
|
|
|
|
|
{
|
|
|
|
|
role: 'user',
|
|
|
|
@ -215,4 +231,4 @@ export class OpenAiProvider extends MultiModalModel {
|
|
|
|
|
|
|
|
|
|
return result.choices[0].message.content || '';
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|