fix(classes.smartai): use IOllamaModelOptions type for defaultOptions instead of inline type
This commit is contained in:
@@ -3,7 +3,7 @@ import * as plugins from './plugins.js';
|
|||||||
import { AnthropicProvider } from './provider.anthropic.js';
|
import { AnthropicProvider } from './provider.anthropic.js';
|
||||||
import { ElevenLabsProvider } from './provider.elevenlabs.js';
|
import { ElevenLabsProvider } from './provider.elevenlabs.js';
|
||||||
import { MistralProvider } from './provider.mistral.js';
|
import { MistralProvider } from './provider.mistral.js';
|
||||||
import { OllamaProvider } from './provider.ollama.js';
|
import { OllamaProvider, type IOllamaModelOptions } from './provider.ollama.js';
|
||||||
import { OpenAiProvider } from './provider.openai.js';
|
import { OpenAiProvider } from './provider.openai.js';
|
||||||
import { PerplexityProvider } from './provider.perplexity.js';
|
import { PerplexityProvider } from './provider.perplexity.js';
|
||||||
import { ExoProvider } from './provider.exo.js';
|
import { ExoProvider } from './provider.exo.js';
|
||||||
@@ -32,16 +32,7 @@ export interface ISmartAiOptions {
|
|||||||
baseUrl?: string;
|
baseUrl?: string;
|
||||||
model?: string;
|
model?: string;
|
||||||
visionModel?: string;
|
visionModel?: string;
|
||||||
defaultOptions?: {
|
defaultOptions?: IOllamaModelOptions;
|
||||||
num_ctx?: number;
|
|
||||||
temperature?: number;
|
|
||||||
top_k?: number;
|
|
||||||
top_p?: number;
|
|
||||||
repeat_penalty?: number;
|
|
||||||
num_predict?: number;
|
|
||||||
stop?: string[];
|
|
||||||
seed?: number;
|
|
||||||
};
|
|
||||||
defaultTimeout?: number;
|
defaultTimeout?: number;
|
||||||
};
|
};
|
||||||
elevenlabs?: {
|
elevenlabs?: {
|
||||||
|
|||||||
Reference in New Issue
Block a user