feat(llm): resolve sending double headers in responses, and not being able to send requests to ollama

This commit is contained in:
perf3ct 2025-06-07 00:02:26 +00:00
parent 20ec294774
commit 6bc9b3c184
No known key found for this signature in database
GPG Key ID: 569C4EEC436F5232
6 changed files with 179 additions and 231 deletions

View File

@ -376,7 +376,7 @@ export default class AiSettingsWidget extends OptionsWidget {
embeddingWarnings.push(t("ai_llm.empty_key_warning.voyage")); embeddingWarnings.push(t("ai_llm.empty_key_warning.voyage"));
} }
if (selectedEmbeddingProvider === 'ollama' && !this.$widget.find('.ollama-base-url').val()) { if (selectedEmbeddingProvider === 'ollama' && !this.$widget.find('.ollama-embedding-base-url').val()) {
embeddingWarnings.push(t("ai_llm.empty_key_warning.ollama")); embeddingWarnings.push(t("ai_llm.empty_key_warning.ollama"));
} }
} }

View File

@ -240,40 +240,65 @@ export class ProviderService {
} }
try { try {
const ollamaBaseUrl = this.$widget.find('.ollama-base-url').val() as string; // Determine which URL to use based on the current context
// If we're in the embedding provider context, use the embedding base URL
// Otherwise, use the general base URL
const selectedAiProvider = this.$widget.find('.ai-selected-provider').val() as string;
const selectedEmbeddingProvider = this.$widget.find('.embedding-selected-provider').val() as string;
let ollamaBaseUrl: string;
// If embedding provider is Ollama and it's visible, use embedding URL
const $embeddingOllamaSettings = this.$widget.find('.ollama-embedding-provider-settings');
if (selectedEmbeddingProvider === 'ollama' && $embeddingOllamaSettings.is(':visible')) {
ollamaBaseUrl = this.$widget.find('.ollama-embedding-base-url').val() as string;
} else {
ollamaBaseUrl = this.$widget.find('.ollama-base-url').val() as string;
}
const response = await server.get<OllamaModelResponse>(`llm/providers/ollama/models?baseUrl=${encodeURIComponent(ollamaBaseUrl)}`); const response = await server.get<OllamaModelResponse>(`llm/providers/ollama/models?baseUrl=${encodeURIComponent(ollamaBaseUrl)}`);
if (response && response.success && response.models && response.models.length > 0) { if (response && response.success && response.models && response.models.length > 0) {
// Update both embedding model dropdowns
const $embedModelSelect = this.$widget.find('.ollama-embedding-model'); const $embedModelSelect = this.$widget.find('.ollama-embedding-model');
const $chatEmbedModelSelect = this.$widget.find('.ollama-chat-embedding-model');
const currentValue = $embedModelSelect.val(); const currentValue = $embedModelSelect.val();
const currentChatEmbedValue = $chatEmbedModelSelect.val();
// Clear existing options // Prepare embedding models
$embedModelSelect.empty();
// Add embedding-specific models first
const embeddingModels = response.models.filter(model => const embeddingModels = response.models.filter(model =>
model.name.includes('embed') || model.name.includes('bert')); model.name.includes('embed') || model.name.includes('bert'));
embeddingModels.forEach(model => {
$embedModelSelect.append(`<option value="${model.name}">${model.name}</option>`);
});
if (embeddingModels.length > 0) {
// Add separator if we have embedding models
$embedModelSelect.append(`<option disabled>─────────────</option>`);
}
// Then add general models which can be used for embeddings too
const generalModels = response.models.filter(model => const generalModels = response.models.filter(model =>
!model.name.includes('embed') && !model.name.includes('bert')); !model.name.includes('embed') && !model.name.includes('bert'));
// Update .ollama-embedding-model dropdown (embedding provider settings)
$embedModelSelect.empty();
embeddingModels.forEach(model => {
$embedModelSelect.append(`<option value="${model.name}">${model.name}</option>`);
});
if (embeddingModels.length > 0) {
$embedModelSelect.append(`<option disabled>─────────────</option>`);
}
generalModels.forEach(model => { generalModels.forEach(model => {
$embedModelSelect.append(`<option value="${model.name}">${model.name}</option>`); $embedModelSelect.append(`<option value="${model.name}">${model.name}</option>`);
}); });
// Try to restore the previously selected value
this.ensureSelectedValue($embedModelSelect, currentValue, 'ollamaEmbeddingModel'); this.ensureSelectedValue($embedModelSelect, currentValue, 'ollamaEmbeddingModel');
// Update .ollama-chat-embedding-model dropdown (general Ollama provider settings)
$chatEmbedModelSelect.empty();
embeddingModels.forEach(model => {
$chatEmbedModelSelect.append(`<option value="${model.name}">${model.name}</option>`);
});
if (embeddingModels.length > 0) {
$chatEmbedModelSelect.append(`<option disabled>─────────────</option>`);
}
generalModels.forEach(model => {
$chatEmbedModelSelect.append(`<option value="${model.name}">${model.name}</option>`);
});
this.ensureSelectedValue($chatEmbedModelSelect, currentChatEmbedValue, 'ollamaEmbeddingModel');
// Also update the LLM model dropdown // Also update the LLM model dropdown
const $modelSelect = this.$widget.find('.ollama-default-model'); const $modelSelect = this.$widget.find('.ollama-default-model');
const currentModelValue = $modelSelect.val(); const currentModelValue = $modelSelect.val();

View File

@ -825,7 +825,10 @@ async function streamMessage(req: Request, res: Response) {
success: true, success: true,
message: 'Streaming initiated successfully' message: 'Streaming initiated successfully'
}); });
log.info(`Sent immediate success response for streaming setup`);
// Mark response as handled to prevent apiResultHandler from processing it again
(res as any).triliumResponseHandled = true;
// Create a new response object for streaming through WebSocket only // Create a new response object for streaming through WebSocket only
// We won't use HTTP streaming since we've already sent the HTTP response // We won't use HTTP streaming since we've already sent the HTTP response
@ -889,42 +892,10 @@ async function streamMessage(req: Request, res: Response) {
thinking: showThinking ? 'Initializing streaming LLM response...' : undefined thinking: showThinking ? 'Initializing streaming LLM response...' : undefined
}); });
// Instead of trying to reimplement the streaming logic ourselves, // Process the LLM request using the existing service but with streaming setup
// delegate to restChatService but set up the correct protocol: // Since we've already sent the initial HTTP response, we'll use the WebSocket for streaming
// 1. We've already sent a success response to the initial POST
// 2. Now we'll have restChatService process the actual streaming through WebSocket
try { try {
// Import the WebSocket service for sending messages // Call restChatService with streaming mode enabled
const wsService = (await import('../../services/ws.js')).default;
// Create a simple pass-through response object that won't write to the HTTP response
// but will allow restChatService to send WebSocket messages
const dummyResponse = {
writableEnded: false,
// Implement methods that would normally be used by restChatService
write: (_chunk: string) => {
// Silent no-op - we're only using WebSocket
return true;
},
end: (_chunk?: string) => {
// Log when streaming is complete via WebSocket
log.info(`[${chatNoteId}] Completed HTTP response handling during WebSocket streaming`);
return dummyResponse;
},
setHeader: (name: string, _value: string) => {
// Only log for content-type to reduce noise
if (name.toLowerCase() === 'content-type') {
log.info(`[${chatNoteId}] Setting up streaming for WebSocket only`);
}
return dummyResponse;
}
};
// Process the streaming now through WebSocket only
try {
log.info(`[${chatNoteId}] Processing LLM streaming through WebSocket after successful initiation at ${new Date().toISOString()}`);
// Call restChatService with our enhanced request and dummy response
// The important part is setting method to GET to indicate streaming mode // The important part is setting method to GET to indicate streaming mode
await restChatService.handleSendMessage({ await restChatService.handleSendMessage({
...req, ...req,
@ -939,11 +910,9 @@ async function streamMessage(req: Request, res: Response) {
showThinking: showThinking === true showThinking: showThinking === true
}, },
params: { chatNoteId } params: { chatNoteId }
} as unknown as Request, dummyResponse as unknown as Response); } as unknown as Request, res);
log.info(`[${chatNoteId}] WebSocket streaming completed at ${new Date().toISOString()}`);
} catch (streamError) { } catch (streamError) {
log.error(`[${chatNoteId}] Error during WebSocket streaming: ${streamError}`); log.error(`Error during WebSocket streaming: ${streamError}`);
// Send error message through WebSocket // Send error message through WebSocket
wsService.sendMessageToAllClients({ wsService.sendMessageToAllClients({
@ -953,17 +922,6 @@ async function streamMessage(req: Request, res: Response) {
done: true done: true
}); });
} }
} catch (error) {
log.error(`Error during streaming: ${error}`);
// Send error to client via WebSocket
wsService.sendMessageToAllClients({
type: 'llm-stream',
chatNoteId: chatNoteId,
error: `Error processing message: ${error}`,
done: true
});
}
} catch (error: any) { } catch (error: any) {
log.error(`Error starting message stream: ${error.message}`); log.error(`Error starting message stream: ${error.message}`);
log.error(`Error starting message stream, can't communicate via WebSocket: ${error.message}`); log.error(`Error starting message stream, can't communicate via WebSocket: ${error.message}`);

View File

@ -45,13 +45,9 @@ interface NoteContext {
export class AIServiceManager implements IAIServiceManager { export class AIServiceManager implements IAIServiceManager {
private services: Partial<Record<ServiceProviders, AIService>> = {}; private services: Partial<Record<ServiceProviders, AIService>> = {};
private providerOrder: ServiceProviders[] = []; // Will be populated from configuration
private initialized = false; private initialized = false;
constructor() { constructor() {
// Initialize provider order immediately
this.updateProviderOrder();
// Initialize tools immediately // Initialize tools immediately
this.initializeTools().catch(error => { this.initializeTools().catch(error => {
log.error(`Error initializing LLM tools during AIServiceManager construction: ${error.message || String(error)}`); log.error(`Error initializing LLM tools during AIServiceManager construction: ${error.message || String(error)}`);
@ -59,6 +55,8 @@ export class AIServiceManager implements IAIServiceManager {
// Set up event listener for provider changes // Set up event listener for provider changes
this.setupProviderChangeListener(); this.setupProviderChangeListener();
this.initialized = true;
} }
/** /**
@ -83,44 +81,18 @@ export class AIServiceManager implements IAIServiceManager {
} }
/** /**
* Update the provider order using the new configuration system (single provider) * Get the currently selected provider using the new configuration system
*/ */
async updateProviderOrderAsync(): Promise<void> { async getSelectedProviderAsync(): Promise<ServiceProviders | null> {
try { try {
const selectedProvider = await getSelectedProvider(); const selectedProvider = await getSelectedProvider();
if (selectedProvider) { return selectedProvider as ServiceProviders || null;
this.providerOrder = [selectedProvider as ServiceProviders];
log.info(`Updated provider order: ${selectedProvider}`);
} else {
this.providerOrder = [];
log.info('No provider selected');
}
this.initialized = true;
} catch (error) { } catch (error) {
log.error(`Failed to get selected provider: ${error}`); log.error(`Failed to get selected provider: ${error}`);
// Keep empty order, will be handled gracefully by other methods return null;
this.providerOrder = [];
this.initialized = true;
} }
} }
/**
* Update the provider precedence order (legacy sync version)
* Returns true if successful, false if options not available yet
*/
updateProviderOrder(): boolean {
if (this.initialized) {
return true;
}
// Use async version but don't wait
this.updateProviderOrderAsync().catch(error => {
log.error(`Error in async provider order update: ${error}`);
});
return true;
}
/** /**
* Validate AI configuration using the new configuration system * Validate AI configuration using the new configuration system
*/ */
@ -162,16 +134,44 @@ export class AIServiceManager implements IAIServiceManager {
* Ensure manager is initialized before using * Ensure manager is initialized before using
*/ */
private ensureInitialized() { private ensureInitialized() {
if (!this.initialized) { // No longer needed with simplified approach
this.updateProviderOrder(); }
/**
* Get or create any available AI service following the simplified pattern
* Returns a service or throws a meaningful error
*/
async getOrCreateAnyService(): Promise<AIService> {
this.ensureInitialized();
// Get the selected provider using the new configuration system
const selectedProvider = await this.getSelectedProviderAsync();
if (!selectedProvider) {
throw new Error('No AI provider is selected. Please select a provider (OpenAI, Anthropic, or Ollama) in your AI settings.');
}
try {
const service = await this.getOrCreateChatProvider(selectedProvider);
if (service) {
return service;
}
throw new Error(`Failed to create ${selectedProvider} service`);
} catch (error) {
log.error(`Provider ${selectedProvider} not available: ${error}`);
throw new Error(`Selected AI provider (${selectedProvider}) is not available. Please check your configuration: ${error}`);
} }
} }
/** /**
* Check if any AI service is available * Check if any AI service is available (legacy method for backward compatibility)
*/ */
isAnyServiceAvailable(): boolean { isAnyServiceAvailable(): boolean {
return Object.values(this.services).some(service => service.isAvailable()); this.ensureInitialized();
// Check if we have the selected provider available
return this.getAvailableProviders().length > 0;
} }
/** /**
@ -235,25 +235,27 @@ export class AIServiceManager implements IAIServiceManager {
throw new Error('No messages provided for chat completion'); throw new Error('No messages provided for chat completion');
} }
// Try providers in order of preference // Get the selected provider
const availableProviders = this.getAvailableProviders(); const selectedProvider = await this.getSelectedProviderAsync();
if (availableProviders.length === 0) { if (!selectedProvider) {
throw new Error('No AI providers are available. Please check your AI settings.'); throw new Error('No AI provider is selected. Please select a provider in your AI settings.');
} }
// Sort available providers by precedence // Check if the selected provider is available
const sortedProviders = this.providerOrder const availableProviders = this.getAvailableProviders();
.filter(provider => availableProviders.includes(provider)); if (!availableProviders.includes(selectedProvider)) {
throw new Error(`Selected AI provider (${selectedProvider}) is not available. Please check your configuration.`);
}
// If a specific provider is requested and available, use it // If a specific provider is requested and available, use it
if (options.model && options.model.includes(':')) { if (options.model && options.model.includes(':')) {
// Use the new configuration system to parse model identifier // Use the new configuration system to parse model identifier
const modelIdentifier = parseModelIdentifier(options.model); const modelIdentifier = parseModelIdentifier(options.model);
if (modelIdentifier.provider && availableProviders.includes(modelIdentifier.provider as ServiceProviders)) { if (modelIdentifier.provider && modelIdentifier.provider === selectedProvider) {
try { try {
const service = this.services[modelIdentifier.provider as ServiceProviders]; const service = await this.getOrCreateChatProvider(modelIdentifier.provider as ServiceProviders);
if (service) { if (service) {
const modifiedOptions = { ...options, model: modelIdentifier.modelId }; const modifiedOptions = { ...options, model: modelIdentifier.modelId };
log.info(`[AIServiceManager] Using provider ${modelIdentifier.provider} from model prefix with modifiedOptions.stream: ${modifiedOptions.stream}`); log.info(`[AIServiceManager] Using provider ${modelIdentifier.provider} from model prefix with modifiedOptions.stream: ${modifiedOptions.stream}`);
@ -261,44 +263,28 @@ export class AIServiceManager implements IAIServiceManager {
} }
} catch (error) { } catch (error) {
log.error(`Error with specified provider ${modelIdentifier.provider}: ${error}`); log.error(`Error with specified provider ${modelIdentifier.provider}: ${error}`);
// If the specified provider fails, continue with the fallback providers throw new Error(`Failed to use specified provider ${modelIdentifier.provider}: ${error}`);
} }
} else if (modelIdentifier.provider && modelIdentifier.provider !== selectedProvider) {
throw new Error(`Model specifies provider '${modelIdentifier.provider}' but selected provider is '${selectedProvider}'. Please select the correct provider or use a model without provider prefix.`);
} }
// If not a provider prefix, treat the entire string as a model name and continue with normal provider selection // If not a provider prefix, treat the entire string as a model name and continue with normal provider selection
} }
// If user has a specific provider selected, try only that one and fail fast // Use the selected provider
if (this.providerOrder.length === 1 && sortedProviders.length === 1) { try {
const selectedProvider = sortedProviders[0];
const service = await this.getOrCreateChatProvider(selectedProvider); const service = await this.getOrCreateChatProvider(selectedProvider);
if (!service) { if (!service) {
throw new Error(`Failed to create selected chat provider: ${selectedProvider}. Please check your configuration.`); throw new Error(`Failed to create selected chat provider: ${selectedProvider}. Please check your configuration.`);
} }
log.info(`[AIServiceManager] Using selected provider ${selectedProvider} with options.stream: ${options.stream}`); log.info(`[AIServiceManager] Using selected provider ${selectedProvider} with options.stream: ${options.stream}`);
return await service.generateChatCompletion(messages, options); return await service.generateChatCompletion(messages, options);
}
// If no specific provider selected, try each provider in order until one succeeds
let lastError: Error | null = null;
for (const provider of sortedProviders) {
try {
const service = await this.getOrCreateChatProvider(provider);
if (service) {
log.info(`[AIServiceManager] Trying provider ${provider} with options.stream: ${options.stream}`);
return await service.generateChatCompletion(messages, options);
}
} catch (error) { } catch (error) {
log.error(`Error with provider ${provider}: ${error}`); log.error(`Error with selected provider ${selectedProvider}: ${error}`);
lastError = error as Error; throw new Error(`Selected AI provider (${selectedProvider}) failed: ${error}`);
// Continue to the next provider
} }
} }
// If we get here, all providers failed
throw new Error(`All AI providers failed: ${lastError?.message || 'Unknown error'}`);
}
setupEventListeners() { setupEventListeners() {
// Setup event listeners for AI services // Setup event listeners for AI services
} }
@ -408,8 +394,8 @@ export class AIServiceManager implements IAIServiceManager {
switch (providerName) { switch (providerName) {
case 'openai': { case 'openai': {
const apiKey = await options.getOption('openaiApiKey'); const apiKey = options.getOption('openaiApiKey');
const baseUrl = await options.getOption('openaiBaseUrl'); const baseUrl = options.getOption('openaiBaseUrl');
if (!apiKey && !baseUrl) return null; if (!apiKey && !baseUrl) return null;
service = new OpenAIService(); service = new OpenAIService();
@ -421,7 +407,7 @@ export class AIServiceManager implements IAIServiceManager {
} }
case 'anthropic': { case 'anthropic': {
const apiKey = await options.getOption('anthropicApiKey'); const apiKey = options.getOption('anthropicApiKey');
if (!apiKey) return null; if (!apiKey) return null;
service = new AnthropicService(); service = new AnthropicService();
@ -432,7 +418,7 @@ export class AIServiceManager implements IAIServiceManager {
} }
case 'ollama': { case 'ollama': {
const baseUrl = await options.getOption('ollamaBaseUrl'); const baseUrl = options.getOption('ollamaBaseUrl');
if (!baseUrl) return null; if (!baseUrl) return null;
service = new OllamaService(); service = new OllamaService();
@ -445,7 +431,6 @@ export class AIServiceManager implements IAIServiceManager {
if (service) { if (service) {
this.services[providerName] = service; this.services[providerName] = service;
log.info(`Created and validated ${providerName} chat provider`);
return service; return service;
} }
} catch (error: any) { } catch (error: any) {
@ -470,9 +455,6 @@ export class AIServiceManager implements IAIServiceManager {
return; return;
} }
// Update provider order from configuration
await this.updateProviderOrderAsync();
// Initialize index service // Initialize index service
await this.getIndexService().initialize(); await this.getIndexService().initialize();
@ -590,18 +572,22 @@ export class AIServiceManager implements IAIServiceManager {
if (service && service.isAvailable()) { if (service && service.isAvailable()) {
return service; return service;
} }
throw new Error(`Specified provider ${provider} is not available`);
} }
// Otherwise, try providers in the configured order // Otherwise, use the selected provider
for (const providerName of this.providerOrder) { const selectedProvider = await this.getSelectedProviderAsync();
const service = await this.getOrCreateChatProvider(providerName); if (!selectedProvider) {
throw new Error('No AI provider is selected. Please select a provider in your AI settings.');
}
const service = await this.getOrCreateChatProvider(selectedProvider);
if (service && service.isAvailable()) { if (service && service.isAvailable()) {
return service; return service;
} }
}
// If no provider is available, throw a clear error // If no provider is available, throw a clear error
throw new Error('No AI chat providers are available. Please check your AI settings.'); throw new Error(`Selected AI provider (${selectedProvider}) is not available. Please check your AI settings.`);
} }
/** /**
@ -611,14 +597,14 @@ export class AIServiceManager implements IAIServiceManager {
try { try {
const selectedProvider = await getSelectedProvider(); const selectedProvider = await getSelectedProvider();
if (selectedProvider === null) { if (selectedProvider === null) {
// No provider selected, fallback to first available // No provider selected, fallback to default
log.info('No provider selected, using first available provider'); log.info('No provider selected, using default provider');
return this.providerOrder[0]; return 'openai';
} }
return selectedProvider; return selectedProvider;
} catch (error) { } catch (error) {
log.error(`Error getting preferred provider: ${error}`); log.error(`Error getting preferred provider: ${error}`);
return this.providerOrder[0]; return 'openai';
} }
} }
@ -628,16 +614,18 @@ export class AIServiceManager implements IAIServiceManager {
getSelectedProvider(): string { getSelectedProvider(): string {
this.ensureInitialized(); this.ensureInitialized();
// Return the first available provider in the order // Try to get the selected provider synchronously
for (const providerName of this.providerOrder) { try {
const service = this.services[providerName]; const selectedProvider = options.getOption('aiSelectedProvider');
if (service && service.isAvailable()) { if (selectedProvider) {
return providerName; return selectedProvider;
} }
} catch (error) {
log.error(`Error getting selected provider: ${error}`);
} }
// Return the first provider as fallback // Return a default if nothing is selected (for backward compatibility)
return this.providerOrder[0]; return 'openai';
} }
/** /**
@ -746,9 +734,6 @@ export class AIServiceManager implements IAIServiceManager {
const providerManager = await import('./providers/providers.js'); const providerManager = await import('./providers/providers.js');
providerManager.clearAllEmbeddingProviders(); providerManager.clearAllEmbeddingProviders();
// Update provider order with new configuration
await this.updateProviderOrderAsync();
log.info('LLM services recreated successfully'); log.info('LLM services recreated successfully');
} catch (error) { } catch (error) {
log.error(`Error recreating LLM services: ${this.handleError(error)}`); log.error(`Error recreating LLM services: ${this.handleError(error)}`);
@ -776,6 +761,9 @@ export default {
isAnyServiceAvailable(): boolean { isAnyServiceAvailable(): boolean {
return getInstance().isAnyServiceAvailable(); return getInstance().isAnyServiceAvailable();
}, },
async getOrCreateAnyService(): Promise<AIService> {
return getInstance().getOrCreateAnyService();
},
getAvailableProviders() { getAvailableProviders() {
return getInstance().getAvailableProviders(); return getInstance().getAvailableProviders();
}, },

View File

@ -5,7 +5,7 @@
import log from "../../log.js"; import log from "../../log.js";
import type { Request, Response } from "express"; import type { Request, Response } from "express";
import type { Message, ChatCompletionOptions } from "../ai_interface.js"; import type { Message, ChatCompletionOptions } from "../ai_interface.js";
import { AIServiceManager } from "../ai_service_manager.js"; import aiServiceManager from "../ai_service_manager.js";
import { ChatPipeline } from "../pipeline/chat_pipeline.js"; import { ChatPipeline } from "../pipeline/chat_pipeline.js";
import type { ChatPipelineInput } from "../pipeline/interfaces.js"; import type { ChatPipelineInput } from "../pipeline/interfaces.js";
import options from "../../options.js"; import options from "../../options.js";
@ -33,25 +33,6 @@ class RestChatService {
} }
} }
/**
* Check if AI services are available
*/
safelyUseAIManager(): boolean {
if (!this.isDatabaseInitialized()) {
log.info("AI check failed: Database is not initialized");
return false;
}
try {
const aiManager = new AIServiceManager();
const isAvailable = aiManager.isAnyServiceAvailable();
log.info(`AI service availability check result: ${isAvailable}`);
return isAvailable;
} catch (error) {
log.error(`Error accessing AI service manager: ${error}`);
return false;
}
}
/** /**
* Handle a message sent to an LLM and get a response * Handle a message sent to an LLM and get a response
@ -93,10 +74,14 @@ class RestChatService {
return { error: "AI features are disabled. Please enable them in the settings." }; return { error: "AI features are disabled. Please enable them in the settings." };
} }
if (!this.safelyUseAIManager()) { // Check database initialization first
return { error: "AI services are currently unavailable. Please check your configuration." }; if (!this.isDatabaseInitialized()) {
throw new Error("Database is not initialized");
} }
// Get or create AI service - will throw meaningful error if not possible
await aiServiceManager.getOrCreateAnyService();
// Load or create chat directly from storage // Load or create chat directly from storage
let chat = await chatStorageService.getChat(chatNoteId); let chat = await chatStorageService.getChat(chatNoteId);

View File

@ -70,7 +70,7 @@ export class ConfigurationManager {
*/ */
public async getSelectedProvider(): Promise<ProviderType | null> { public async getSelectedProvider(): Promise<ProviderType | null> {
try { try {
const selectedProvider = await options.getOption('aiSelectedProvider'); const selectedProvider = options.getOption('aiSelectedProvider');
return selectedProvider as ProviderType || null; return selectedProvider as ProviderType || null;
} catch (error) { } catch (error) {
log.error(`Error getting selected provider: ${error}`); log.error(`Error getting selected provider: ${error}`);
@ -83,7 +83,7 @@ export class ConfigurationManager {
*/ */
public async getSelectedEmbeddingProvider(): Promise<EmbeddingProviderType | null> { public async getSelectedEmbeddingProvider(): Promise<EmbeddingProviderType | null> {
try { try {
const selectedProvider = await options.getOption('embeddingSelectedProvider'); const selectedProvider = options.getOption('embeddingSelectedProvider');
return selectedProvider as EmbeddingProviderType || null; return selectedProvider as EmbeddingProviderType || null;
} catch (error) { } catch (error) {
log.error(`Error getting selected embedding provider: ${error}`); log.error(`Error getting selected embedding provider: ${error}`);
@ -155,11 +155,9 @@ export class ConfigurationManager {
*/ */
public async getDefaultModels(): Promise<Record<ProviderType, string | undefined>> { public async getDefaultModels(): Promise<Record<ProviderType, string | undefined>> {
try { try {
const [openaiModel, anthropicModel, ollamaModel] = await Promise.all([ const openaiModel = options.getOption('openaiDefaultModel');
options.getOption('openaiDefaultModel'), const anthropicModel = options.getOption('anthropicDefaultModel');
options.getOption('anthropicDefaultModel'), const ollamaModel = options.getOption('ollamaDefaultModel');
options.getOption('ollamaDefaultModel')
]);
return { return {
openai: openaiModel || undefined, openai: openaiModel || undefined,
@ -182,20 +180,14 @@ export class ConfigurationManager {
*/ */
public async getProviderSettings(): Promise<ProviderSettings> { public async getProviderSettings(): Promise<ProviderSettings> {
try { try {
const [ const openaiApiKey = options.getOption('openaiApiKey');
openaiApiKey, openaiBaseUrl, openaiDefaultModel, const openaiBaseUrl = options.getOption('openaiBaseUrl');
anthropicApiKey, anthropicBaseUrl, anthropicDefaultModel, const openaiDefaultModel = options.getOption('openaiDefaultModel');
ollamaBaseUrl, ollamaDefaultModel const anthropicApiKey = options.getOption('anthropicApiKey');
] = await Promise.all([ const anthropicBaseUrl = options.getOption('anthropicBaseUrl');
options.getOption('openaiApiKey'), const anthropicDefaultModel = options.getOption('anthropicDefaultModel');
options.getOption('openaiBaseUrl'), const ollamaBaseUrl = options.getOption('ollamaBaseUrl');
options.getOption('openaiDefaultModel'), const ollamaDefaultModel = options.getOption('ollamaDefaultModel');
options.getOption('anthropicApiKey'),
options.getOption('anthropicBaseUrl'),
options.getOption('anthropicDefaultModel'),
options.getOption('ollamaBaseUrl'),
options.getOption('ollamaDefaultModel')
]);
const settings: ProviderSettings = {}; const settings: ProviderSettings = {};
@ -302,7 +294,7 @@ export class ConfigurationManager {
private async getAIEnabled(): Promise<boolean> { private async getAIEnabled(): Promise<boolean> {
try { try {
return await options.getOptionBool('aiEnabled'); return options.getOptionBool('aiEnabled');
} catch { } catch {
return false; return false;
} }