diff --git a/src/public/app/widgets/llm_chat_panel.ts b/src/public/app/widgets/llm_chat_panel.ts index 4c9eb5c4d..aaedb1d21 100644 --- a/src/public/app/widgets/llm_chat_panel.ts +++ b/src/public/app/widgets/llm_chat_panel.ts @@ -757,9 +757,6 @@ export default class LlmChatPanel extends BasicWidget { return; } - // Get the default embedding provider - const defaultProvider = options.get('embeddingsDefaultProvider') || 'openai'; - // Get provider precedence const precedenceStr = options.get('aiProviderPrecedence') || 'openai,anthropic,ollama'; let precedenceList: string[] = []; @@ -800,8 +797,6 @@ export default class LlmChatPanel extends BasicWidget { enabledProviders.push('local'); // Perform validation checks - const defaultInPrecedence = precedenceList.includes(defaultProvider); - const defaultIsEnabled = enabledProviders.includes(defaultProvider); const allPrecedenceEnabled = precedenceList.every((p: string) => enabledProviders.includes(p)); // Get embedding queue status @@ -820,19 +815,11 @@ export default class LlmChatPanel extends BasicWidget { const hasEmbeddingsInQueue = queuedNotes > 0; // Show warning if there are issues - if (!defaultInPrecedence || !defaultIsEnabled || !allPrecedenceEnabled || hasEmbeddingsInQueue) { + if (!allPrecedenceEnabled || hasEmbeddingsInQueue) { let message = 'AI Provider Configuration Issues'; message += '