fix: Ollama is unable to create a knowledge base using a local embedding model #630

This commit is contained in:
kangfenmao 2025-01-06 15:43:20 +08:00
parent 158f9bf1ad
commit bb50447a98
2 changed files with 3 additions and 6 deletions

View File

@ -72,6 +72,7 @@ const PopupContainer: React.FC<Props> = ({ title, resolve }) => {
} catch (error) {
console.error('Error getting embedding dimensions:', error)
window.message.error(t('message.error.get_embedding_dimensions'))
setLoading(false)
return
}

View File

@ -1,7 +1,7 @@
import type { ExtractChunkData } from '@llm-tools/embedjs-interfaces'
import AiProvider from '@renderer/providers/AiProvider'
import { FileType, KnowledgeBase, KnowledgeBaseParams, Message } from '@renderer/types'
import { isEmpty, take } from 'lodash'
import { take } from 'lodash'
import { getProviderByModel } from './AssistantService'
import FileManager from './FileManager'
@ -10,10 +10,6 @@ export const getKnowledgeBaseParams = (base: KnowledgeBase): KnowledgeBaseParams
const provider = getProviderByModel(base.model)
const aiProvider = new AiProvider(provider)
if (provider.id === 'ollama' && isEmpty(provider.apiKey)) {
provider.apiKey = 'empty'
}
let host = aiProvider.getBaseURL()
if (provider.type === 'gemini') {
@ -24,7 +20,7 @@ export const getKnowledgeBaseParams = (base: KnowledgeBase): KnowledgeBaseParams
id: base.id,
model: base.model.id,
dimensions: base.dimensions,
apiKey: aiProvider.getApiKey(),
apiKey: aiProvider.getApiKey() || 'secret',
apiVersion: provider.apiVersion,
baseURL: host
}