fix(memory): correct embedBatch comment — /api/embed supports batched input

This commit is contained in:
Varun Chopra 2026-03-08 16:25:48 +00:00
parent 96f413f68b
commit bddd6d8239

View File

@ -118,7 +118,8 @@ export async function createOllamaEmbeddingProvider(
model: client.model,
embedQuery: embedOne,
embedBatch: async (texts: string[]) => {
// Ollama /api/embed accepts one input per request.
// Ollama /api/embed supports batched input, but we fan-out here to
// keep error handling and response normalisation consistent per text.
return await Promise.all(texts.map(embedOne));
},
};