diff --git a/src/agents/ollama-stream.test.ts b/src/agents/ollama-stream.test.ts index 03d35834db7..3e465ec8368 100644 --- a/src/agents/ollama-stream.test.ts +++ b/src/agents/ollama-stream.test.ts @@ -598,33 +598,6 @@ describe("createOllamaStreamFn", () => { }, ); }); - - it("sends think:false when reasoning is 'off'", async () => { - await withMockNdjsonFetch( - [ - '{"model":"m","created_at":"t","message":{"role":"assistant","content":"ok"},"done":false}', - '{"model":"m","created_at":"t","message":{"role":"assistant","content":""},"done":true,"prompt_eval_count":1,"eval_count":1}', - ], - async (fetchMock) => { - const streamFn = createOllamaStreamFn("http://ollama-host:11434"); - const stream = await streamFn( - { - id: "deepseek-r1:32b", - api: "ollama", - provider: "ollama", - contextWindow: 131072, - } as never, - { messages: [{ role: "user", content: "hello" }] } as never, - { reasoning: "off" } as never, - ); - await collectStreamEvents(stream); - - const [, reqInit] = fetchMock.mock.calls[0] as unknown as [string, RequestInit]; - const body = JSON.parse(reqInit.body as string) as { think?: boolean }; - expect(body.think).toBe(false); - }, - ); - }); }); describe("resolveOllamaBaseUrlForRun", () => { diff --git a/src/agents/ollama-stream.ts b/src/agents/ollama-stream.ts index 1c3048b67a1..6a8fd608914 100644 --- a/src/agents/ollama-stream.ts +++ b/src/agents/ollama-stream.ts @@ -464,7 +464,7 @@ export function createOllamaStreamFn( // `think` boolean. Forward the reasoning level so `think: false` is // sent explicitly when thinking is disabled (#46680). const thinkParam: { think?: boolean } = {}; - if (options?.reasoning && options.reasoning !== "off") { + if (options?.reasoning) { thinkParam.think = true; } else if (options) { // Thinking explicitly disabled – tell Ollama not to think.