Vincent Koc f16ecd1dac
fix(ollama): unify context window handling across discovery, merge, and OpenAI-compat transport (#29205)
* fix(ollama): inject num_ctx for OpenAI-compatible transport

* fix(ollama): discover per-model context and preserve higher limits

* fix(agents): prefer matching provider model for fallback limits

* fix(types): require numeric token limits in provider model merge

* fix(types): accept unknown payload in ollama num_ctx wrapper

* fix(types): simplify ollama settled-result extraction

* config(models): add provider flag for Ollama OpenAI num_ctx injection

* config(schema): allow provider num_ctx injection flag

* config(labels): label provider num_ctx injection flag

* config(help): document provider num_ctx injection flag

* agents(ollama): gate OpenAI num_ctx injection with provider config

* tests(ollama): cover provider num_ctx injection flag behavior

* docs(config): list provider num_ctx injection option

* docs(ollama): document OpenAI num_ctx injection toggle

* docs(config): clarify merge token-limit precedence

* config(help): note merge uses higher model token limits

* fix(ollama): cap /api/show discovery concurrency

* fix(ollama): restrict num_ctx injection to OpenAI compat

* tests(ollama): cover ipv6 and compat num_ctx gating

* fix(ollama): detect remote compat endpoints for ollama-labeled providers

* fix(ollama): cap per-model /api/show lookups to bound discovery load
2026-02-27 17:20:47 -08:00

337 lines
9.8 KiB
TypeScript

import { describe, expect, it, vi } from "vitest";
import type { OpenClawConfig } from "../../../config/config.js";
import {
isOllamaCompatProvider,
resolveAttemptFsWorkspaceOnly,
resolveOllamaCompatNumCtxEnabled,
resolvePromptBuildHookResult,
resolvePromptModeForSession,
shouldInjectOllamaCompatNumCtx,
wrapOllamaCompatNumCtx,
wrapStreamFnTrimToolCallNames,
} from "./attempt.js";
describe("resolvePromptBuildHookResult", () => {
function createLegacyOnlyHookRunner() {
return {
hasHooks: vi.fn(
(hookName: "before_prompt_build" | "before_agent_start") =>
hookName === "before_agent_start",
),
runBeforePromptBuild: vi.fn(async () => undefined),
runBeforeAgentStart: vi.fn(async () => ({ prependContext: "from-hook" })),
};
}
it("reuses precomputed legacy before_agent_start result without invoking hook again", async () => {
const hookRunner = createLegacyOnlyHookRunner();
const result = await resolvePromptBuildHookResult({
prompt: "hello",
messages: [],
hookCtx: {},
hookRunner,
legacyBeforeAgentStartResult: { prependContext: "from-cache", systemPrompt: "legacy-system" },
});
expect(hookRunner.runBeforeAgentStart).not.toHaveBeenCalled();
expect(result).toEqual({
prependContext: "from-cache",
systemPrompt: "legacy-system",
});
});
it("calls legacy hook when precomputed result is absent", async () => {
const hookRunner = createLegacyOnlyHookRunner();
const messages = [{ role: "user", content: "ctx" }];
const result = await resolvePromptBuildHookResult({
prompt: "hello",
messages,
hookCtx: {},
hookRunner,
});
expect(hookRunner.runBeforeAgentStart).toHaveBeenCalledTimes(1);
expect(hookRunner.runBeforeAgentStart).toHaveBeenCalledWith({ prompt: "hello", messages }, {});
expect(result.prependContext).toBe("from-hook");
});
});
describe("resolvePromptModeForSession", () => {
it("uses minimal mode for subagent sessions", () => {
expect(resolvePromptModeForSession("agent:main:subagent:child")).toBe("minimal");
});
it("uses full mode for cron sessions", () => {
expect(resolvePromptModeForSession("agent:main:cron:job-1")).toBe("full");
expect(resolvePromptModeForSession("agent:main:cron:job-1:run:run-abc")).toBe("full");
});
});
describe("resolveAttemptFsWorkspaceOnly", () => {
it("uses global tools.fs.workspaceOnly when agent has no override", () => {
const cfg: OpenClawConfig = {
tools: {
fs: { workspaceOnly: true },
},
};
expect(
resolveAttemptFsWorkspaceOnly({
config: cfg,
sessionAgentId: "main",
}),
).toBe(true);
});
it("prefers agent-specific tools.fs.workspaceOnly override", () => {
const cfg: OpenClawConfig = {
tools: {
fs: { workspaceOnly: true },
},
agents: {
list: [
{
id: "main",
tools: {
fs: { workspaceOnly: false },
},
},
],
},
};
expect(
resolveAttemptFsWorkspaceOnly({
config: cfg,
sessionAgentId: "main",
}),
).toBe(false);
});
});
describe("wrapStreamFnTrimToolCallNames", () => {
function createFakeStream(params: { events: unknown[]; resultMessage: unknown }): {
result: () => Promise<unknown>;
[Symbol.asyncIterator]: () => AsyncIterator<unknown>;
} {
return {
async result() {
return params.resultMessage;
},
[Symbol.asyncIterator]() {
return (async function* () {
for (const event of params.events) {
yield event;
}
})();
},
};
}
it("trims whitespace from live streamed tool call names and final result message", async () => {
const partialToolCall = { type: "toolCall", name: " read " };
const messageToolCall = { type: "toolCall", name: " exec " };
const finalToolCall = { type: "toolCall", name: " write " };
const event = {
type: "toolcall_delta",
partial: { role: "assistant", content: [partialToolCall] },
message: { role: "assistant", content: [messageToolCall] },
};
const finalMessage = { role: "assistant", content: [finalToolCall] };
const baseFn = vi.fn(() => createFakeStream({ events: [event], resultMessage: finalMessage }));
const wrappedFn = wrapStreamFnTrimToolCallNames(baseFn as never);
const stream = wrappedFn({} as never, {} as never, {} as never) as Awaited<
ReturnType<typeof wrappedFn>
>;
const seenEvents: unknown[] = [];
for await (const item of stream) {
seenEvents.push(item);
}
const result = await stream.result();
expect(seenEvents).toHaveLength(1);
expect(partialToolCall.name).toBe("read");
expect(messageToolCall.name).toBe("exec");
expect(finalToolCall.name).toBe("write");
expect(result).toBe(finalMessage);
expect(baseFn).toHaveBeenCalledTimes(1);
});
it("supports async stream functions that return a promise", async () => {
const finalToolCall = { type: "toolCall", name: " browser " };
const finalMessage = { role: "assistant", content: [finalToolCall] };
const baseFn = vi.fn(async () =>
createFakeStream({
events: [],
resultMessage: finalMessage,
}),
);
const wrappedFn = wrapStreamFnTrimToolCallNames(baseFn as never);
const stream = await wrappedFn({} as never, {} as never, {} as never);
const result = await stream.result();
expect(finalToolCall.name).toBe("browser");
expect(result).toBe(finalMessage);
expect(baseFn).toHaveBeenCalledTimes(1);
});
});
describe("isOllamaCompatProvider", () => {
it("detects native ollama provider id", () => {
expect(
isOllamaCompatProvider({
provider: "ollama",
api: "openai-completions",
baseUrl: "https://example.com/v1",
}),
).toBe(true);
});
it("detects localhost Ollama OpenAI-compatible endpoint", () => {
expect(
isOllamaCompatProvider({
provider: "custom",
api: "openai-completions",
baseUrl: "http://127.0.0.1:11434/v1",
}),
).toBe(true);
});
it("does not misclassify non-local OpenAI-compatible providers", () => {
expect(
isOllamaCompatProvider({
provider: "custom",
api: "openai-completions",
baseUrl: "https://api.openrouter.ai/v1",
}),
).toBe(false);
});
it("detects remote Ollama-compatible endpoint when provider id hints ollama", () => {
expect(
isOllamaCompatProvider({
provider: "my-ollama",
api: "openai-completions",
baseUrl: "http://ollama-host:11434/v1",
}),
).toBe(true);
});
it("detects IPv6 loopback Ollama OpenAI-compatible endpoint", () => {
expect(
isOllamaCompatProvider({
provider: "custom",
api: "openai-completions",
baseUrl: "http://[::1]:11434/v1",
}),
).toBe(true);
});
it("does not classify arbitrary remote hosts on 11434 without ollama provider hint", () => {
expect(
isOllamaCompatProvider({
provider: "custom",
api: "openai-completions",
baseUrl: "http://example.com:11434/v1",
}),
).toBe(false);
});
});
describe("wrapOllamaCompatNumCtx", () => {
it("injects num_ctx and preserves downstream onPayload hooks", () => {
let payloadSeen: Record<string, unknown> | undefined;
const baseFn = vi.fn((_model, _context, options) => {
const payload: Record<string, unknown> = { options: { temperature: 0.1 } };
options?.onPayload?.(payload);
payloadSeen = payload;
return {} as never;
});
const downstream = vi.fn();
const wrapped = wrapOllamaCompatNumCtx(baseFn as never, 202752);
void wrapped({} as never, {} as never, { onPayload: downstream } as never);
expect(baseFn).toHaveBeenCalledTimes(1);
expect((payloadSeen?.options as Record<string, unknown> | undefined)?.num_ctx).toBe(202752);
expect(downstream).toHaveBeenCalledTimes(1);
});
});
describe("resolveOllamaCompatNumCtxEnabled", () => {
it("defaults to true when config is missing", () => {
expect(resolveOllamaCompatNumCtxEnabled({ providerId: "ollama" })).toBe(true);
});
it("defaults to true when provider config is missing", () => {
expect(
resolveOllamaCompatNumCtxEnabled({
config: { models: { providers: {} } },
providerId: "ollama",
}),
).toBe(true);
});
it("returns false when provider flag is explicitly disabled", () => {
expect(
resolveOllamaCompatNumCtxEnabled({
config: {
models: {
providers: {
ollama: {
baseUrl: "http://127.0.0.1:11434/v1",
api: "openai-completions",
injectNumCtxForOpenAICompat: false,
models: [],
},
},
},
},
providerId: "ollama",
}),
).toBe(false);
});
});
describe("shouldInjectOllamaCompatNumCtx", () => {
it("requires openai-completions adapter", () => {
expect(
shouldInjectOllamaCompatNumCtx({
model: {
provider: "ollama",
api: "openai-responses",
baseUrl: "http://127.0.0.1:11434/v1",
},
}),
).toBe(false);
});
it("respects provider flag disablement", () => {
expect(
shouldInjectOllamaCompatNumCtx({
model: {
provider: "ollama",
api: "openai-completions",
baseUrl: "http://127.0.0.1:11434/v1",
},
config: {
models: {
providers: {
ollama: {
baseUrl: "http://127.0.0.1:11434/v1",
api: "openai-completions",
injectNumCtxForOpenAICompat: false,
models: [],
},
},
},
},
providerId: "ollama",
}),
).toBe(false);
});
});