* fix(ollama): inject num_ctx for OpenAI-compatible transport * fix(ollama): discover per-model context and preserve higher limits * fix(agents): prefer matching provider model for fallback limits * fix(types): require numeric token limits in provider model merge * fix(types): accept unknown payload in ollama num_ctx wrapper * fix(types): simplify ollama settled-result extraction * config(models): add provider flag for Ollama OpenAI num_ctx injection * config(schema): allow provider num_ctx injection flag * config(labels): label provider num_ctx injection flag * config(help): document provider num_ctx injection flag * agents(ollama): gate OpenAI num_ctx injection with provider config * tests(ollama): cover provider num_ctx injection flag behavior * docs(config): list provider num_ctx injection option * docs(ollama): document OpenAI num_ctx injection toggle * docs(config): clarify merge token-limit precedence * config(help): note merge uses higher model token limits * fix(ollama): cap /api/show discovery concurrency * fix(ollama): restrict num_ctx injection to OpenAI compat * tests(ollama): cover ipv6 and compat num_ctx gating * fix(ollama): detect remote compat endpoints for ollama-labeled providers * fix(ollama): cap per-model /api/show lookups to bound discovery load
337 lines
9.8 KiB
TypeScript
337 lines
9.8 KiB
TypeScript
import { describe, expect, it, vi } from "vitest";
|
|
import type { OpenClawConfig } from "../../../config/config.js";
|
|
import {
|
|
isOllamaCompatProvider,
|
|
resolveAttemptFsWorkspaceOnly,
|
|
resolveOllamaCompatNumCtxEnabled,
|
|
resolvePromptBuildHookResult,
|
|
resolvePromptModeForSession,
|
|
shouldInjectOllamaCompatNumCtx,
|
|
wrapOllamaCompatNumCtx,
|
|
wrapStreamFnTrimToolCallNames,
|
|
} from "./attempt.js";
|
|
|
|
describe("resolvePromptBuildHookResult", () => {
|
|
function createLegacyOnlyHookRunner() {
|
|
return {
|
|
hasHooks: vi.fn(
|
|
(hookName: "before_prompt_build" | "before_agent_start") =>
|
|
hookName === "before_agent_start",
|
|
),
|
|
runBeforePromptBuild: vi.fn(async () => undefined),
|
|
runBeforeAgentStart: vi.fn(async () => ({ prependContext: "from-hook" })),
|
|
};
|
|
}
|
|
|
|
it("reuses precomputed legacy before_agent_start result without invoking hook again", async () => {
|
|
const hookRunner = createLegacyOnlyHookRunner();
|
|
const result = await resolvePromptBuildHookResult({
|
|
prompt: "hello",
|
|
messages: [],
|
|
hookCtx: {},
|
|
hookRunner,
|
|
legacyBeforeAgentStartResult: { prependContext: "from-cache", systemPrompt: "legacy-system" },
|
|
});
|
|
|
|
expect(hookRunner.runBeforeAgentStart).not.toHaveBeenCalled();
|
|
expect(result).toEqual({
|
|
prependContext: "from-cache",
|
|
systemPrompt: "legacy-system",
|
|
});
|
|
});
|
|
|
|
it("calls legacy hook when precomputed result is absent", async () => {
|
|
const hookRunner = createLegacyOnlyHookRunner();
|
|
const messages = [{ role: "user", content: "ctx" }];
|
|
const result = await resolvePromptBuildHookResult({
|
|
prompt: "hello",
|
|
messages,
|
|
hookCtx: {},
|
|
hookRunner,
|
|
});
|
|
|
|
expect(hookRunner.runBeforeAgentStart).toHaveBeenCalledTimes(1);
|
|
expect(hookRunner.runBeforeAgentStart).toHaveBeenCalledWith({ prompt: "hello", messages }, {});
|
|
expect(result.prependContext).toBe("from-hook");
|
|
});
|
|
});
|
|
|
|
describe("resolvePromptModeForSession", () => {
|
|
it("uses minimal mode for subagent sessions", () => {
|
|
expect(resolvePromptModeForSession("agent:main:subagent:child")).toBe("minimal");
|
|
});
|
|
|
|
it("uses full mode for cron sessions", () => {
|
|
expect(resolvePromptModeForSession("agent:main:cron:job-1")).toBe("full");
|
|
expect(resolvePromptModeForSession("agent:main:cron:job-1:run:run-abc")).toBe("full");
|
|
});
|
|
});
|
|
|
|
describe("resolveAttemptFsWorkspaceOnly", () => {
|
|
it("uses global tools.fs.workspaceOnly when agent has no override", () => {
|
|
const cfg: OpenClawConfig = {
|
|
tools: {
|
|
fs: { workspaceOnly: true },
|
|
},
|
|
};
|
|
|
|
expect(
|
|
resolveAttemptFsWorkspaceOnly({
|
|
config: cfg,
|
|
sessionAgentId: "main",
|
|
}),
|
|
).toBe(true);
|
|
});
|
|
|
|
it("prefers agent-specific tools.fs.workspaceOnly override", () => {
|
|
const cfg: OpenClawConfig = {
|
|
tools: {
|
|
fs: { workspaceOnly: true },
|
|
},
|
|
agents: {
|
|
list: [
|
|
{
|
|
id: "main",
|
|
tools: {
|
|
fs: { workspaceOnly: false },
|
|
},
|
|
},
|
|
],
|
|
},
|
|
};
|
|
|
|
expect(
|
|
resolveAttemptFsWorkspaceOnly({
|
|
config: cfg,
|
|
sessionAgentId: "main",
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
});
|
|
|
|
describe("wrapStreamFnTrimToolCallNames", () => {
|
|
function createFakeStream(params: { events: unknown[]; resultMessage: unknown }): {
|
|
result: () => Promise<unknown>;
|
|
[Symbol.asyncIterator]: () => AsyncIterator<unknown>;
|
|
} {
|
|
return {
|
|
async result() {
|
|
return params.resultMessage;
|
|
},
|
|
[Symbol.asyncIterator]() {
|
|
return (async function* () {
|
|
for (const event of params.events) {
|
|
yield event;
|
|
}
|
|
})();
|
|
},
|
|
};
|
|
}
|
|
|
|
it("trims whitespace from live streamed tool call names and final result message", async () => {
|
|
const partialToolCall = { type: "toolCall", name: " read " };
|
|
const messageToolCall = { type: "toolCall", name: " exec " };
|
|
const finalToolCall = { type: "toolCall", name: " write " };
|
|
const event = {
|
|
type: "toolcall_delta",
|
|
partial: { role: "assistant", content: [partialToolCall] },
|
|
message: { role: "assistant", content: [messageToolCall] },
|
|
};
|
|
const finalMessage = { role: "assistant", content: [finalToolCall] };
|
|
const baseFn = vi.fn(() => createFakeStream({ events: [event], resultMessage: finalMessage }));
|
|
|
|
const wrappedFn = wrapStreamFnTrimToolCallNames(baseFn as never);
|
|
const stream = wrappedFn({} as never, {} as never, {} as never) as Awaited<
|
|
ReturnType<typeof wrappedFn>
|
|
>;
|
|
|
|
const seenEvents: unknown[] = [];
|
|
for await (const item of stream) {
|
|
seenEvents.push(item);
|
|
}
|
|
const result = await stream.result();
|
|
|
|
expect(seenEvents).toHaveLength(1);
|
|
expect(partialToolCall.name).toBe("read");
|
|
expect(messageToolCall.name).toBe("exec");
|
|
expect(finalToolCall.name).toBe("write");
|
|
expect(result).toBe(finalMessage);
|
|
expect(baseFn).toHaveBeenCalledTimes(1);
|
|
});
|
|
|
|
it("supports async stream functions that return a promise", async () => {
|
|
const finalToolCall = { type: "toolCall", name: " browser " };
|
|
const finalMessage = { role: "assistant", content: [finalToolCall] };
|
|
const baseFn = vi.fn(async () =>
|
|
createFakeStream({
|
|
events: [],
|
|
resultMessage: finalMessage,
|
|
}),
|
|
);
|
|
|
|
const wrappedFn = wrapStreamFnTrimToolCallNames(baseFn as never);
|
|
const stream = await wrappedFn({} as never, {} as never, {} as never);
|
|
const result = await stream.result();
|
|
|
|
expect(finalToolCall.name).toBe("browser");
|
|
expect(result).toBe(finalMessage);
|
|
expect(baseFn).toHaveBeenCalledTimes(1);
|
|
});
|
|
});
|
|
|
|
describe("isOllamaCompatProvider", () => {
|
|
it("detects native ollama provider id", () => {
|
|
expect(
|
|
isOllamaCompatProvider({
|
|
provider: "ollama",
|
|
api: "openai-completions",
|
|
baseUrl: "https://example.com/v1",
|
|
}),
|
|
).toBe(true);
|
|
});
|
|
|
|
it("detects localhost Ollama OpenAI-compatible endpoint", () => {
|
|
expect(
|
|
isOllamaCompatProvider({
|
|
provider: "custom",
|
|
api: "openai-completions",
|
|
baseUrl: "http://127.0.0.1:11434/v1",
|
|
}),
|
|
).toBe(true);
|
|
});
|
|
|
|
it("does not misclassify non-local OpenAI-compatible providers", () => {
|
|
expect(
|
|
isOllamaCompatProvider({
|
|
provider: "custom",
|
|
api: "openai-completions",
|
|
baseUrl: "https://api.openrouter.ai/v1",
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
|
|
it("detects remote Ollama-compatible endpoint when provider id hints ollama", () => {
|
|
expect(
|
|
isOllamaCompatProvider({
|
|
provider: "my-ollama",
|
|
api: "openai-completions",
|
|
baseUrl: "http://ollama-host:11434/v1",
|
|
}),
|
|
).toBe(true);
|
|
});
|
|
|
|
it("detects IPv6 loopback Ollama OpenAI-compatible endpoint", () => {
|
|
expect(
|
|
isOllamaCompatProvider({
|
|
provider: "custom",
|
|
api: "openai-completions",
|
|
baseUrl: "http://[::1]:11434/v1",
|
|
}),
|
|
).toBe(true);
|
|
});
|
|
|
|
it("does not classify arbitrary remote hosts on 11434 without ollama provider hint", () => {
|
|
expect(
|
|
isOllamaCompatProvider({
|
|
provider: "custom",
|
|
api: "openai-completions",
|
|
baseUrl: "http://example.com:11434/v1",
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
});
|
|
|
|
describe("wrapOllamaCompatNumCtx", () => {
|
|
it("injects num_ctx and preserves downstream onPayload hooks", () => {
|
|
let payloadSeen: Record<string, unknown> | undefined;
|
|
const baseFn = vi.fn((_model, _context, options) => {
|
|
const payload: Record<string, unknown> = { options: { temperature: 0.1 } };
|
|
options?.onPayload?.(payload);
|
|
payloadSeen = payload;
|
|
return {} as never;
|
|
});
|
|
const downstream = vi.fn();
|
|
|
|
const wrapped = wrapOllamaCompatNumCtx(baseFn as never, 202752);
|
|
void wrapped({} as never, {} as never, { onPayload: downstream } as never);
|
|
|
|
expect(baseFn).toHaveBeenCalledTimes(1);
|
|
expect((payloadSeen?.options as Record<string, unknown> | undefined)?.num_ctx).toBe(202752);
|
|
expect(downstream).toHaveBeenCalledTimes(1);
|
|
});
|
|
});
|
|
|
|
describe("resolveOllamaCompatNumCtxEnabled", () => {
|
|
it("defaults to true when config is missing", () => {
|
|
expect(resolveOllamaCompatNumCtxEnabled({ providerId: "ollama" })).toBe(true);
|
|
});
|
|
|
|
it("defaults to true when provider config is missing", () => {
|
|
expect(
|
|
resolveOllamaCompatNumCtxEnabled({
|
|
config: { models: { providers: {} } },
|
|
providerId: "ollama",
|
|
}),
|
|
).toBe(true);
|
|
});
|
|
|
|
it("returns false when provider flag is explicitly disabled", () => {
|
|
expect(
|
|
resolveOllamaCompatNumCtxEnabled({
|
|
config: {
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "http://127.0.0.1:11434/v1",
|
|
api: "openai-completions",
|
|
injectNumCtxForOpenAICompat: false,
|
|
models: [],
|
|
},
|
|
},
|
|
},
|
|
},
|
|
providerId: "ollama",
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
});
|
|
|
|
describe("shouldInjectOllamaCompatNumCtx", () => {
|
|
it("requires openai-completions adapter", () => {
|
|
expect(
|
|
shouldInjectOllamaCompatNumCtx({
|
|
model: {
|
|
provider: "ollama",
|
|
api: "openai-responses",
|
|
baseUrl: "http://127.0.0.1:11434/v1",
|
|
},
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
|
|
it("respects provider flag disablement", () => {
|
|
expect(
|
|
shouldInjectOllamaCompatNumCtx({
|
|
model: {
|
|
provider: "ollama",
|
|
api: "openai-completions",
|
|
baseUrl: "http://127.0.0.1:11434/v1",
|
|
},
|
|
config: {
|
|
models: {
|
|
providers: {
|
|
ollama: {
|
|
baseUrl: "http://127.0.0.1:11434/v1",
|
|
api: "openai-completions",
|
|
injectNumCtxForOpenAICompat: false,
|
|
models: [],
|
|
},
|
|
},
|
|
},
|
|
},
|
|
providerId: "ollama",
|
|
}),
|
|
).toBe(false);
|
|
});
|
|
});
|