test(openai): broaden live model coverage

This commit is contained in:
Vincent Koc 2026-03-20 15:15:15 -07:00
parent f1802a5bc7
commit d1d46c6cfb

View File

@ -3,9 +3,71 @@ import { describe, expect, it } from "vitest";
import { buildOpenAIProvider } from "./openai-provider.js"; import { buildOpenAIProvider } from "./openai-provider.js";
const OPENAI_API_KEY = process.env.OPENAI_API_KEY ?? ""; const OPENAI_API_KEY = process.env.OPENAI_API_KEY ?? "";
const DEFAULT_LIVE_MODEL_IDS = ["gpt-5.4-mini", "gpt-5.4-nano"] as const;
const liveEnabled = OPENAI_API_KEY.trim().length > 0 && process.env.OPENCLAW_LIVE_TEST === "1"; const liveEnabled = OPENAI_API_KEY.trim().length > 0 && process.env.OPENCLAW_LIVE_TEST === "1";
const describeLive = liveEnabled ? describe : describe.skip; const describeLive = liveEnabled ? describe : describe.skip;
type LiveModelCase = {
modelId: string;
templateId: string;
templateName: string;
cost: { input: number; output: number; cacheRead: number; cacheWrite: number };
contextWindow: number;
maxTokens: number;
};
function resolveLiveModelCase(modelId: string): LiveModelCase {
switch (modelId) {
case "gpt-5.4":
return {
modelId,
templateId: "gpt-5.2",
templateName: "GPT-5.2",
cost: { input: 1, output: 2, cacheRead: 0, cacheWrite: 0 },
contextWindow: 400_000,
maxTokens: 128_000,
};
case "gpt-5.4-pro":
return {
modelId,
templateId: "gpt-5.2-pro",
templateName: "GPT-5.2 Pro",
cost: { input: 15, output: 60, cacheRead: 0, cacheWrite: 0 },
contextWindow: 400_000,
maxTokens: 128_000,
};
case "gpt-5.4-mini":
return {
modelId,
templateId: "gpt-5-mini",
templateName: "GPT-5 mini",
cost: { input: 1, output: 2, cacheRead: 0, cacheWrite: 0 },
contextWindow: 400_000,
maxTokens: 128_000,
};
case "gpt-5.4-nano":
return {
modelId,
templateId: "gpt-5-nano",
templateName: "GPT-5 nano",
cost: { input: 0.5, output: 1, cacheRead: 0, cacheWrite: 0 },
contextWindow: 200_000,
maxTokens: 64_000,
};
default:
throw new Error(`Unsupported live OpenAI model: ${modelId}`);
}
}
function resolveLiveModelCases(raw?: string): LiveModelCase[] {
const requested = raw
?.split(",")
.map((value) => value.trim())
.filter(Boolean);
const modelIds = requested?.length ? requested : [...DEFAULT_LIVE_MODEL_IDS];
return [...new Set(modelIds)].map((modelId) => resolveLiveModelCase(modelId));
}
describe("buildOpenAIProvider", () => { describe("buildOpenAIProvider", () => {
it("resolves gpt-5.4 mini and nano from GPT-5 small-model templates", () => { it("resolves gpt-5.4 mini and nano from GPT-5 small-model templates", () => {
const provider = buildOpenAIProvider(); const provider = buildOpenAIProvider();
@ -113,63 +175,67 @@ describe("buildOpenAIProvider", () => {
}); });
describeLive("buildOpenAIProvider live", () => { describeLive("buildOpenAIProvider live", () => {
it("resolves a live model and completes through the OpenAI responses API", async () => { it.each(resolveLiveModelCases(process.env.OPENCLAW_LIVE_OPENAI_MODELS))(
const provider = buildOpenAIProvider(); "resolves %s and completes through the OpenAI responses API",
const registry = { async (liveCase) => {
find(providerId: string, id: string) { const provider = buildOpenAIProvider();
if (providerId !== "openai") { const registry = {
find(providerId: string, id: string) {
if (providerId !== "openai") {
return null;
}
if (id === liveCase.templateId) {
return {
id: liveCase.templateId,
name: liveCase.templateName,
provider: "openai",
api: "openai-completions",
baseUrl: "https://api.openai.com/v1",
reasoning: true,
input: ["text", "image"],
cost: liveCase.cost,
contextWindow: liveCase.contextWindow,
maxTokens: liveCase.maxTokens,
};
}
return null; return null;
} },
if (id === "gpt-5-nano") { };
return {
id,
name: "GPT-5 nano",
provider: "openai",
api: "openai-completions",
baseUrl: "https://api.openai.com/v1",
reasoning: true,
input: ["text", "image"],
cost: { input: 0.5, output: 1, cacheRead: 0, cacheWrite: 0 },
contextWindow: 200_000,
maxTokens: 64_000,
};
}
return null;
},
};
const resolved = provider.resolveDynamicModel?.({ const resolved = provider.resolveDynamicModel?.({
provider: "openai", provider: "openai",
modelId: "gpt-5.4-nano", modelId: liveCase.modelId,
modelRegistry: registry as never, modelRegistry: registry as never,
}); });
expect(resolved).toBeDefined(); expect(resolved).toBeDefined();
const normalized = provider.normalizeResolvedModel?.({ const normalized = provider.normalizeResolvedModel?.({
provider: "openai", provider: "openai",
modelId: resolved!.id, modelId: resolved!.id,
model: resolved!, model: resolved!,
}); });
expect(normalized).toMatchObject({ expect(normalized).toMatchObject({
provider: "openai", provider: "openai",
id: "gpt-5.4-nano", id: liveCase.modelId,
api: "openai-responses", api: "openai-responses",
baseUrl: "https://api.openai.com/v1", baseUrl: "https://api.openai.com/v1",
}); });
const client = new OpenAI({ const client = new OpenAI({
apiKey: OPENAI_API_KEY, apiKey: OPENAI_API_KEY,
baseURL: normalized?.baseUrl, baseURL: normalized?.baseUrl,
}); });
const response = await client.responses.create({ const response = await client.responses.create({
model: normalized?.id ?? "gpt-5.4-nano", model: normalized?.id ?? liveCase.modelId,
input: "Reply with exactly OK.", input: "Reply with exactly OK.",
max_output_tokens: 16, max_output_tokens: 16,
}); });
expect(response.output_text.trim()).toBe("OK"); expect(response.output_text.trim()).toMatch(/^OK[.!]?$/);
}, 30_000); },
30_000,
);
}); });