test(openai): broaden live model coverage
This commit is contained in:
parent
f1802a5bc7
commit
d1d46c6cfb
@ -3,9 +3,71 @@ import { describe, expect, it } from "vitest";
|
|||||||
import { buildOpenAIProvider } from "./openai-provider.js";
|
import { buildOpenAIProvider } from "./openai-provider.js";
|
||||||
|
|
||||||
const OPENAI_API_KEY = process.env.OPENAI_API_KEY ?? "";
|
const OPENAI_API_KEY = process.env.OPENAI_API_KEY ?? "";
|
||||||
|
const DEFAULT_LIVE_MODEL_IDS = ["gpt-5.4-mini", "gpt-5.4-nano"] as const;
|
||||||
const liveEnabled = OPENAI_API_KEY.trim().length > 0 && process.env.OPENCLAW_LIVE_TEST === "1";
|
const liveEnabled = OPENAI_API_KEY.trim().length > 0 && process.env.OPENCLAW_LIVE_TEST === "1";
|
||||||
const describeLive = liveEnabled ? describe : describe.skip;
|
const describeLive = liveEnabled ? describe : describe.skip;
|
||||||
|
|
||||||
|
type LiveModelCase = {
|
||||||
|
modelId: string;
|
||||||
|
templateId: string;
|
||||||
|
templateName: string;
|
||||||
|
cost: { input: number; output: number; cacheRead: number; cacheWrite: number };
|
||||||
|
contextWindow: number;
|
||||||
|
maxTokens: number;
|
||||||
|
};
|
||||||
|
|
||||||
|
function resolveLiveModelCase(modelId: string): LiveModelCase {
|
||||||
|
switch (modelId) {
|
||||||
|
case "gpt-5.4":
|
||||||
|
return {
|
||||||
|
modelId,
|
||||||
|
templateId: "gpt-5.2",
|
||||||
|
templateName: "GPT-5.2",
|
||||||
|
cost: { input: 1, output: 2, cacheRead: 0, cacheWrite: 0 },
|
||||||
|
contextWindow: 400_000,
|
||||||
|
maxTokens: 128_000,
|
||||||
|
};
|
||||||
|
case "gpt-5.4-pro":
|
||||||
|
return {
|
||||||
|
modelId,
|
||||||
|
templateId: "gpt-5.2-pro",
|
||||||
|
templateName: "GPT-5.2 Pro",
|
||||||
|
cost: { input: 15, output: 60, cacheRead: 0, cacheWrite: 0 },
|
||||||
|
contextWindow: 400_000,
|
||||||
|
maxTokens: 128_000,
|
||||||
|
};
|
||||||
|
case "gpt-5.4-mini":
|
||||||
|
return {
|
||||||
|
modelId,
|
||||||
|
templateId: "gpt-5-mini",
|
||||||
|
templateName: "GPT-5 mini",
|
||||||
|
cost: { input: 1, output: 2, cacheRead: 0, cacheWrite: 0 },
|
||||||
|
contextWindow: 400_000,
|
||||||
|
maxTokens: 128_000,
|
||||||
|
};
|
||||||
|
case "gpt-5.4-nano":
|
||||||
|
return {
|
||||||
|
modelId,
|
||||||
|
templateId: "gpt-5-nano",
|
||||||
|
templateName: "GPT-5 nano",
|
||||||
|
cost: { input: 0.5, output: 1, cacheRead: 0, cacheWrite: 0 },
|
||||||
|
contextWindow: 200_000,
|
||||||
|
maxTokens: 64_000,
|
||||||
|
};
|
||||||
|
default:
|
||||||
|
throw new Error(`Unsupported live OpenAI model: ${modelId}`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
function resolveLiveModelCases(raw?: string): LiveModelCase[] {
|
||||||
|
const requested = raw
|
||||||
|
?.split(",")
|
||||||
|
.map((value) => value.trim())
|
||||||
|
.filter(Boolean);
|
||||||
|
const modelIds = requested?.length ? requested : [...DEFAULT_LIVE_MODEL_IDS];
|
||||||
|
return [...new Set(modelIds)].map((modelId) => resolveLiveModelCase(modelId));
|
||||||
|
}
|
||||||
|
|
||||||
describe("buildOpenAIProvider", () => {
|
describe("buildOpenAIProvider", () => {
|
||||||
it("resolves gpt-5.4 mini and nano from GPT-5 small-model templates", () => {
|
it("resolves gpt-5.4 mini and nano from GPT-5 small-model templates", () => {
|
||||||
const provider = buildOpenAIProvider();
|
const provider = buildOpenAIProvider();
|
||||||
@ -113,63 +175,67 @@ describe("buildOpenAIProvider", () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
describeLive("buildOpenAIProvider live", () => {
|
describeLive("buildOpenAIProvider live", () => {
|
||||||
it("resolves a live model and completes through the OpenAI responses API", async () => {
|
it.each(resolveLiveModelCases(process.env.OPENCLAW_LIVE_OPENAI_MODELS))(
|
||||||
const provider = buildOpenAIProvider();
|
"resolves %s and completes through the OpenAI responses API",
|
||||||
const registry = {
|
async (liveCase) => {
|
||||||
find(providerId: string, id: string) {
|
const provider = buildOpenAIProvider();
|
||||||
if (providerId !== "openai") {
|
const registry = {
|
||||||
|
find(providerId: string, id: string) {
|
||||||
|
if (providerId !== "openai") {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
if (id === liveCase.templateId) {
|
||||||
|
return {
|
||||||
|
id: liveCase.templateId,
|
||||||
|
name: liveCase.templateName,
|
||||||
|
provider: "openai",
|
||||||
|
api: "openai-completions",
|
||||||
|
baseUrl: "https://api.openai.com/v1",
|
||||||
|
reasoning: true,
|
||||||
|
input: ["text", "image"],
|
||||||
|
cost: liveCase.cost,
|
||||||
|
contextWindow: liveCase.contextWindow,
|
||||||
|
maxTokens: liveCase.maxTokens,
|
||||||
|
};
|
||||||
|
}
|
||||||
return null;
|
return null;
|
||||||
}
|
},
|
||||||
if (id === "gpt-5-nano") {
|
};
|
||||||
return {
|
|
||||||
id,
|
|
||||||
name: "GPT-5 nano",
|
|
||||||
provider: "openai",
|
|
||||||
api: "openai-completions",
|
|
||||||
baseUrl: "https://api.openai.com/v1",
|
|
||||||
reasoning: true,
|
|
||||||
input: ["text", "image"],
|
|
||||||
cost: { input: 0.5, output: 1, cacheRead: 0, cacheWrite: 0 },
|
|
||||||
contextWindow: 200_000,
|
|
||||||
maxTokens: 64_000,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
return null;
|
|
||||||
},
|
|
||||||
};
|
|
||||||
|
|
||||||
const resolved = provider.resolveDynamicModel?.({
|
const resolved = provider.resolveDynamicModel?.({
|
||||||
provider: "openai",
|
provider: "openai",
|
||||||
modelId: "gpt-5.4-nano",
|
modelId: liveCase.modelId,
|
||||||
modelRegistry: registry as never,
|
modelRegistry: registry as never,
|
||||||
});
|
});
|
||||||
|
|
||||||
expect(resolved).toBeDefined();
|
expect(resolved).toBeDefined();
|
||||||
|
|
||||||
const normalized = provider.normalizeResolvedModel?.({
|
const normalized = provider.normalizeResolvedModel?.({
|
||||||
provider: "openai",
|
provider: "openai",
|
||||||
modelId: resolved!.id,
|
modelId: resolved!.id,
|
||||||
model: resolved!,
|
model: resolved!,
|
||||||
});
|
});
|
||||||
|
|
||||||
expect(normalized).toMatchObject({
|
expect(normalized).toMatchObject({
|
||||||
provider: "openai",
|
provider: "openai",
|
||||||
id: "gpt-5.4-nano",
|
id: liveCase.modelId,
|
||||||
api: "openai-responses",
|
api: "openai-responses",
|
||||||
baseUrl: "https://api.openai.com/v1",
|
baseUrl: "https://api.openai.com/v1",
|
||||||
});
|
});
|
||||||
|
|
||||||
const client = new OpenAI({
|
const client = new OpenAI({
|
||||||
apiKey: OPENAI_API_KEY,
|
apiKey: OPENAI_API_KEY,
|
||||||
baseURL: normalized?.baseUrl,
|
baseURL: normalized?.baseUrl,
|
||||||
});
|
});
|
||||||
|
|
||||||
const response = await client.responses.create({
|
const response = await client.responses.create({
|
||||||
model: normalized?.id ?? "gpt-5.4-nano",
|
model: normalized?.id ?? liveCase.modelId,
|
||||||
input: "Reply with exactly OK.",
|
input: "Reply with exactly OK.",
|
||||||
max_output_tokens: 16,
|
max_output_tokens: 16,
|
||||||
});
|
});
|
||||||
|
|
||||||
expect(response.output_text.trim()).toBe("OK");
|
expect(response.output_text.trim()).toMatch(/^OK[.!]?$/);
|
||||||
}, 30_000);
|
},
|
||||||
|
30_000,
|
||||||
|
);
|
||||||
});
|
});
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user