diff --git a/src/agents/pi-embedded-runner/model.test.ts b/src/agents/pi-embedded-runner/model.test.ts index 72e4d3decf3..0f0d410ec8d 100644 --- a/src/agents/pi-embedded-runner/model.test.ts +++ b/src/agents/pi-embedded-runner/model.test.ts @@ -844,7 +844,8 @@ describe("resolveModel", () => { expect(result.error).toBeUndefined(); expect(result.model).toMatchObject({ ...buildOpenAICodexForwardCompatExpectation("gpt-5.4"), - api: "openai-responses", + // api: "openai-responses" from config is normalized to "openai-codex-responses" + // by normalizeCodexTransport when baseUrl is the codex endpoint contextWindow: 1_050_000, maxTokens: 128_000, }); @@ -910,7 +911,6 @@ describe("resolveModel", () => { } return { ...buildOpenAICodexForwardCompatExpectation("gpt-5.4"), - api: "openai-responses", input: ["text"], contextWindow: 272000, maxTokens: 128000, @@ -924,7 +924,7 @@ describe("resolveModel", () => { expect(result.error).toBeUndefined(); expect(result.model).toMatchObject({ ...buildOpenAICodexForwardCompatExpectation("gpt-5.4"), - api: "openai-responses", + // api stays "openai-codex-responses" after plugin normalization input: ["text"], contextWindow: 272000, maxTokens: 128000, diff --git a/src/agents/pi-embedded-runner/model.ts b/src/agents/pi-embedded-runner/model.ts index 04642eb747c..2219bc3d2b9 100644 --- a/src/agents/pi-embedded-runner/model.ts +++ b/src/agents/pi-embedded-runner/model.ts @@ -304,6 +304,7 @@ function preserveDiscoveredTransportMetadata(params: { input: configuredModel?.input ?? discoveredModel.input ?? dynamicModel.input, compat: configuredModel?.compat ?? discoveredModel.compat ?? dynamicModel.compat, maxTokens: configuredModel?.maxTokens ?? discoveredModel.maxTokens ?? dynamicModel.maxTokens, + cost: discoveredModel.cost ?? dynamicModel.cost, headers: discoveredHeaders || dynamicHeaders ? {