Fix cost preservation and test api expectations for codex normalization
This commit is contained in:
parent
8dfd3a1875
commit
0467f05067
@ -844,7 +844,8 @@ describe("resolveModel", () => {
|
||||
expect(result.error).toBeUndefined();
|
||||
expect(result.model).toMatchObject({
|
||||
...buildOpenAICodexForwardCompatExpectation("gpt-5.4"),
|
||||
api: "openai-responses",
|
||||
// api: "openai-responses" from config is normalized to "openai-codex-responses"
|
||||
// by normalizeCodexTransport when baseUrl is the codex endpoint
|
||||
contextWindow: 1_050_000,
|
||||
maxTokens: 128_000,
|
||||
});
|
||||
@ -910,7 +911,6 @@ describe("resolveModel", () => {
|
||||
}
|
||||
return {
|
||||
...buildOpenAICodexForwardCompatExpectation("gpt-5.4"),
|
||||
api: "openai-responses",
|
||||
input: ["text"],
|
||||
contextWindow: 272000,
|
||||
maxTokens: 128000,
|
||||
@ -924,7 +924,7 @@ describe("resolveModel", () => {
|
||||
expect(result.error).toBeUndefined();
|
||||
expect(result.model).toMatchObject({
|
||||
...buildOpenAICodexForwardCompatExpectation("gpt-5.4"),
|
||||
api: "openai-responses",
|
||||
// api stays "openai-codex-responses" after plugin normalization
|
||||
input: ["text"],
|
||||
contextWindow: 272000,
|
||||
maxTokens: 128000,
|
||||
|
||||
@ -304,6 +304,7 @@ function preserveDiscoveredTransportMetadata(params: {
|
||||
input: configuredModel?.input ?? discoveredModel.input ?? dynamicModel.input,
|
||||
compat: configuredModel?.compat ?? discoveredModel.compat ?? dynamicModel.compat,
|
||||
maxTokens: configuredModel?.maxTokens ?? discoveredModel.maxTokens ?? dynamicModel.maxTokens,
|
||||
cost: discoveredModel.cost ?? dynamicModel.cost,
|
||||
headers:
|
||||
discoveredHeaders || dynamicHeaders
|
||||
? {
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user