From f736a2504f202df1a9bae35b858d2777997ac0b6 Mon Sep 17 00:00:00 2001 From: Luca Forstner Date: Wed, 8 Apr 2026 12:35:48 +0200 Subject: [PATCH 1/3] chore: Pin models and increase max tokens in e2e/canary tests --- .../__snapshots__/ai-sdk-v3.log-payloads.json | 10 ++--- .../__snapshots__/ai-sdk-v3.span-events.json | 10 ++--- .../__snapshots__/ai-sdk-v4.log-payloads.json | 10 ++--- .../__snapshots__/ai-sdk-v4.span-events.json | 10 ++--- .../__snapshots__/ai-sdk-v5.log-payloads.json | 14 +++--- .../__snapshots__/ai-sdk-v5.span-events.json | 14 +++--- .../__snapshots__/ai-sdk-v6.log-payloads.json | 14 +++--- .../__snapshots__/ai-sdk-v6.span-events.json | 14 +++--- .../ai-sdk-instrumentation/scenario.impl.mjs | 10 ++--- .../ai-sdk-otel-export/scenario.impl.ts | 4 +- .../anthropic-v0712.log-payloads.json | 2 +- .../anthropic-v0712.span-events.json | 2 +- .../anthropic-v0730.log-payloads.json | 2 +- .../anthropic-v0730.span-events.json | 2 +- .../anthropic-v0780.log-payloads.json | 2 +- .../anthropic-v0780.span-events.json | 2 +- .../anthropic-v0800.log-payloads.json | 2 +- .../anthropic-v0800.span-events.json | 2 +- .../scenario.impl.mjs | 8 ++-- .../claude-agent-sdk-v0.1.span-events.json | 2 +- .../scenario.impl.mjs | 2 +- .../google-genai-v1300.log-payloads.json | 8 ++-- .../google-genai-v1440.log-payloads.json | 8 ++-- .../google-genai-v1450.log-payloads.json | 8 ++-- .../google-genai-v1460.log-payloads.json | 8 ++-- .../assertions.ts | 12 ++--- .../scenario.impl.mjs | 10 ++--- .../mistral-instrumentation/constants.mjs | 4 +- .../mistral-instrumentation/scenario.impl.mjs | 6 +-- .../__snapshots__/openai-v4.log-payloads.json | 16 +++---- .../__snapshots__/openai-v4.span-events.json | 16 +++---- .../__snapshots__/openai-v5.log-payloads.json | 16 +++---- .../__snapshots__/openai-v5.span-events.json | 16 +++---- .../__snapshots__/openai-v6.log-payloads.json | 16 +++---- .../__snapshots__/openai-v6.span-events.json | 16 +++---- .../openai-instrumentation/scenario.impl.mjs | 24 +++++----- .../__snapshots__/span-events.json | 10 ++--- .../openrouter-instrumentation/constants.mjs | 2 +- .../scenario.impl.mjs | 10 ++--- .../__snapshots__/log-payloads.json | 44 +++++++++---------- .../wrap-langchain-js-traces/scenario.ts | 4 +- 41 files changed, 196 insertions(+), 196 deletions(-) diff --git a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v3.log-payloads.json b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v3.log-payloads.json index 296e955c6..88bc57fb6 100644 --- a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v3.log-payloads.json +++ b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v3.log-payloads.json @@ -21,7 +21,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateText", @@ -40,7 +40,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "streamText", @@ -59,7 +59,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateText", @@ -90,7 +90,7 @@ "schema": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateObject", @@ -110,7 +110,7 @@ "schema": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "streamObject", diff --git a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v3.span-events.json b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v3.span-events.json index d268ebe0b..f3e698e12 100644 --- a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v3.span-events.json +++ b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v3.span-events.json @@ -28,7 +28,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateText", @@ -55,7 +55,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "streamText", @@ -82,7 +82,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateText", @@ -120,7 +120,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateObject", @@ -147,7 +147,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "streamObject", diff --git a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v4.log-payloads.json b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v4.log-payloads.json index f46eda730..3fd5885d3 100644 --- a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v4.log-payloads.json +++ b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v4.log-payloads.json @@ -21,7 +21,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateText", @@ -40,7 +40,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "streamText", @@ -59,7 +59,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateText", @@ -98,7 +98,7 @@ "schema": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateObject", @@ -118,7 +118,7 @@ "schema": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "streamObject", diff --git a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v4.span-events.json b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v4.span-events.json index d268ebe0b..f3e698e12 100644 --- a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v4.span-events.json +++ b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v4.span-events.json @@ -28,7 +28,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateText", @@ -55,7 +55,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "streamText", @@ -82,7 +82,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateText", @@ -120,7 +120,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "generateObject", @@ -147,7 +147,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.chat" }, "name": "streamObject", diff --git a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v5.log-payloads.json b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v5.log-payloads.json index 438b4beb1..f2716f4b2 100644 --- a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v5.log-payloads.json +++ b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v5.log-payloads.json @@ -21,7 +21,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateText", @@ -40,7 +40,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "streamText", @@ -59,7 +59,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateText", @@ -347,7 +347,7 @@ "schema": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateObject", @@ -367,7 +367,7 @@ "schema": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "streamObject", @@ -390,7 +390,7 @@ ] }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "Agent.generate", @@ -413,7 +413,7 @@ ] }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "Agent.stream", diff --git a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v5.span-events.json b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v5.span-events.json index 5fe97b20e..94e80a747 100644 --- a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v5.span-events.json +++ b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v5.span-events.json @@ -28,7 +28,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateText", @@ -55,7 +55,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "streamText", @@ -82,7 +82,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateText", @@ -153,7 +153,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateObject", @@ -180,7 +180,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "streamObject", @@ -207,7 +207,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "Agent.generate", @@ -234,7 +234,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "Agent.stream", diff --git a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v6.log-payloads.json b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v6.log-payloads.json index 2262c7b08..466001273 100644 --- a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v6.log-payloads.json +++ b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v6.log-payloads.json @@ -21,7 +21,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateText", @@ -40,7 +40,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "streamText", @@ -59,7 +59,7 @@ "prompt": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateText", @@ -377,7 +377,7 @@ "schema": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateObject", @@ -397,7 +397,7 @@ "schema": "" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "streamObject", @@ -420,7 +420,7 @@ ] }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "ToolLoopAgent.generate", @@ -443,7 +443,7 @@ ] }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "ToolLoopAgent.stream", diff --git a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v6.span-events.json b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v6.span-events.json index 84d55a6e7..96c99d2f7 100644 --- a/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v6.span-events.json +++ b/e2e/scenarios/ai-sdk-instrumentation/__snapshots__/ai-sdk-v6.span-events.json @@ -28,7 +28,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateText", @@ -55,7 +55,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "streamText", @@ -82,7 +82,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateText", @@ -153,7 +153,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "generateObject", @@ -180,7 +180,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "streamObject", @@ -207,7 +207,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "ToolLoopAgent.generate", @@ -234,7 +234,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai.responses" }, "name": "ToolLoopAgent.stream", diff --git a/e2e/scenarios/ai-sdk-instrumentation/scenario.impl.mjs b/e2e/scenarios/ai-sdk-instrumentation/scenario.impl.mjs index 79c00f54d..1ce29c7a4 100644 --- a/e2e/scenarios/ai-sdk-instrumentation/scenario.impl.mjs +++ b/e2e/scenarios/ai-sdk-instrumentation/scenario.impl.mjs @@ -124,7 +124,7 @@ async function runAISDKInstrumentationScenario( { decorateAI, flushCount, flushDelayMs } = {}, ) { const instrumentedAI = decorateAI ? decorateAI(options.ai) : options.ai; - const openaiModel = options.openai("gpt-4o-mini"); + const openaiModel = options.openai("gpt-4o-mini-2024-07-18"); const sdkMajorVersion = parseMajorVersion(options.sdkVersion); const supportsRichInputScenarios = sdkMajorVersion >= 5; const outputObject = createOutputObjectIfSupported(options.ai); @@ -136,7 +136,7 @@ async function runAISDKInstrumentationScenario( model: openaiModel, prompt: "Reply with the single token PARIS and no punctuation.", temperature: 0, - ...tokenLimit(options.maxTokensKey, 16), + ...tokenLimit(options.maxTokensKey, 24), }); }); @@ -200,7 +200,7 @@ async function runAISDKInstrumentationScenario( model: openaiModel, prompt: "Reply with the word DENIED and nothing else.", temperature: 0, - ...tokenLimit(options.maxTokensKey, 16), + ...tokenLimit(options.maxTokensKey, 24), }; params[DENY_OUTPUT_PATHS_SYMBOL] = ["text", "_output"]; await instrumentedAI.generateText(params); @@ -264,7 +264,7 @@ async function runAISDKInstrumentationScenario( content: "Reply with exactly HELLO and no punctuation.", }, ], - ...tokenLimit(options.maxTokensKey, 16), + ...tokenLimit(options.maxTokensKey, 24), }); }, ); @@ -286,7 +286,7 @@ async function runAISDKInstrumentationScenario( "Reply with exactly STREAM HELLO and no punctuation.", }, ], - ...tokenLimit(options.maxTokensKey, 16), + ...tokenLimit(options.maxTokensKey, 24), }); for await (const _chunk of result.textStream) { } diff --git a/e2e/scenarios/ai-sdk-otel-export/scenario.impl.ts b/e2e/scenarios/ai-sdk-otel-export/scenario.impl.ts index a4b71b870..e11baa6b7 100644 --- a/e2e/scenarios/ai-sdk-otel-export/scenario.impl.ts +++ b/e2e/scenarios/ai-sdk-otel-export/scenario.impl.ts @@ -43,7 +43,7 @@ export async function runAISDKOtelExport(options: AISDKOtelScenarioOptions) { const { trace } = await import("@opentelemetry/api"); trace.setGlobalTracerProvider(provider); - const model = options.openai("gpt-4o-mini") as any; + const model = options.openai("gpt-4o-mini-2024-07-18") as any; const telemetryBase = { isEnabled: true, metadata: { @@ -58,7 +58,7 @@ export async function runAISDKOtelExport(options: AISDKOtelScenarioOptions) { model, prompt: "Reply with the single token PARIS and no punctuation.", temperature: 0, - ...tokenLimit(options.maxTokensKey, 16), + ...tokenLimit(options.maxTokensKey, 24), experimental_telemetry: { ...telemetryBase, functionId: "otel-generate", diff --git a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0712.log-payloads.json b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0712.log-payloads.json index 65b25fc47..b46bfc0c0 100644 --- a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0712.log-payloads.json +++ b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0712.log-payloads.json @@ -321,7 +321,7 @@ } ], "metadata": { - "model": "claude-sonnet-4-5", + "model": "claude-sonnet-4-5-20250929", "provider": "anthropic", "stop_reason": "end_turn", "stop_sequence": null diff --git a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0712.span-events.json b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0712.span-events.json index c5d430b8d..8d20ad958 100644 --- a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0712.span-events.json +++ b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0712.span-events.json @@ -259,7 +259,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "claude-sonnet-4-5", + "model": "claude-sonnet-4-5-20250929", "provider": "anthropic" }, "metric_keys": [ diff --git a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0730.log-payloads.json b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0730.log-payloads.json index 65b25fc47..b46bfc0c0 100644 --- a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0730.log-payloads.json +++ b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0730.log-payloads.json @@ -321,7 +321,7 @@ } ], "metadata": { - "model": "claude-sonnet-4-5", + "model": "claude-sonnet-4-5-20250929", "provider": "anthropic", "stop_reason": "end_turn", "stop_sequence": null diff --git a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0730.span-events.json b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0730.span-events.json index c5d430b8d..8d20ad958 100644 --- a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0730.span-events.json +++ b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0730.span-events.json @@ -259,7 +259,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "claude-sonnet-4-5", + "model": "claude-sonnet-4-5-20250929", "provider": "anthropic" }, "metric_keys": [ diff --git a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0780.log-payloads.json b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0780.log-payloads.json index 65b25fc47..b46bfc0c0 100644 --- a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0780.log-payloads.json +++ b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0780.log-payloads.json @@ -321,7 +321,7 @@ } ], "metadata": { - "model": "claude-sonnet-4-5", + "model": "claude-sonnet-4-5-20250929", "provider": "anthropic", "stop_reason": "end_turn", "stop_sequence": null diff --git a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0780.span-events.json b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0780.span-events.json index c5d430b8d..8d20ad958 100644 --- a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0780.span-events.json +++ b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0780.span-events.json @@ -259,7 +259,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "claude-sonnet-4-5", + "model": "claude-sonnet-4-5-20250929", "provider": "anthropic" }, "metric_keys": [ diff --git a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0800.log-payloads.json b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0800.log-payloads.json index 65b25fc47..b46bfc0c0 100644 --- a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0800.log-payloads.json +++ b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0800.log-payloads.json @@ -321,7 +321,7 @@ } ], "metadata": { - "model": "claude-sonnet-4-5", + "model": "claude-sonnet-4-5-20250929", "provider": "anthropic", "stop_reason": "end_turn", "stop_sequence": null diff --git a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0800.span-events.json b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0800.span-events.json index c5d430b8d..8d20ad958 100644 --- a/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0800.span-events.json +++ b/e2e/scenarios/anthropic-instrumentation/__snapshots__/anthropic-v0800.span-events.json @@ -259,7 +259,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "claude-sonnet-4-5", + "model": "claude-sonnet-4-5-20250929", "provider": "anthropic" }, "metric_keys": [ diff --git a/e2e/scenarios/anthropic-instrumentation/scenario.impl.mjs b/e2e/scenarios/anthropic-instrumentation/scenario.impl.mjs index ce27aefe5..3cc3f4e49 100644 --- a/e2e/scenarios/anthropic-instrumentation/scenario.impl.mjs +++ b/e2e/scenarios/anthropic-instrumentation/scenario.impl.mjs @@ -41,7 +41,7 @@ async function runAnthropicInstrumentationScenario( await runOperation("anthropic-create-operation", "create", async () => { await client.messages.create({ model: ANTHROPIC_MODEL, - max_tokens: 16, + max_tokens: 24, temperature: 0, messages: [{ role: "user", content: "Reply with exactly OK." }], }); @@ -54,7 +54,7 @@ async function runAnthropicInstrumentationScenario( const response = await client.messages .create({ model: ANTHROPIC_MODEL, - max_tokens: 16, + max_tokens: 24, temperature: 0, messages: [ { @@ -184,7 +184,7 @@ async function runAnthropicInstrumentationScenario( "stream-thinking", async () => { const stream = await client.messages.create({ - model: "claude-sonnet-4-5", + model: "claude-sonnet-4-5-20250929", max_tokens: 2048, temperature: 1, thinking: { @@ -211,7 +211,7 @@ async function runAnthropicInstrumentationScenario( async () => { await client.beta.messages.create({ model: ANTHROPIC_MODEL, - max_tokens: 16, + max_tokens: 24, temperature: 0, messages: [{ role: "user", content: "Reply with exactly BETA." }], }); diff --git a/e2e/scenarios/claude-agent-sdk-instrumentation/__snapshots__/claude-agent-sdk-v0.1.span-events.json b/e2e/scenarios/claude-agent-sdk-instrumentation/__snapshots__/claude-agent-sdk-v0.1.span-events.json index 57289ec3f..07440521d 100644 --- a/e2e/scenarios/claude-agent-sdk-instrumentation/__snapshots__/claude-agent-sdk-v0.1.span-events.json +++ b/e2e/scenarios/claude-agent-sdk-instrumentation/__snapshots__/claude-agent-sdk-v0.1.span-events.json @@ -240,7 +240,7 @@ "has_output": false, "metadata": null, "metric_keys": [], - "name": "Agent: math-expert", + "name": "Agent: general-purpose", "root_span_id": "", "span_id": "", "span_parents": [ diff --git a/e2e/scenarios/claude-agent-sdk-instrumentation/scenario.impl.mjs b/e2e/scenarios/claude-agent-sdk-instrumentation/scenario.impl.mjs index 879c14416..f5417f4c4 100644 --- a/e2e/scenarios/claude-agent-sdk-instrumentation/scenario.impl.mjs +++ b/e2e/scenarios/claude-agent-sdk-instrumentation/scenario.impl.mjs @@ -121,7 +121,7 @@ async function runClaudeAgentSDKScenario({ decorateSDK, sdk }) { agents: { "math-expert": { description: "Math specialist", - model: "haiku", + model: CLAUDE_AGENT_MODEL, prompt: "You are a math expert. Use the calculator tool for calculations. Be concise.", }, diff --git a/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1300.log-payloads.json b/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1300.log-payloads.json index ac9c77164..3916b54df 100644 --- a/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1300.log-payloads.json +++ b/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1300.log-payloads.json @@ -24,7 +24,7 @@ { "input": { "config": { - "maxOutputTokens": 16, + "maxOutputTokens": 24, "temperature": 0 }, "contents": { @@ -105,7 +105,7 @@ { "input": { "config": { - "maxOutputTokens": 24, + "maxOutputTokens": 48, "temperature": 0 }, "contents": [ @@ -207,7 +207,7 @@ { "input": { "config": { - "maxOutputTokens": 32, + "maxOutputTokens": 64, "temperature": 0 }, "contents": { @@ -271,7 +271,7 @@ { "input": { "config": { - "maxOutputTokens": 16, + "maxOutputTokens": 24, "temperature": 0 }, "contents": { diff --git a/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1440.log-payloads.json b/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1440.log-payloads.json index ac9c77164..3916b54df 100644 --- a/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1440.log-payloads.json +++ b/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1440.log-payloads.json @@ -24,7 +24,7 @@ { "input": { "config": { - "maxOutputTokens": 16, + "maxOutputTokens": 24, "temperature": 0 }, "contents": { @@ -105,7 +105,7 @@ { "input": { "config": { - "maxOutputTokens": 24, + "maxOutputTokens": 48, "temperature": 0 }, "contents": [ @@ -207,7 +207,7 @@ { "input": { "config": { - "maxOutputTokens": 32, + "maxOutputTokens": 64, "temperature": 0 }, "contents": { @@ -271,7 +271,7 @@ { "input": { "config": { - "maxOutputTokens": 16, + "maxOutputTokens": 24, "temperature": 0 }, "contents": { diff --git a/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1450.log-payloads.json b/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1450.log-payloads.json index ac9c77164..3916b54df 100644 --- a/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1450.log-payloads.json +++ b/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1450.log-payloads.json @@ -24,7 +24,7 @@ { "input": { "config": { - "maxOutputTokens": 16, + "maxOutputTokens": 24, "temperature": 0 }, "contents": { @@ -105,7 +105,7 @@ { "input": { "config": { - "maxOutputTokens": 24, + "maxOutputTokens": 48, "temperature": 0 }, "contents": [ @@ -207,7 +207,7 @@ { "input": { "config": { - "maxOutputTokens": 32, + "maxOutputTokens": 64, "temperature": 0 }, "contents": { @@ -271,7 +271,7 @@ { "input": { "config": { - "maxOutputTokens": 16, + "maxOutputTokens": 24, "temperature": 0 }, "contents": { diff --git a/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1460.log-payloads.json b/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1460.log-payloads.json index ac9c77164..3916b54df 100644 --- a/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1460.log-payloads.json +++ b/e2e/scenarios/google-genai-instrumentation/__snapshots__/google-genai-v1460.log-payloads.json @@ -24,7 +24,7 @@ { "input": { "config": { - "maxOutputTokens": 16, + "maxOutputTokens": 24, "temperature": 0 }, "contents": { @@ -105,7 +105,7 @@ { "input": { "config": { - "maxOutputTokens": 24, + "maxOutputTokens": 48, "temperature": 0 }, "contents": [ @@ -207,7 +207,7 @@ { "input": { "config": { - "maxOutputTokens": 32, + "maxOutputTokens": 64, "temperature": 0 }, "contents": { @@ -271,7 +271,7 @@ { "input": { "config": { - "maxOutputTokens": 16, + "maxOutputTokens": 24, "temperature": 0 }, "contents": { diff --git a/e2e/scenarios/google-genai-instrumentation/assertions.ts b/e2e/scenarios/google-genai-instrumentation/assertions.ts index f0baf930e..dac27b78a 100644 --- a/e2e/scenarios/google-genai-instrumentation/assertions.ts +++ b/e2e/scenarios/google-genai-instrumentation/assertions.ts @@ -9,7 +9,7 @@ import { withScenarioHarness } from "../../helpers/scenario-harness"; import { findChildSpans, findLatestSpan } from "../../helpers/trace-selectors"; import { summarizeWrapperContract } from "../../helpers/wrapper-contract"; -import { ROOT_NAME, SCENARIO_NAME } from "./scenario.impl.mjs"; +import { GOOGLE_MODEL, ROOT_NAME, SCENARIO_NAME } from "./scenario.impl.mjs"; type RunGoogleGenAIScenario = (harness: { runNodeScenarioDir: (options: { @@ -351,7 +351,7 @@ export function defineGoogleGenAIInstrumentationAssertions(options: { expect(span).toBeDefined(); expect(operation?.span.parentIds).toEqual([root?.span.id ?? ""]); expect(span?.row.metadata).toMatchObject({ - model: "gemini-2.5-flash-lite", + model: GOOGLE_MODEL, }); }, ); @@ -368,7 +368,7 @@ export function defineGoogleGenAIInstrumentationAssertions(options: { expect(span).toBeDefined(); expect(operation?.span.parentIds).toEqual([root?.span.id ?? ""]); expect(span?.row.metadata).toMatchObject({ - model: "gemini-2.5-flash-lite", + model: GOOGLE_MODEL, }); expect(JSON.stringify(span?.input)).toContain("file.png"); }); @@ -388,7 +388,7 @@ export function defineGoogleGenAIInstrumentationAssertions(options: { expect(span).toBeDefined(); expect(operation?.span.parentIds).toEqual([root?.span.id ?? ""]); expect(span?.row.metadata).toMatchObject({ - model: "gemini-2.5-flash-lite", + model: GOOGLE_MODEL, }); expect(span?.metrics).toMatchObject({ time_to_first_token: expect.any(Number), @@ -416,7 +416,7 @@ export function defineGoogleGenAIInstrumentationAssertions(options: { expect(span).toBeDefined(); expect(operation?.span.parentIds).toEqual([root?.span.id ?? ""]); expect(span?.row.metadata).toMatchObject({ - model: "gemini-2.5-flash-lite", + model: GOOGLE_MODEL, }); expect(span?.metrics).toMatchObject({ time_to_first_token: expect.any(Number), @@ -449,7 +449,7 @@ export function defineGoogleGenAIInstrumentationAssertions(options: { expect(span).toBeDefined(); expect(operation?.span.parentIds).toEqual([root?.span.id ?? ""]); expect(span?.row.metadata).toMatchObject({ - model: "gemini-2.5-flash-lite", + model: GOOGLE_MODEL, }); expect( output?.functionCalls?.some((call) => call.name === "get_weather") || diff --git a/e2e/scenarios/google-genai-instrumentation/scenario.impl.mjs b/e2e/scenarios/google-genai-instrumentation/scenario.impl.mjs index 31ce210dd..1ff56ef23 100644 --- a/e2e/scenarios/google-genai-instrumentation/scenario.impl.mjs +++ b/e2e/scenarios/google-genai-instrumentation/scenario.impl.mjs @@ -45,7 +45,7 @@ async function runGoogleGenAIInstrumentationScenario(sdk, options = {}) { model: GOOGLE_MODEL, contents: "Reply with exactly PARIS.", config: { - maxOutputTokens: 16, + maxOutputTokens: 24, temperature: 0, }, }); @@ -74,7 +74,7 @@ async function runGoogleGenAIInstrumentationScenario(sdk, options = {}) { }, ], config: { - maxOutputTokens: 24, + maxOutputTokens: 48, temperature: 0, }, }); @@ -86,7 +86,7 @@ async function runGoogleGenAIInstrumentationScenario(sdk, options = {}) { model: GOOGLE_MODEL, contents: "Count from 1 to 3 and include the words one two three.", config: { - maxOutputTokens: 32, + maxOutputTokens: 64, temperature: 0, }, }); @@ -101,7 +101,7 @@ async function runGoogleGenAIInstrumentationScenario(sdk, options = {}) { model: GOOGLE_MODEL, contents: "Reply with exactly BONJOUR.", config: { - maxOutputTokens: 16, + maxOutputTokens: 24, temperature: 0, }, }); @@ -149,4 +149,4 @@ export async function runAutoGoogleGenAIInstrumentation(sdk) { await runGoogleGenAIInstrumentationScenario(sdk); } -export { ROOT_NAME, SCENARIO_NAME }; +export { GOOGLE_MODEL, ROOT_NAME, SCENARIO_NAME }; diff --git a/e2e/scenarios/mistral-instrumentation/constants.mjs b/e2e/scenarios/mistral-instrumentation/constants.mjs index b4eb0ab83..c3266b467 100644 --- a/e2e/scenarios/mistral-instrumentation/constants.mjs +++ b/e2e/scenarios/mistral-instrumentation/constants.mjs @@ -1,6 +1,6 @@ -const CHAT_MODEL = "mistral-small-latest"; +const CHAT_MODEL = "mistral-small-2506"; const EMBEDDING_MODEL = "mistral-embed"; -const FIM_MODEL = "codestral-latest"; +const FIM_MODEL = "codestral-2508"; const AGENT_MODEL = CHAT_MODEL; const ROOT_NAME = "mistral-root"; const SCENARIO_NAME = "mistral-instrumentation"; diff --git a/e2e/scenarios/mistral-instrumentation/scenario.impl.mjs b/e2e/scenarios/mistral-instrumentation/scenario.impl.mjs index ee61b661c..47d0fe2e4 100644 --- a/e2e/scenarios/mistral-instrumentation/scenario.impl.mjs +++ b/e2e/scenarios/mistral-instrumentation/scenario.impl.mjs @@ -370,7 +370,7 @@ async function runMistralInstrumentationScenario( content: "Reply with exactly: observability", }, ], - maxTokens: 16, + maxTokens: 24, temperature: 0, }), MISTRAL_REQUEST_RETRY_OPTIONS, @@ -569,7 +569,7 @@ async function runMistralInstrumentationScenario( responseFormat: { type: "text", }, - maxTokens: 12, + maxTokens: 16, temperature: 0, }), MISTRAL_REQUEST_RETRY_OPTIONS, @@ -635,7 +635,7 @@ async function runMistralInstrumentationScenario( responseFormat: { type: "text", }, - maxTokens: 12, + maxTokens: 16, stream: true, temperature: 0, }); diff --git a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v4.log-payloads.json b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v4.log-payloads.json index f91eda777..db9f2d590 100644 --- a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v4.log-payloads.json +++ b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v4.log-payloads.json @@ -36,7 +36,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -73,7 +73,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -110,7 +110,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -147,7 +147,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -184,7 +184,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -223,7 +223,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -288,7 +288,7 @@ "kind": "text" }, "metadata": { - "model": "omni-moderation-latest", + "model": "omni-moderation-2024-09-26", "provider": "openai" }, "metrics": { @@ -476,7 +476,7 @@ "kind": "text" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { diff --git a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v4.span-events.json b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v4.span-events.json index 4e97348fa..4dc6b4bde 100644 --- a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v4.span-events.json +++ b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v4.span-events.json @@ -28,7 +28,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -53,7 +53,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -78,7 +78,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -103,7 +103,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -128,7 +128,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -153,7 +153,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -203,7 +203,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "omni-moderation-latest", + "model": "omni-moderation-2024-09-26", "provider": "openai" }, "metrics": { @@ -328,7 +328,7 @@ "has_input": true, "has_output": false, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { diff --git a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v5.log-payloads.json b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v5.log-payloads.json index f91eda777..db9f2d590 100644 --- a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v5.log-payloads.json +++ b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v5.log-payloads.json @@ -36,7 +36,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -73,7 +73,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -110,7 +110,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -147,7 +147,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -184,7 +184,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -223,7 +223,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -288,7 +288,7 @@ "kind": "text" }, "metadata": { - "model": "omni-moderation-latest", + "model": "omni-moderation-2024-09-26", "provider": "openai" }, "metrics": { @@ -476,7 +476,7 @@ "kind": "text" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { diff --git a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v5.span-events.json b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v5.span-events.json index 4e97348fa..4dc6b4bde 100644 --- a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v5.span-events.json +++ b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v5.span-events.json @@ -28,7 +28,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -53,7 +53,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -78,7 +78,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -103,7 +103,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -128,7 +128,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -153,7 +153,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -203,7 +203,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "omni-moderation-latest", + "model": "omni-moderation-2024-09-26", "provider": "openai" }, "metrics": { @@ -328,7 +328,7 @@ "has_input": true, "has_output": false, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { diff --git a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.log-payloads.json b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.log-payloads.json index f91eda777..db9f2d590 100644 --- a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.log-payloads.json +++ b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.log-payloads.json @@ -36,7 +36,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -73,7 +73,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -110,7 +110,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -147,7 +147,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -184,7 +184,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -223,7 +223,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -288,7 +288,7 @@ "kind": "text" }, "metadata": { - "model": "omni-moderation-latest", + "model": "omni-moderation-2024-09-26", "provider": "openai" }, "metrics": { @@ -476,7 +476,7 @@ "kind": "text" }, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { diff --git a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.span-events.json b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.span-events.json index 4e97348fa..4dc6b4bde 100644 --- a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.span-events.json +++ b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.span-events.json @@ -28,7 +28,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -53,7 +53,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -78,7 +78,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -103,7 +103,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -128,7 +128,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -153,7 +153,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { @@ -203,7 +203,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "omni-moderation-latest", + "model": "omni-moderation-2024-09-26", "provider": "openai" }, "metrics": { @@ -328,7 +328,7 @@ "has_input": true, "has_output": false, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { diff --git a/e2e/scenarios/openai-instrumentation/scenario.impl.mjs b/e2e/scenarios/openai-instrumentation/scenario.impl.mjs index 99771e5b0..968c855cb 100644 --- a/e2e/scenarios/openai-instrumentation/scenario.impl.mjs +++ b/e2e/scenarios/openai-instrumentation/scenario.impl.mjs @@ -4,9 +4,9 @@ import { runTracedScenario, } from "../../helpers/provider-runtime.mjs"; -const OPENAI_MODEL = "gpt-4o-mini"; +const OPENAI_MODEL = "gpt-4o-mini-2024-07-18"; const EMBEDDING_MODEL = "text-embedding-3-small"; -const MODERATION_MODEL = "omni-moderation-latest"; +const MODERATION_MODEL = "omni-moderation-2024-09-26"; const ROOT_NAME = "openai-instrumentation-root"; const SCENARIO_NAME = "openai-instrumentation"; @@ -110,7 +110,7 @@ export async function runOpenAIInstrumentationScenario(options) { await client.chat.completions.create({ model: OPENAI_MODEL, messages: [{ role: "user", content: "Reply with exactly OK." }], - max_tokens: 8, + max_tokens: 12, temperature: 0, }); }); @@ -123,7 +123,7 @@ export async function runOpenAIInstrumentationScenario(options) { client.chat.completions.create({ model: OPENAI_MODEL, messages: [{ role: "user", content: "Reply with exactly FOUR." }], - max_tokens: 8, + max_tokens: 12, temperature: 0, }), ); @@ -135,7 +135,7 @@ export async function runOpenAIInstrumentationScenario(options) { model: OPENAI_MODEL, messages: [{ role: "user", content: "Reply with exactly STREAM." }], stream: true, - max_tokens: 8, + max_tokens: 12, temperature: 0, stream_options: { include_usage: true, @@ -158,7 +158,7 @@ export async function runOpenAIInstrumentationScenario(options) { }, ], stream: true, - max_tokens: 16, + max_tokens: 24, temperature: 0, stream_options: { include_usage: true, @@ -200,7 +200,7 @@ export async function runOpenAIInstrumentationScenario(options) { messages: [ { role: "user", content: "Reply with exactly SYNC STREAM." }, ], - max_tokens: 16, + max_tokens: 24, temperature: 0, }; @@ -252,7 +252,7 @@ export async function runOpenAIInstrumentationScenario(options) { await client.responses.create({ model: OPENAI_MODEL, input: "Reply with exactly PARIS.", - max_output_tokens: 16, + max_output_tokens: 24, }); }, ); @@ -265,7 +265,7 @@ export async function runOpenAIInstrumentationScenario(options) { client.responses.create({ model: OPENAI_MODEL, input: "What is 2 + 2? Reply with just the number.", - max_output_tokens: 16, + max_output_tokens: 24, }), ); }, @@ -279,7 +279,7 @@ export async function runOpenAIInstrumentationScenario(options) { client.responses.create({ model: OPENAI_MODEL, input: "Reply with exactly RESPONSE STREAM.", - max_output_tokens: 16, + max_output_tokens: 24, stream: true, }), ); @@ -294,7 +294,7 @@ export async function runOpenAIInstrumentationScenario(options) { const stream = client.responses.stream({ model: OPENAI_MODEL, input: "What is 6 x 6? Reply with just the number.", - max_output_tokens: 16, + max_output_tokens: 24, }); await collectAsync(stream); await stream.finalResponse(); @@ -308,7 +308,7 @@ export async function runOpenAIInstrumentationScenario(options) { const stream = client.responses.stream({ model: OPENAI_MODEL, input: "Reply with exactly PARTIAL.", - max_output_tokens: 16, + max_output_tokens: 24, }); await collectOneAndReturn(stream); }, diff --git a/e2e/scenarios/openrouter-instrumentation/__snapshots__/span-events.json b/e2e/scenarios/openrouter-instrumentation/__snapshots__/span-events.json index 551614c24..860eb5c70 100644 --- a/e2e/scenarios/openrouter-instrumentation/__snapshots__/span-events.json +++ b/e2e/scenarios/openrouter-instrumentation/__snapshots__/span-events.json @@ -31,7 +31,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metric_keys": [ @@ -73,7 +73,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metric_keys": [ @@ -151,7 +151,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metric_keys": [ @@ -193,7 +193,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metric_keys": [ @@ -235,7 +235,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metric_keys": [ diff --git a/e2e/scenarios/openrouter-instrumentation/constants.mjs b/e2e/scenarios/openrouter-instrumentation/constants.mjs index ca393804d..c961d045b 100644 --- a/e2e/scenarios/openrouter-instrumentation/constants.mjs +++ b/e2e/scenarios/openrouter-instrumentation/constants.mjs @@ -1,4 +1,4 @@ -const CHAT_MODEL = "openai/gpt-4o-mini"; +const CHAT_MODEL = "openai/gpt-4o-mini-2024-07-18"; const EMBEDDING_MODEL = "openai/text-embedding-3-small"; const ROOT_NAME = "openrouter-root"; const SCENARIO_NAME = "openrouter-instrumentation"; diff --git a/e2e/scenarios/openrouter-instrumentation/scenario.impl.mjs b/e2e/scenarios/openrouter-instrumentation/scenario.impl.mjs index 49117a548..b9661cc56 100644 --- a/e2e/scenarios/openrouter-instrumentation/scenario.impl.mjs +++ b/e2e/scenarios/openrouter-instrumentation/scenario.impl.mjs @@ -61,7 +61,7 @@ async function runOpenRouterInstrumentationScenario( withCompatibleChatRequest({ model: CHAT_MODEL, messages: [{ role: "user", content: "Reply with exactly OK." }], - maxTokens: 16, + maxTokens: 24, temperature: 0, }), ); @@ -77,7 +77,7 @@ async function runOpenRouterInstrumentationScenario( messages: [ { role: "user", content: "Reply with exactly STREAM." }, ], - maxTokens: 16, + maxTokens: 24, stream: true, streamOptions: { includeUsage: true, @@ -110,7 +110,7 @@ async function runOpenRouterInstrumentationScenario( await client.beta.responses.send( withCompatibleResponsesRequest({ input: "Reply with exactly OBSERVABILITY.", - maxOutputTokens: 16, + maxOutputTokens: 24, model: CHAT_MODEL, temperature: 0, }), @@ -125,7 +125,7 @@ async function runOpenRouterInstrumentationScenario( const stream = await client.beta.responses.send( withCompatibleResponsesRequest({ input: "Reply with exactly STREAMED RESPONSE.", - maxOutputTokens: 16, + maxOutputTokens: 24, model: CHAT_MODEL, stream: true, temperature: 0, @@ -142,7 +142,7 @@ async function runOpenRouterInstrumentationScenario( const result = client.callModel({ input: "Use the lookup_weather tool for Vienna exactly once, then answer with only the forecast.", - maxOutputTokens: 16, + maxOutputTokens: 24, maxToolCalls: 1, model: CHAT_MODEL, temperature: 0, diff --git a/e2e/scenarios/wrap-langchain-js-traces/__snapshots__/log-payloads.json b/e2e/scenarios/wrap-langchain-js-traces/__snapshots__/log-payloads.json index b3852b50a..5811b1866 100644 --- a/e2e/scenarios/wrap-langchain-js-traces/__snapshots__/log-payloads.json +++ b/e2e/scenarios/wrap-langchain-js-traces/__snapshots__/log-payloads.json @@ -114,15 +114,15 @@ "sdk_language": "javascript" }, "invocation_params": { - "max_tokens": 16, - "model": "gpt-4o-mini", + "max_tokens": 24, + "model": "gpt-4o-mini-2024-07-18", "stream": false, "temperature": 0 }, "metadata": { "ls_integration": "langchain_chat_model", - "ls_max_tokens": 16, - "ls_model_name": "gpt-4o-mini", + "ls_max_tokens": 24, + "ls_model_name": "gpt-4o-mini-2024-07-18", "ls_model_type": "chat", "ls_provider": "openai", "ls_temperature": 0, @@ -141,8 +141,8 @@ "ChatOpenAI" ], "kwargs": { - "max_tokens": 16, - "model": "gpt-4o-mini", + "max_tokens": 24, + "model": "gpt-4o-mini-2024-07-18", "openai_api_key": { "id": [ "OPENAI_API_KEY" @@ -418,7 +418,7 @@ ], "kwargs": { "max_tokens": 32, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "openai_api_key": { "id": [ "OPENAI_API_KEY" @@ -689,14 +689,14 @@ }, "invocation_params": { "max_tokens": 32, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "stream": false, "temperature": 0 }, "metadata": { "ls_integration": "langchain_chat_model", "ls_max_tokens": 32, - "ls_model_name": "gpt-4o-mini", + "ls_model_name": "gpt-4o-mini-2024-07-18", "ls_model_type": "chat", "ls_provider": "openai", "ls_temperature": 0, @@ -717,7 +717,7 @@ ], "kwargs": { "max_tokens": 32, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "openai_api_key": { "id": [ "OPENAI_API_KEY" @@ -951,7 +951,7 @@ }, "invocation_params": { "max_tokens": 32, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "stream": true, "stream_options": { "include_usage": true @@ -961,7 +961,7 @@ "metadata": { "ls_integration": "langchain_chat_model", "ls_max_tokens": 32, - "ls_model_name": "gpt-4o-mini", + "ls_model_name": "gpt-4o-mini-2024-07-18", "ls_model_type": "chat", "ls_provider": "openai", "ls_temperature": 0, @@ -981,7 +981,7 @@ ], "kwargs": { "max_tokens": 32, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "openai_api_key": { "id": [ "OPENAI_API_KEY" @@ -1216,7 +1216,7 @@ }, "invocation_params": { "max_tokens": 128, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "stream": false, "temperature": 0, "tools": [ @@ -1246,7 +1246,7 @@ "metadata": { "ls_integration": "langchain_chat_model", "ls_max_tokens": 128, - "ls_model_name": "gpt-4o-mini", + "ls_model_name": "gpt-4o-mini-2024-07-18", "ls_model_type": "chat", "ls_provider": "openai", "ls_temperature": 0, @@ -1290,7 +1290,7 @@ ], "kwargs": { "max_tokens": 128, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "openai_api_key": { "id": [ "OPENAI_API_KEY" @@ -1540,7 +1540,7 @@ }, "invocation_params": { "max_tokens": 128, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "stream": false, "temperature": 0, "tools": [ @@ -1583,7 +1583,7 @@ "metadata": { "ls_integration": "langchain_chat_model", "ls_max_tokens": 128, - "ls_model_name": "gpt-4o-mini", + "ls_model_name": "gpt-4o-mini-2024-07-18", "ls_model_type": "chat", "ls_provider": "openai", "ls_temperature": 0, @@ -1640,7 +1640,7 @@ ], "kwargs": { "max_tokens": 128, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "openai_api_key": { "id": [ "OPENAI_API_KEY" @@ -1892,7 +1892,7 @@ }, "invocation_params": { "max_tokens": 128, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "stream": false, "temperature": 0, "tools": [ @@ -1935,7 +1935,7 @@ "metadata": { "ls_integration": "langchain_chat_model", "ls_max_tokens": 128, - "ls_model_name": "gpt-4o-mini", + "ls_model_name": "gpt-4o-mini-2024-07-18", "ls_model_type": "chat", "ls_provider": "openai", "ls_temperature": 0, @@ -1992,7 +1992,7 @@ ], "kwargs": { "max_tokens": 128, - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "openai_api_key": { "id": [ "OPENAI_API_KEY" diff --git a/e2e/scenarios/wrap-langchain-js-traces/scenario.ts b/e2e/scenarios/wrap-langchain-js-traces/scenario.ts index 0f9c0fb61..ff5b00592 100644 --- a/e2e/scenarios/wrap-langchain-js-traces/scenario.ts +++ b/e2e/scenarios/wrap-langchain-js-traces/scenario.ts @@ -11,7 +11,7 @@ import { } from "../../helpers/provider-runtime.mjs"; import { runMain } from "../../helpers/scenario-runtime"; -const OPENAI_MODEL = "gpt-4o-mini"; +const OPENAI_MODEL = "gpt-4o-mini-2024-07-18"; runMain(async () => { await runTracedScenario({ @@ -23,7 +23,7 @@ runMain(async () => { await runOperation("langchain-invoke-operation", "invoke", async () => { const model = new ChatOpenAI({ model: OPENAI_MODEL, - maxTokens: 16, + maxTokens: 24, temperature: 0, callbacks: [handler], }); From 6dc20603a4e40bc7e4465894975019802dab477c Mon Sep 17 00:00:00 2001 From: Luca Forstner Date: Wed, 8 Apr 2026 13:19:32 +0200 Subject: [PATCH 2/3] snapshots --- .../mistral-v1-10-0.log-payloads.json | 20 +++++++++---------- .../mistral-v1-10-0.span-events.json | 20 +++++++++---------- .../mistral-v1-14-1.log-payloads.json | 20 +++++++++---------- .../mistral-v1-14-1.span-events.json | 20 +++++++++---------- .../mistral-v1-15-1.log-payloads.json | 20 +++++++++---------- .../mistral-v1-15-1.span-events.json | 20 +++++++++---------- .../mistral-v1-3-4.log-payloads.json | 20 +++++++++---------- .../mistral-v1-3-4.span-events.json | 20 +++++++++---------- .../mistral-v1.log-payloads.json | 20 +++++++++---------- .../__snapshots__/mistral-v1.span-events.json | 20 +++++++++---------- .../mistral-v2.log-payloads.json | 20 +++++++++---------- .../__snapshots__/mistral-v2.span-events.json | 20 +++++++++---------- 12 files changed, 120 insertions(+), 120 deletions(-) diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-10-0.log-payloads.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-10-0.log-payloads.json index c435f5788..e02cea9c9 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-10-0.log-payloads.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-10-0.log-payloads.json @@ -33,7 +33,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -74,7 +74,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -115,7 +115,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -176,7 +176,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -211,7 +211,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -246,7 +246,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -287,7 +287,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -328,7 +328,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -369,7 +369,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-10-0.span-events.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-10-0.span-events.json index b7d2919bc..ab1e0df51 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-10-0.span-events.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-10-0.span-events.json @@ -31,7 +31,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -67,7 +67,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -103,7 +103,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -124,7 +124,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -181,7 +181,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -217,7 +217,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -253,7 +253,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -289,7 +289,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -325,7 +325,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-14-1.log-payloads.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-14-1.log-payloads.json index c435f5788..e02cea9c9 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-14-1.log-payloads.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-14-1.log-payloads.json @@ -33,7 +33,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -74,7 +74,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -115,7 +115,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -176,7 +176,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -211,7 +211,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -246,7 +246,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -287,7 +287,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -328,7 +328,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -369,7 +369,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-14-1.span-events.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-14-1.span-events.json index b7d2919bc..ab1e0df51 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-14-1.span-events.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-14-1.span-events.json @@ -31,7 +31,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -67,7 +67,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -103,7 +103,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -124,7 +124,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -181,7 +181,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -217,7 +217,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -253,7 +253,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -289,7 +289,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -325,7 +325,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-15-1.log-payloads.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-15-1.log-payloads.json index c435f5788..e02cea9c9 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-15-1.log-payloads.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-15-1.log-payloads.json @@ -33,7 +33,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -74,7 +74,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -115,7 +115,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -176,7 +176,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -211,7 +211,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -246,7 +246,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -287,7 +287,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -328,7 +328,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -369,7 +369,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-15-1.span-events.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-15-1.span-events.json index b7d2919bc..ab1e0df51 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-15-1.span-events.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-15-1.span-events.json @@ -31,7 +31,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -67,7 +67,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -103,7 +103,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -124,7 +124,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -181,7 +181,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -217,7 +217,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -253,7 +253,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -289,7 +289,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -325,7 +325,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-3-4.log-payloads.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-3-4.log-payloads.json index cf15e9eec..19d6a3806 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-3-4.log-payloads.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-3-4.log-payloads.json @@ -33,7 +33,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -74,7 +74,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -115,7 +115,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -176,7 +176,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -211,7 +211,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -246,7 +246,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -284,7 +284,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -325,7 +325,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -366,7 +366,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-3-4.span-events.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-3-4.span-events.json index 909d73766..09caba8c2 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-3-4.span-events.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1-3-4.span-events.json @@ -31,7 +31,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -67,7 +67,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -103,7 +103,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -124,7 +124,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -181,7 +181,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -217,7 +217,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -250,7 +250,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -286,7 +286,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -322,7 +322,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1.log-payloads.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1.log-payloads.json index c435f5788..e02cea9c9 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1.log-payloads.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1.log-payloads.json @@ -33,7 +33,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -74,7 +74,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -115,7 +115,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -176,7 +176,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -211,7 +211,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -246,7 +246,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -287,7 +287,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -328,7 +328,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -369,7 +369,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1.span-events.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1.span-events.json index b7d2919bc..ab1e0df51 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1.span-events.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v1.span-events.json @@ -31,7 +31,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -67,7 +67,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -103,7 +103,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -124,7 +124,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -145,7 +145,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -181,7 +181,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -217,7 +217,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -253,7 +253,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -289,7 +289,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -325,7 +325,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v2.log-payloads.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v2.log-payloads.json index 85b50c53d..dec74bdb6 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v2.log-payloads.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v2.log-payloads.json @@ -33,7 +33,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -75,7 +75,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -117,7 +117,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -148,7 +148,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -180,7 +180,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -216,7 +216,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -252,7 +252,7 @@ "has_output": true, "input": "", "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -294,7 +294,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -336,7 +336,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -378,7 +378,7 @@ "type": "messages" }, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ diff --git a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v2.span-events.json b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v2.span-events.json index 6227c9374..1f09eecd5 100644 --- a/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v2.span-events.json +++ b/e2e/scenarios/mistral-instrumentation/__snapshots__/mistral-v2.span-events.json @@ -31,7 +31,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -68,7 +68,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -105,7 +105,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -127,7 +127,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -149,7 +149,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -186,7 +186,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -223,7 +223,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "codestral-latest", + "model": "codestral-2508", "provider": "mistral" }, "metric_keys": [ @@ -260,7 +260,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -297,7 +297,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ @@ -334,7 +334,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "mistral-small-latest", + "model": "mistral-small-2506", "provider": "mistral" }, "metric_keys": [ From 836820e08ccd86f23668998caa6f09558e807115 Mon Sep 17 00:00:00 2001 From: Luca Forstner Date: Wed, 8 Apr 2026 15:43:56 +0200 Subject: [PATCH 3/3] snapshots --- .../__snapshots__/openai-v6.log-payloads.json | 2 +- .../__snapshots__/openai-v6.span-events.json | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.log-payloads.json b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.log-payloads.json index f92a4f320..982b5ae25 100644 --- a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.log-payloads.json +++ b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.log-payloads.json @@ -554,7 +554,7 @@ } ], "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": { diff --git a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.span-events.json b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.span-events.json index 9086c4654..dcbd1c10a 100644 --- a/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.span-events.json +++ b/e2e/scenarios/openai-instrumentation/__snapshots__/openai-v6.span-events.json @@ -378,7 +378,7 @@ "has_input": true, "has_output": true, "metadata": { - "model": "gpt-4o-mini", + "model": "gpt-4o-mini-2024-07-18", "provider": "openai" }, "metrics": {