fix(openai): preserve .withResponse() on create() return value #42601
Annotations
10 errors and 1 warning
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "c3e7079cff5c30f5",
+ "span_id": "44d3a01175de7587",
+ "start_timestamp": 1769681199.424,
"status": "ok",
+ "timestamp": 1769681199.53712,
+ "trace_id": "46343ef57a10f98695eca5537bae2b28",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45583,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45583,
+ "url": "http://localhost:45583/openai/chat/completions",
+ "url.full": "http://localhost:45583/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:45583/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "44d3a01175de7587",
+ "span_id": "e7b055790d0b6217",
+ "start_timestamp": 1769681199.462,
+ "status": "ok",
+ "timestamp": 1769681199.5262444,
+ "trace_id": "46343ef57a10f98695eca5537bae2b28",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "c3e7079cff5c30f5",
+ "span_id": "73075c71b8e75e86",
+ "start_timestamp": 1769681199.537,
"status": "ok",
+ "timestamp": 1769681199.558887,
+ "trace_id": "46343ef57a10f98695eca5537bae2b28",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45583,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45583,
+ "url": "http://localhost:45583/openai/responses",
+ "url.full": "http://localhost:45583/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:45583/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "73075c71b8e75e86",
+ "span_id": "f8c7cb6c069b0d51",
+ "start_timestamp": 1769681199.54,
+ "status": "ok",
+ "timestamp": 1769681199.5562608,
+ "trace_id": "46343ef57a10f98695eca5537bae2b28",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "9904c7f757e71db8",
+ "span_id": "b0cae8bb27c77936",
+ "start_timestamp": 1769681195.564,
"status": "ok",
+ "timestamp": 1769681195.679575,
+ "trace_id": "9194f0be04ed2d854fc07db53c90deef",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37103,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37103,
+ "url": "http://localhost:37103/openai/chat/completions",
+ "url.full": "http://localhost:37103/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:37103/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b0cae8bb27c77936",
+ "span_id": "b5fdd1853b4e55d4",
+ "start_timestamp": 1769681195.606,
+ "status": "ok",
+ "timestamp": 1769681195.668236,
+ "trace_id": "9194f0be04ed2d854fc07db53c90deef",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "9904c7f757e71db8",
+ "span_id": "3273de9ef9ebc337",
+ "start_timestamp": 1769681195.68,
"status": "ok",
+ "timestamp": 1769681195.7046642,
+ "trace_id": "9194f0be04ed2d854fc07db53c90deef",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37103,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37103,
+ "url": "http://localhost:37103/openai/responses",
+ "url.full": "http://localhost:37103/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:37103/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "3273de9ef9ebc337",
+ "span_id": "23b8c34ff67c7691",
+ "start_timestamp": 1769681195.683,
+ "status": "ok",
+ "timestamp": 1769681195.7025626,
+ "trace_id": "9194f0be04ed2d854fc07db53c90deef",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "8cc22f5f981efec9",
+ "span_id": "bc42d74f18535344",
+ "start_timestamp": 1769681189.464,
"status": "ok",
+ "timestamp": 1769681189.557604,
+ "trace_id": "55363283386425ea2249f09e4408a2bc",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41239,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41239,
+ "url": "http://localhost:41239/openai/embeddings",
+ "url.full": "http://localhost:41239/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:41239/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "bc42d74f18535344",
+ "span_id": "c01149431d36a37d",
+ "start_timestamp": 1769681189.502,
+ "status": "ok",
+ "timestamp": 1769681189.5464246,
+ "trace_id": "55363283386425ea2249f09e4408a2bc",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "8cc22f5f981efec9",
+ "span_id": "6a4b098c5b5d9b04",
+ "start_timestamp": 1769681189.558,
+ "status": "ok",
+ "timestamp": 1769681189.5884566,
+ "trace_id": "55363283386425ea2249f09e4408a2bc",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 41239,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41239,
+ "url": "http://localhost:41239/openai/embeddings",
+ "url.full": "http://localhost:41239/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:41239/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "6a4b098c5b5d9b04",
+ "span_id": "16328fb18edaa586",
+ "start_timestamp": 1769681189.561,
+ "status": "not_found",
+ "timestamp": 1769681189.5793388,
+ "trac
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "4d77e8380c833886",
+ "span_id": "6fbc8a0bb29348da",
+ "start_timestamp": 1769681188.54,
"status": "ok",
+ "timestamp": 1769681188.650649,
+ "trace_id": "6ea87c5ea0f501852b23b5ca7c1037ef",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 34023,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34023,
+ "url": "http://localhost:34023/openai/embeddings",
+ "url.full": "http://localhost:34023/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:34023/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "6fbc8a0bb29348da",
+ "span_id": "b057a1dbed79356b",
+ "start_timestamp": 1769681188.579,
+ "status": "ok",
+ "timestamp": 1769681188.6391163,
+ "trace_id": "6ea87c5ea0f501852b23b5ca7c1037ef",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "4d77e8380c833886",
+ "span_id": "0a9906727801a36a",
+ "start_timestamp": 1769681188.651,
+ "status": "ok",
+ "timestamp": 1769681188.6745465,
+ "trace_id": "6ea87c5ea0f501852b23b5ca7c1037ef",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 34023,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34023,
+ "url": "http://localhost:34023/openai/embeddings",
+ "url.full": "http://localhost:34023/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:34023/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "0a9906727801a36a",
+ "span_id": "a7f7c320c0b69238",
+ "start_timestamp": 1769681188.654,
+ "status": "not_found",
+ "timestamp": 1769681188.6649377,
+ "trace
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "701b352866c2f6e5",
+ "span_id": "d66ff3e36660dc81",
+ "start_timestamp": 1769681187.312,
"status": "ok",
+ "timestamp": 1769681187.404201,
+ "trace_id": "4505baa739121e35eac84d2d8e5214bb",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 34203,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34203,
+ "url": "http://localhost:34203/openai/embeddings",
+ "url.full": "http://localhost:34203/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:34203/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "d66ff3e36660dc81",
+ "span_id": "59e2023d25a8a9a1",
+ "start_timestamp": 1769681187.344,
+ "status": "ok",
+ "timestamp": 1769681187.3906586,
+ "trace_id": "4505baa739121e35eac84d2d8e5214bb",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "701b352866c2f6e5",
+ "span_id": "165e6774630dd2bc",
+ "start_timestamp": 1769681187.404,
+ "status": "ok",
+ "timestamp": 1769681187.4332256,
+ "trace_id": "4505baa739121e35eac84d2d8e5214bb",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 34203,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34203,
+ "url": "http://localhost:34203/openai/embeddings",
+ "url.full": "http://localhost:34203/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:34203/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "165e6774630dd2bc",
+ "span_id": "6e846b022c66b000",
+ "start_timestamp": 1769681187.407,
+ "status": "not_found",
+ "timestamp": 1769681187.424001,
+ "trace_id": "4505baa739121e35eac84d2d8e5214bb",
+ },
+ {
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "23588b173e0216d1",
+ "span_id": "f259befdbfd3319a",
+ "start_timestamp": 1769681186.357,
"status": "ok",
+ "timestamp": 1769681186.451843,
+ "trace_id": "09cb6c29bc9f411689e2ae79a6dce527",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41459,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41459,
+ "url": "http://localhost:41459/openai/embeddings",
+ "url.full": "http://localhost:41459/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:41459/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "f259befdbfd3319a",
+ "span_id": "c26a6a754d2c7665",
+ "start_timestamp": 1769681186.398,
+ "status": "ok",
+ "timestamp": 1769681186.445073,
+ "trace_id": "09cb6c29bc9f411689e2ae79a6dce527",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "23588b173e0216d1",
+ "span_id": "cb82d245b66585b2",
+ "start_timestamp": 1769681186.452,
+ "status": "ok",
+ "timestamp": 1769681186.4720857,
+ "trace_id": "09cb6c29bc9f411689e2ae79a6dce527",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 41459,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41459,
+ "url": "http://localhost:41459/openai/embeddings",
+ "url.full": "http://localhost:41459/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:41459/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "cb82d245b66585b2",
+ "span_id": "60f45971b3baa838",
+ "start_timestamp": 1769681186.454,
+ "status": "not_found",
+ "timestamp": 1769681186.4659886,
+ "trace_id": "09cb6c29bc9f411689e2ae79a6dce527",
+ },
+ {
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "049a3003b3521ae1",
+ "span_id": "835e9cca7d17b27d",
+ "start_timestamp": 1769681175.18,
"status": "ok",
+ "timestamp": 1769681175.2980664,
+ "trace_id": "a197f22c17f7ce0af823b69c7b9fd728",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40973,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40973,
+ "url": "http://localhost:40973/openai/chat/completions",
+ "url.full": "http://localhost:40973/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40973/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "835e9cca7d17b27d",
+ "span_id": "34b067ff60229fe0",
+ "start_timestamp": 1769681175.231,
+ "status": "ok",
+ "timestamp": 1769681175.2874978,
+ "trace_id": "a197f22c17f7ce0af823b69c7b9fd728",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "049a3003b3521ae1",
+ "span_id": "bf26dc1cf9aa4e86",
+ "start_timestamp": 1769681175.298,
"status": "ok",
+ "timestamp": 1769681175.320671,
+ "trace_id": "a197f22c17f7ce0af823b69c7b9fd728",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40973,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40973,
+ "url": "http://localhost:40973/openai/responses",
+ "url.full": "http://localhost:40973/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40973/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "bf26dc1cf9aa4e86",
+ "span_id": "aa7cc84ef1462d83",
+ "start_timestamp": 1769681175.301,
+ "
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "9874531f3a75b0e2",
+ "span_id": "0d6b13f27b7d4f0a",
+ "start_timestamp": 1769681171.478,
"status": "ok",
+ "timestamp": 1769681171.5865169,
+ "trace_id": "7ac1a9b16d82525a33245e4819cfbe8e",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38037,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38037,
+ "url": "http://localhost:38037/openai/chat/completions",
+ "url.full": "http://localhost:38037/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38037/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "0d6b13f27b7d4f0a",
+ "span_id": "a1b050801e4365e6",
+ "start_timestamp": 1769681171.518,
+ "status": "ok",
+ "timestamp": 1769681171.5756862,
+ "trace_id": "7ac1a9b16d82525a33245e4819cfbe8e",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "9874531f3a75b0e2",
+ "span_id": "e5988c8afe1e5154",
+ "start_timestamp": 1769681171.587,
"status": "ok",
+ "timestamp": 1769681171.6074572,
+ "trace_id": "7ac1a9b16d82525a33245e4819cfbe8e",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38037,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38037,
+ "url": "http://localhost:38037/openai/responses",
+ "url.full": "http://localhost:38037/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38037/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e5988c8afe1e5154",
+ "span_id": "d4272eeb4db568e8",
+ "start_timestamp": 1769681171.589,
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "dca9ae87fc242023",
+ "span_id": "b12499ded9a91c7f",
+ "start_timestamp": 1769681167.303,
"status": "ok",
+ "timestamp": 1769681167.3846638,
+ "trace_id": "df24d1ff79c996b565cfd7867dcd0ac3",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41857,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41857,
+ "url": "http://localhost:41857/openai/chat/completions",
+ "url.full": "http://localhost:41857/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:41857/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b12499ded9a91c7f",
+ "span_id": "69721c910bd6b8d9",
+ "start_timestamp": 1769681167.33,
+ "status": "ok",
+ "timestamp": 1769681167.3744564,
+ "trace_id": "df24d1ff79c996b565cfd7867dcd0ac3",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "dca9ae87fc242023",
+ "span_id": "c8a65a5261e1f241",
+ "start_timestamp": 1769681167.385,
"status": "ok",
+ "timestamp": 1769681167.405051,
+ "trace_id": "df24d1ff79c996b565cfd7867dcd0ac3",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41857,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41857,
+ "url": "http://localhost:41857/openai/responses",
+ "url.full": "http://localhost:41857/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:41857/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c8a65a5261e1f241",
+ "span_id": "abd2f2dfe59bf7b9",
+ "start_timestamp": 1769681167.388,
+ "status": "ok",
+ "timestamp": 1769681167.4030743,
+ "trace_id": "df24d1ff79c996b565cfd7867dcd0ac3",
+ },
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "0eeab71d3a93c90d",
+ "span_id": "86dfdda5ab2388b3",
+ "start_timestamp": 1769681163.72,
"status": "ok",
+ "timestamp": 1769681163.8230975,
+ "trace_id": "a66ecb242e72e16521ca6153bda4d164",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40915,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40915,
+ "url": "http://localhost:40915/openai/chat/completions",
+ "url.full": "http://localhost:40915/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40915/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "86dfdda5ab2388b3",
+ "span_id": "bfcb0e89bae84f23",
+ "start_timestamp": 1769681163.759,
+ "status": "ok",
+ "timestamp": 1769681163.812684,
+ "trace_id": "a66ecb242e72e16521ca6153bda4d164",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "0eeab71d3a93c90d",
+ "span_id": "eb19a19ac319edfc",
+ "start_timestamp": 1769681163.823,
"status": "ok",
+ "timestamp": 1769681163.8447652,
+ "trace_id": "a66ecb242e72e16521ca6153bda4d164",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40915,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40915,
+ "url": "http://localhost:40915/openai/responses",
+ "url.full": "http://localhost:40915/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40915/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "eb19a19ac319edfc",
+ "span_id": "b96776d9ad262f5d",
+ "start_timestamp": 1769681163.826,
+ "status": "ok",
+ "timestamp": 1769681163.8424857,
+ "trace_id": "a66ecb242e72e16521ca6153bda4d164",
+ },
|
|
|
The logs for this run have expired and are no longer available.
Loading