fix(openai): preserve .withResponse() on create() return value #42601
Annotations
10 errors and 1 warning
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "17fc84b6ad16a527",
+ "span_id": "ea1a4e8768d5c6eb",
+ "start_timestamp": 1769681200.712,
"status": "ok",
+ "timestamp": 1769681200.7903967,
+ "trace_id": "9f1ae84b4b9560f308afd5b8054fd88c",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40697,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40697,
+ "url": "http://localhost:40697/openai/chat/completions",
+ "url.full": "http://localhost:40697/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:40697/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "ea1a4e8768d5c6eb",
+ "span_id": "19a0b36ab35f5e9c",
+ "start_timestamp": 1769681200.741,
+ "status": "ok",
+ "timestamp": 1769681200.786222,
+ "trace_id": "9f1ae84b4b9560f308afd5b8054fd88c",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "17fc84b6ad16a527",
+ "span_id": "750ccba0e3eb580b",
+ "start_timestamp": 1769681200.791,
"status": "ok",
+ "timestamp": 1769681200.8039656,
+ "trace_id": "9f1ae84b4b9560f308afd5b8054fd88c",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40697,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40697,
+ "url": "http://localhost:40697/openai/responses",
+ "url.full": "http://localhost:40697/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:40697/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "750ccba0e3eb580b",
+ "span_id": "1ae420988563c11c",
+ "start_timestamp": 1769681200.792,
+ "status": "ok",
+ "timestamp": 1769681200.802272,
+ "trace_id": "9f1ae84b4b9560f308afd5b8054fd88c",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "6c74d30c6b48d85f",
+ "span_id": "fa0bb88ad47635e9",
+ "start_timestamp": 1769681197.186,
"status": "ok",
+ "timestamp": 1769681197.2433636,
+ "trace_id": "77b4e6e14b3f8eafc4e0d1f0212abb5d",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37371,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37371,
+ "url": "http://localhost:37371/openai/chat/completions",
+ "url.full": "http://localhost:37371/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:37371/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "fa0bb88ad47635e9",
+ "span_id": "12f00fc78e3f9d79",
+ "start_timestamp": 1769681197.195,
+ "status": "ok",
+ "timestamp": 1769681197.2391734,
+ "trace_id": "77b4e6e14b3f8eafc4e0d1f0212abb5d",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "6c74d30c6b48d85f",
+ "span_id": "b61ed84d440e304f",
+ "start_timestamp": 1769681197.244,
"status": "ok",
+ "timestamp": 1769681197.2585058,
+ "trace_id": "77b4e6e14b3f8eafc4e0d1f0212abb5d",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37371,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37371,
+ "url": "http://localhost:37371/openai/responses",
+ "url.full": "http://localhost:37371/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:37371/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b61ed84d440e304f",
+ "span_id": "a803c320b93b4ed5",
+ "start_timestamp": 1769681197.245,
+ "status": "ok",
+ "timestamp": 1769681197.256912,
+ "trace_id": "77b4e6e14b3f8eafc4e0d1f0212abb5d",
+ },
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "8aca7fd22ea98834",
+ "span_id": "c062ac7d3fe70599",
+ "start_timestamp": 1769681192.302,
"status": "ok",
+ "timestamp": 1769681192.395445,
+ "trace_id": "2326fd0e023bbc1f4bc956d54ec8640e",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 42709,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 42709,
+ "url": "http://localhost:42709/openai/embeddings",
+ "url.full": "http://localhost:42709/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:42709/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c062ac7d3fe70599",
+ "span_id": "de31f5af24c56e31",
+ "start_timestamp": 1769681192.338,
+ "status": "ok",
+ "timestamp": 1769681192.3924751,
+ "trace_id": "2326fd0e023bbc1f4bc956d54ec8640e",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "8aca7fd22ea98834",
+ "span_id": "bd24148a42e272af",
+ "start_timestamp": 1769681192.395,
+ "status": "ok",
+ "timestamp": 1769681192.4128492,
+ "trace_id": "2326fd0e023bbc1f4bc956d54ec8640e",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 42709,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 42709,
+ "url": "http://localhost:42709/openai/embeddings",
+ "url.full": "http://localhost:42709/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:42709/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "bd24148a42e272af",
+ "span_id": "c8e99740d727f4bc",
+ "start_timestamp": 1769681192.397,
+ "status": "not_found",
+ "timestamp": 1769681192.4063864,
+ "trac
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "09d4d0129be31d73",
+ "span_id": "e3e7474eaaa22834",
+ "start_timestamp": 1769681191.531,
"status": "ok",
+ "timestamp": 1769681191.5870159,
+ "trace_id": "6cdb75fd50e50b4f0ab9066db39afdd3",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 46257,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46257,
+ "url": "http://localhost:46257/openai/embeddings",
+ "url.full": "http://localhost:46257/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:46257/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e3e7474eaaa22834",
+ "span_id": "69a586875a543fb1",
+ "start_timestamp": 1769681191.544,
+ "status": "ok",
+ "timestamp": 1769681191.5843897,
+ "trace_id": "6cdb75fd50e50b4f0ab9066db39afdd3",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "09d4d0129be31d73",
+ "span_id": "cf4aa01f14c66e19",
+ "start_timestamp": 1769681191.587,
+ "status": "ok",
+ "timestamp": 1769681191.604793,
+ "trace_id": "6cdb75fd50e50b4f0ab9066db39afdd3",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 46257,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46257,
+ "url": "http://localhost:46257/openai/embeddings",
+ "url.full": "http://localhost:46257/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:46257/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "cf4aa01f14c66e19",
+ "span_id": "182d14d95c2bf91e",
+ "start_timestamp": 1769681191.589,
+ "status": "not_found",
+ "timestamp": 1769681191.598585,
+ "trace
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "40f78de7bce2bf6d",
+ "span_id": "e3191f56eeebcb75",
+ "start_timestamp": 1769681190.451,
"status": "ok",
+ "timestamp": 1769681190.5443287,
+ "trace_id": "bae65a8da19a1f4f8678f13a2adf4a1a",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37775,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37775,
+ "url": "http://localhost:37775/openai/embeddings",
+ "url.full": "http://localhost:37775/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:37775/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e3191f56eeebcb75",
+ "span_id": "9c32b651b537b925",
+ "start_timestamp": 1769681190.484,
+ "status": "ok",
+ "timestamp": 1769681190.5414324,
+ "trace_id": "bae65a8da19a1f4f8678f13a2adf4a1a",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "40f78de7bce2bf6d",
+ "span_id": "4e385529c1935b1e",
+ "start_timestamp": 1769681190.545,
+ "status": "ok",
+ "timestamp": 1769681190.5684304,
+ "trace_id": "bae65a8da19a1f4f8678f13a2adf4a1a",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 37775,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37775,
+ "url": "http://localhost:37775/openai/embeddings",
+ "url.full": "http://localhost:37775/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:37775/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "4e385529c1935b1e",
+ "span_id": "8ee835375028f4df",
+ "start_timestamp": 1769681190.546,
+ "status": "not_found",
+ "timestamp": 1769681190.559056,
+ "trace_id": "bae65a8da19a1f4f8678f13a2adf4a1a",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "21be0b2093402b9a",
+ "span_id": "903e29fa245fcae9",
+ "start_timestamp": 1769681189.693,
"status": "ok",
+ "timestamp": 1769681189.7594006,
+ "trace_id": "133681439766662e64fcc5c1e3bbe020",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 43217,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43217,
+ "url": "http://localhost:43217/openai/embeddings",
+ "url.full": "http://localhost:43217/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:43217/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "903e29fa245fcae9",
+ "span_id": "71e215417d1eee4e",
+ "start_timestamp": 1769681189.707,
+ "status": "ok",
+ "timestamp": 1769681189.7552938,
+ "trace_id": "133681439766662e64fcc5c1e3bbe020",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "21be0b2093402b9a",
+ "span_id": "549ad53342fedcc1",
+ "start_timestamp": 1769681189.759,
+ "status": "ok",
+ "timestamp": 1769681189.775845,
+ "trace_id": "133681439766662e64fcc5c1e3bbe020",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 43217,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43217,
+ "url": "http://localhost:43217/openai/embeddings",
+ "url.full": "http://localhost:43217/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:43217/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "549ad53342fedcc1",
+ "span_id": "4eec80c1dcddf27b",
+ "start_timestamp": 1769681189.761,
+ "status": "not_found",
+ "timestamp": 1769681189.7710345,
+ "trace_id": "133681439766662e64fcc5c1e3bbe020",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "1a9331532795a56a",
+ "span_id": "dcec858d12c8c6a9",
+ "start_timestamp": 1769681178.712,
"status": "ok",
+ "timestamp": 1769681178.8195405,
+ "trace_id": "cd0f5fe012b17df03714f7d1e9843198",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 44503,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 44503,
+ "url": "http://localhost:44503/openai/chat/completions",
+ "url.full": "http://localhost:44503/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:44503/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "dcec858d12c8c6a9",
+ "span_id": "cd8e421dc1b2b65b",
+ "start_timestamp": 1769681178.748,
+ "status": "ok",
+ "timestamp": 1769681178.8155932,
+ "trace_id": "cd0f5fe012b17df03714f7d1e9843198",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "1a9331532795a56a",
+ "span_id": "06e9e67d05fd38e6",
+ "start_timestamp": 1769681178.82,
"status": "ok",
+ "timestamp": 1769681178.835089,
+ "trace_id": "cd0f5fe012b17df03714f7d1e9843198",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 44503,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 44503,
+ "url": "http://localhost:44503/openai/responses",
+ "url.full": "http://localhost:44503/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:44503/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "06e9e67d05fd38e6",
+ "span_id": "6dcbe7765f8746b5",
+ "start_timestamp": 1769681178.822,
+ "
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "51864b5d0ac99fa9",
+ "span_id": "b4f0253f286d7a96",
+ "start_timestamp": 1769681175.214,
"status": "ok",
+ "timestamp": 1769681175.2905145,
+ "trace_id": "c934ff2480fd46868b696d8545d3ce9c",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45347,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45347,
+ "url": "http://localhost:45347/openai/chat/completions",
+ "url.full": "http://localhost:45347/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:45347/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b4f0253f286d7a96",
+ "span_id": "2b89e6a7f8682151",
+ "start_timestamp": 1769681175.229,
+ "status": "ok",
+ "timestamp": 1769681175.287197,
+ "trace_id": "c934ff2480fd46868b696d8545d3ce9c",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "51864b5d0ac99fa9",
+ "span_id": "4c1a4fc112c50831",
+ "start_timestamp": 1769681175.291,
"status": "ok",
+ "timestamp": 1769681175.307992,
+ "trace_id": "c934ff2480fd46868b696d8545d3ce9c",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45347,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45347,
+ "url": "http://localhost:45347/openai/responses",
+ "url.full": "http://localhost:45347/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:45347/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "4c1a4fc112c50831",
+ "span_id": "3000dd884a1b2a04",
+ "start_timestamp": 1769681175.293,
+ "
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "1fd002529fb0b994",
+ "span_id": "37ed4b16dca16d51",
+ "start_timestamp": 1769681171.273,
"status": "ok",
+ "timestamp": 1769681171.369784,
+ "trace_id": "3058df605e1871e30e481ec0fddb415b",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 42971,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 42971,
+ "url": "http://localhost:42971/openai/chat/completions",
+ "url.full": "http://localhost:42971/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:42971/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "37ed4b16dca16d51",
+ "span_id": "e162c5b5cd5eab36",
+ "start_timestamp": 1769681171.305,
+ "status": "ok",
+ "timestamp": 1769681171.3668644,
+ "trace_id": "3058df605e1871e30e481ec0fddb415b",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "1fd002529fb0b994",
+ "span_id": "0d3488cac0d0830b",
+ "start_timestamp": 1769681171.37,
"status": "ok",
+ "timestamp": 1769681171.3868203,
+ "trace_id": "3058df605e1871e30e481ec0fddb415b",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 42971,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 42971,
+ "url": "http://localhost:42971/openai/responses",
+ "url.full": "http://localhost:42971/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:42971/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "0d3488cac0d0830b",
+ "span_id": "110fab53c6a80202",
+ "start_timestamp": 1769681171.372,
+ "status": "ok",
+ "timestamp": 1769681171.386164,
+ "trace_id": "3058df605e1871e30e481ec0fddb415b",
+ },
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "79bc50cdbde19e18",
+ "span_id": "60be6b88157cc999",
+ "start_timestamp": 1769681167.618,
"status": "ok",
+ "timestamp": 1769681167.669012,
+ "trace_id": "3ebee3f5ff4af2f8c0a41c0265cc9eb4",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40135,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40135,
+ "url": "http://localhost:40135/openai/chat/completions",
+ "url.full": "http://localhost:40135/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40135/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "60be6b88157cc999",
+ "span_id": "b1f0a3a177e8db58",
+ "start_timestamp": 1769681167.627,
+ "status": "ok",
+ "timestamp": 1769681167.6657922,
+ "trace_id": "3ebee3f5ff4af2f8c0a41c0265cc9eb4",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "79bc50cdbde19e18",
+ "span_id": "c7d33fd1b2b15145",
+ "start_timestamp": 1769681167.669,
"status": "ok",
+ "timestamp": 1769681167.6804419,
+ "trace_id": "3ebee3f5ff4af2f8c0a41c0265cc9eb4",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40135,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40135,
+ "url": "http://localhost:40135/openai/responses",
+ "url.full": "http://localhost:40135/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40135/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c7d33fd1b2b15145",
+ "span_id": "be26cf9e2bc6ade2",
+ "start_timestamp": 1769681167.67,
+ "status": "ok",
+ "timestamp": 1769681167.6793463,
+ "trace_id": "3ebee3f5ff4af2f8c0a41c0265cc9eb4",
+ },
|
|
|
The logs for this run have expired and are no longer available.
Loading