fix(openai): preserve .withResponse() on create() return value #42601
Annotations
10 errors and 1 warning
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "56aad438ff679be5",
+ "span_id": "df47fbd24487c04f",
+ "start_timestamp": 1769681200.319,
"status": "ok",
+ "timestamp": 1769681200.3892071,
+ "trace_id": "d390fef45a3b58ed001071136efe4869",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 39605,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 39605,
+ "url": "http://localhost:39605/openai/chat/completions",
+ "url.full": "http://localhost:39605/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:39605/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "df47fbd24487c04f",
+ "span_id": "417f7b8e1bdcd2d3",
+ "start_timestamp": 1769681200.343,
+ "status": "ok",
+ "timestamp": 1769681200.3861444,
+ "trace_id": "d390fef45a3b58ed001071136efe4869",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "56aad438ff679be5",
+ "span_id": "c8e90692ecc1e343",
+ "start_timestamp": 1769681200.389,
"status": "ok",
+ "timestamp": 1769681200.4013104,
+ "trace_id": "d390fef45a3b58ed001071136efe4869",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 39605,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 39605,
+ "url": "http://localhost:39605/openai/responses",
+ "url.full": "http://localhost:39605/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:39605/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c8e90692ecc1e343",
+ "span_id": "c81880cc4e07dd99",
+ "start_timestamp": 1769681200.391,
+ "status": "ok",
+ "timestamp": 1769681200.4008133,
+ "trace_id": "d390fef45a3b58ed001071136efe4869",
+ },
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "fa808fed4323332e",
+ "span_id": "d21c7306a69e9957",
+ "start_timestamp": 1769681196.826,
"status": "ok",
+ "timestamp": 1769681196.8991163,
+ "trace_id": "468e5bfe8b06fb1db00fb2fb178a9459",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 32799,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 32799,
+ "url": "http://localhost:32799/openai/chat/completions",
+ "url.full": "http://localhost:32799/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:32799/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "d21c7306a69e9957",
+ "span_id": "d51e6db6a62a7675",
+ "start_timestamp": 1769681196.846,
+ "status": "ok",
+ "timestamp": 1769681196.8949687,
+ "trace_id": "468e5bfe8b06fb1db00fb2fb178a9459",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "fa808fed4323332e",
+ "span_id": "eb7c4f5f13ba015e",
+ "start_timestamp": 1769681196.9,
"status": "ok",
+ "timestamp": 1769681196.9180207,
+ "trace_id": "468e5bfe8b06fb1db00fb2fb178a9459",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 32799,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 32799,
+ "url": "http://localhost:32799/openai/responses",
+ "url.full": "http://localhost:32799/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:32799/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "eb7c4f5f13ba015e",
+ "span_id": "50379292ab08a55c",
+ "start_timestamp": 1769681196.902,
+ "status": "ok",
+ "timestamp": 1769681196.9166057,
+ "trace_id": "468e5bfe8b06fb1db00fb2fb178a9459",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "d3bd31dba19242d1",
+ "span_id": "c3c93ff4672f0338",
+ "start_timestamp": 1769681192.247,
"status": "ok",
+ "timestamp": 1769681192.3494518,
+ "trace_id": "932c4dc2eab3f3b0f8e53f02e9c4d3f7",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 33759,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 33759,
+ "url": "http://localhost:33759/openai/embeddings",
+ "url.full": "http://localhost:33759/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:33759/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c3c93ff4672f0338",
+ "span_id": "a016a595051b49fd",
+ "start_timestamp": 1769681192.283,
+ "status": "ok",
+ "timestamp": 1769681192.345116,
+ "trace_id": "932c4dc2eab3f3b0f8e53f02e9c4d3f7",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "d3bd31dba19242d1",
+ "span_id": "a634c5739099f962",
+ "start_timestamp": 1769681192.35,
+ "status": "ok",
+ "timestamp": 1769681192.3685036,
+ "trace_id": "932c4dc2eab3f3b0f8e53f02e9c4d3f7",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 33759,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 33759,
+ "url": "http://localhost:33759/openai/embeddings",
+ "url.full": "http://localhost:33759/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:33759/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "a634c5739099f962",
+ "span_id": "1384bc8b788a785b",
+ "start_timestamp": 1769681192.352,
+ "status": "not_found",
+ "timestamp": 1769681192.361421,
+ "trace_
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "0e47ce157823d859",
+ "span_id": "e67b1e65bb165227",
+ "start_timestamp": 1769681191.422,
"status": "ok",
+ "timestamp": 1769681191.480162,
+ "trace_id": "65e38c2773281eddcf475fb395e12c53",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40883,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40883,
+ "url": "http://localhost:40883/openai/embeddings",
+ "url.full": "http://localhost:40883/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40883/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e67b1e65bb165227",
+ "span_id": "e70ea56aebecb3be",
+ "start_timestamp": 1769681191.434,
+ "status": "ok",
+ "timestamp": 1769681191.4770727,
+ "trace_id": "65e38c2773281eddcf475fb395e12c53",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "0e47ce157823d859",
+ "span_id": "7fc05978863664a2",
+ "start_timestamp": 1769681191.481,
+ "status": "ok",
+ "timestamp": 1769681191.5037403,
+ "trace_id": "65e38c2773281eddcf475fb395e12c53",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 40883,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40883,
+ "url": "http://localhost:40883/openai/embeddings",
+ "url.full": "http://localhost:40883/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:40883/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "7fc05978863664a2",
+ "span_id": "a79e436630a776ce",
+ "start_timestamp": 1769681191.483,
+ "status": "not_found",
+ "timestamp": 1769681191.4973078,
+ "trac
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "b34a9ded3129164d",
+ "span_id": "4436e1b35de85158",
+ "start_timestamp": 1769681190.382,
"status": "ok",
+ "timestamp": 1769681190.4612517,
+ "trace_id": "75423fee90c227a904c2dfc2ab742a7e",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37153,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37153,
+ "url": "http://localhost:37153/openai/embeddings",
+ "url.full": "http://localhost:37153/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:37153/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "4436e1b35de85158",
+ "span_id": "e49e34e768b104b0",
+ "start_timestamp": 1769681190.412,
+ "status": "ok",
+ "timestamp": 1769681190.4583976,
+ "trace_id": "75423fee90c227a904c2dfc2ab742a7e",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "b34a9ded3129164d",
+ "span_id": "49b84ca477f31de0",
+ "start_timestamp": 1769681190.462,
+ "status": "ok",
+ "timestamp": 1769681190.4785292,
+ "trace_id": "75423fee90c227a904c2dfc2ab742a7e",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 37153,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37153,
+ "url": "http://localhost:37153/openai/embeddings",
+ "url.full": "http://localhost:37153/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:37153/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "49b84ca477f31de0",
+ "span_id": "2a35935ac854c1da",
+ "start_timestamp": 1769681190.463,
+ "status": "not_found",
+ "timestamp": 1769681190.4731083,
+ "trace_id": "75423fee90c227a904c2dfc2ab742a7e",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "580b1b522e34c9dc",
+ "span_id": "b17484f5a7a4ce97",
+ "start_timestamp": 1769681189.639,
"status": "ok",
+ "timestamp": 1769681189.7078106,
+ "trace_id": "1f6e48d2437a8ae4694fc46ae9b5fb87",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 39863,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 39863,
+ "url": "http://localhost:39863/openai/embeddings",
+ "url.full": "http://localhost:39863/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:39863/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b17484f5a7a4ce97",
+ "span_id": "951c528aae6aa490",
+ "start_timestamp": 1769681189.653,
+ "status": "ok",
+ "timestamp": 1769681189.7038062,
+ "trace_id": "1f6e48d2437a8ae4694fc46ae9b5fb87",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "580b1b522e34c9dc",
+ "span_id": "ea4bce1bd0a8b8a4",
+ "start_timestamp": 1769681189.708,
+ "status": "ok",
+ "timestamp": 1769681189.7255929,
+ "trace_id": "1f6e48d2437a8ae4694fc46ae9b5fb87",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 39863,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 39863,
+ "url": "http://localhost:39863/openai/embeddings",
+ "url.full": "http://localhost:39863/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:39863/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "ea4bce1bd0a8b8a4",
+ "span_id": "4684778efb5a1e0a",
+ "start_timestamp": 1769681189.71,
+ "status": "not_found",
+ "timestamp": 1769681189.7204163,
+ "trace_id": "1f6e48d2437a8ae4694fc46ae9b5fb87",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "787cc9313a1c3264",
+ "span_id": "fab7ba9be57c87ec",
+ "start_timestamp": 1769681178.748,
"status": "ok",
+ "timestamp": 1769681178.8533788,
+ "trace_id": "aa40a3d025eb7b6bc997728cc09ca459",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 46007,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46007,
+ "url": "http://localhost:46007/openai/chat/completions",
+ "url.full": "http://localhost:46007/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:46007/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "fab7ba9be57c87ec",
+ "span_id": "a496ad4d11129e95",
+ "start_timestamp": 1769681178.784,
+ "status": "ok",
+ "timestamp": 1769681178.8486164,
+ "trace_id": "aa40a3d025eb7b6bc997728cc09ca459",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "787cc9313a1c3264",
+ "span_id": "508db6a1940e472f",
+ "start_timestamp": 1769681178.854,
"status": "ok",
+ "timestamp": 1769681178.882011,
+ "trace_id": "aa40a3d025eb7b6bc997728cc09ca459",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 46007,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46007,
+ "url": "http://localhost:46007/openai/responses",
+ "url.full": "http://localhost:46007/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:46007/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "508db6a1940e472f",
+ "span_id": "5bcd494874ca89d6",
+ "start_timestamp": 1769681178.855,
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "b46ddd7f8be6ab2b",
+ "span_id": "41b781cb9d22e1ab",
+ "start_timestamp": 1769681175.131,
"status": "ok",
+ "timestamp": 1769681175.2160854,
+ "trace_id": "ccdfc1e622c146cb3be5f5ab3db7f449",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 42439,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 42439,
+ "url": "http://localhost:42439/openai/chat/completions",
+ "url.full": "http://localhost:42439/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:42439/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "41b781cb9d22e1ab",
+ "span_id": "643f84ae6bfee02e",
+ "start_timestamp": 1769681175.144,
+ "status": "ok",
+ "timestamp": 1769681175.2093804,
+ "trace_id": "ccdfc1e622c146cb3be5f5ab3db7f449",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "b46ddd7f8be6ab2b",
+ "span_id": "b2f3730a4699065d",
+ "start_timestamp": 1769681175.216,
"status": "ok",
+ "timestamp": 1769681175.2487833,
+ "trace_id": "ccdfc1e622c146cb3be5f5ab3db7f449",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 42439,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 42439,
+ "url": "http://localhost:42439/openai/responses",
+ "url.full": "http://localhost:42439/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:42439/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b2f3730a4699065d",
+ "span_id": "d10cf413d40309d2",
+ "start_timestamp": 1769681175.221,
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "61f69fbf11795e4c",
+ "span_id": "2d478378b02dfd5e",
+ "start_timestamp": 1769681171.06,
"status": "ok",
+ "timestamp": 1769681171.1478574,
+ "trace_id": "3bce0296d13183f728cebf847654f71f",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45437,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45437,
+ "url": "http://localhost:45437/openai/chat/completions",
+ "url.full": "http://localhost:45437/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:45437/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "2d478378b02dfd5e",
+ "span_id": "b70d5eaaa5b1b99c",
+ "start_timestamp": 1769681171.088,
+ "status": "ok",
+ "timestamp": 1769681171.1433053,
+ "trace_id": "3bce0296d13183f728cebf847654f71f",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "61f69fbf11795e4c",
+ "span_id": "7fc21a34d7df177e",
+ "start_timestamp": 1769681171.148,
"status": "ok",
+ "timestamp": 1769681171.164311,
+ "trace_id": "3bce0296d13183f728cebf847654f71f",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45437,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45437,
+ "url": "http://localhost:45437/openai/responses",
+ "url.full": "http://localhost:45437/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:45437/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "7fc21a34d7df177e",
+ "span_id": "dc0cb7e4ea9a40ef",
+ "start_timestamp": 1769681171.15,
+ "status": "ok",
+ "timestamp": 1769681171.1629076,
+ "trace_id": "3bce0296d13183f728cebf847654f71f",
+ },
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "466a0fb2e4b7a2a1",
+ "span_id": "daa3c32896e403aa",
+ "start_timestamp": 1769681167.353,
"status": "ok",
+ "timestamp": 1769681167.4069235,
+ "trace_id": "0851b3552863818932bac8250f2a622b",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37355,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37355,
+ "url": "http://localhost:37355/openai/chat/completions",
+ "url.full": "http://localhost:37355/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:37355/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "daa3c32896e403aa",
+ "span_id": "abc0ee9de1c8c0e8",
+ "start_timestamp": 1769681167.363,
+ "status": "ok",
+ "timestamp": 1769681167.404077,
+ "trace_id": "0851b3552863818932bac8250f2a622b",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "466a0fb2e4b7a2a1",
+ "span_id": "8abc282570411aaa",
+ "start_timestamp": 1769681167.407,
"status": "ok",
+ "timestamp": 1769681167.4159865,
+ "trace_id": "0851b3552863818932bac8250f2a622b",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37355,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37355,
+ "url": "http://localhost:37355/openai/responses",
+ "url.full": "http://localhost:37355/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:37355/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "8abc282570411aaa",
+ "span_id": "c7f2ec04f2d60443",
+ "start_timestamp": 1769681167.409,
+ "status": "ok",
+ "timestamp": 1769681167.4155936,
+ "trace_id": "0851b3552863818932bac8250f2a622b",
+ }
|
|
|
The logs for this run have expired and are no longer available.
Loading