fix(openai): preserve .withResponse() on create() return value #42601
Annotations
10 errors and 1 warning
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "7b4491d08e51d47f",
+ "span_id": "b2b820079dfde358",
+ "start_timestamp": 1769681190.336,
"status": "ok",
+ "timestamp": 1769681190.4185789,
+ "trace_id": "e22eb42eb39469bf1d2b54991115817f",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 43985,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43985,
+ "url": "http://localhost:43985/openai/chat/completions",
+ "url.full": "http://localhost:43985/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:43985/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b2b820079dfde358",
+ "span_id": "42e1bc908d52188a",
+ "start_timestamp": 1769681190.362,
+ "status": "ok",
+ "timestamp": 1769681190.4135578,
+ "trace_id": "e22eb42eb39469bf1d2b54991115817f",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "7b4491d08e51d47f",
+ "span_id": "18af34bcc5d1ca13",
+ "start_timestamp": 1769681190.419,
"status": "ok",
+ "timestamp": 1769681190.4398682,
+ "trace_id": "e22eb42eb39469bf1d2b54991115817f",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 43985,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43985,
+ "url": "http://localhost:43985/openai/responses",
+ "url.full": "http://localhost:43985/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:43985/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "18af34bcc5d1ca13",
+ "span_id": "d46df02ea4ca1457",
+ "start_timestamp": 1769681190.424,
+ "status": "ok",
+ "timestamp": 1769681190.4386458,
+ "trace_id": "e22eb42eb39469bf1d2b54991115817f",
+ },
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "674ab432eea07ce7",
+ "span_id": "8782131b6bbe3dda",
+ "start_timestamp": 1769681186.817,
"status": "ok",
+ "timestamp": 1769681186.8942358,
+ "trace_id": "3b9b6964038939ad2bba382761a2c693",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 34529,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34529,
+ "url": "http://localhost:34529/openai/chat/completions",
+ "url.full": "http://localhost:34529/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:34529/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "8782131b6bbe3dda",
+ "span_id": "83b7aaafeb6c6d88",
+ "start_timestamp": 1769681186.845,
+ "status": "ok",
+ "timestamp": 1769681186.8894174,
+ "trace_id": "3b9b6964038939ad2bba382761a2c693",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "674ab432eea07ce7",
+ "span_id": "a4a0587afadcd531",
+ "start_timestamp": 1769681186.895,
"status": "ok",
+ "timestamp": 1769681186.9158792,
+ "trace_id": "3b9b6964038939ad2bba382761a2c693",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 34529,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34529,
+ "url": "http://localhost:34529/openai/responses",
+ "url.full": "http://localhost:34529/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:34529/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "a4a0587afadcd531",
+ "span_id": "934087bb1e409874",
+ "start_timestamp": 1769681186.899,
+ "status": "ok",
+ "timestamp": 1769681186.914312,
+ "trace_id": "3b9b6964038939ad2bba382761a2c693",
+ },
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "be11aaf387723fce",
+ "span_id": "4f83c6c3e90e9f0c",
+ "start_timestamp": 1769681182.843,
"status": "ok",
+ "timestamp": 1769681182.933857,
+ "trace_id": "01eac071dad4f8b829bf1fe55c700b5e",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40831,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40831,
+ "url": "http://localhost:40831/openai/embeddings",
+ "url.full": "http://localhost:40831/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40831/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "4f83c6c3e90e9f0c",
+ "span_id": "eb349ede6dd07b8c",
+ "start_timestamp": 1769681182.876,
+ "status": "ok",
+ "timestamp": 1769681182.9290695,
+ "trace_id": "01eac071dad4f8b829bf1fe55c700b5e",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "be11aaf387723fce",
+ "span_id": "d0012ddad78a52f5",
+ "start_timestamp": 1769681182.934,
+ "status": "ok",
+ "timestamp": 1769681182.9588113,
+ "trace_id": "01eac071dad4f8b829bf1fe55c700b5e",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 40831,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40831,
+ "url": "http://localhost:40831/openai/embeddings",
+ "url.full": "http://localhost:40831/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:40831/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "d0012ddad78a52f5",
+ "span_id": "0c80d4147fdbeefa",
+ "start_timestamp": 1769681182.936,
+ "status": "not_found",
+ "timestamp": 1769681182.951261,
+ "trace
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "015936a281be3605",
+ "span_id": "51fb60a8b7eb7711",
+ "start_timestamp": 1769681182.013,
"status": "ok",
+ "timestamp": 1769681182.1004598,
+ "trace_id": "d178eea3e1623b63370e120e395181de",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 33541,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 33541,
+ "url": "http://localhost:33541/openai/embeddings",
+ "url.full": "http://localhost:33541/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:33541/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "51fb60a8b7eb7711",
+ "span_id": "a1cbdf0084953815",
+ "start_timestamp": 1769681182.041,
+ "status": "ok",
+ "timestamp": 1769681182.095496,
+ "trace_id": "d178eea3e1623b63370e120e395181de",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "015936a281be3605",
+ "span_id": "e891df2da8e0403c",
+ "start_timestamp": 1769681182.101,
+ "status": "ok",
+ "timestamp": 1769681182.1278675,
+ "trace_id": "d178eea3e1623b63370e120e395181de",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 33541,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 33541,
+ "url": "http://localhost:33541/openai/embeddings",
+ "url.full": "http://localhost:33541/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:33541/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e891df2da8e0403c",
+ "span_id": "6b41d4e5e1ca6acb",
+ "start_timestamp": 1769681182.104,
+ "status": "not_found",
+ "timestamp": 1769681182.118539,
+ "trace
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "1dcbb38ed94e4065",
+ "span_id": "c1ea60ece76ba6b6",
+ "start_timestamp": 1769681180.811,
"status": "ok",
+ "timestamp": 1769681180.932901,
+ "trace_id": "61a9734a9cfdc04834439b925cc727e8",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 35369,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 35369,
+ "url": "http://localhost:35369/openai/embeddings",
+ "url.full": "http://localhost:35369/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:35369/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c1ea60ece76ba6b6",
+ "span_id": "00e4f3980b1eb8e6",
+ "start_timestamp": 1769681180.85,
+ "status": "ok",
+ "timestamp": 1769681180.9247015,
+ "trace_id": "61a9734a9cfdc04834439b925cc727e8",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "1dcbb38ed94e4065",
+ "span_id": "ab8730d3eabe029a",
+ "start_timestamp": 1769681180.934,
+ "status": "ok",
+ "timestamp": 1769681180.9766138,
+ "trace_id": "61a9734a9cfdc04834439b925cc727e8",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 35369,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 35369,
+ "url": "http://localhost:35369/openai/embeddings",
+ "url.full": "http://localhost:35369/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:35369/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "ab8730d3eabe029a",
+ "span_id": "4c05756e85990275",
+ "start_timestamp": 1769681180.936,
+ "status": "not_found",
+ "timestamp": 1769681180.9644485,
+ "trace_id": "61a9734a9cfdc04834439b925cc727e8",
+ },
+ {
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "d5bf444183d29d8f",
+ "span_id": "54260d9b312bb3af",
+ "start_timestamp": 1769681180.058,
"status": "ok",
+ "timestamp": 1769681180.1422598,
+ "trace_id": "35388ec6547d13c5fbcd453514ea9f76",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 37477,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37477,
+ "url": "http://localhost:37477/openai/embeddings",
+ "url.full": "http://localhost:37477/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:37477/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "54260d9b312bb3af",
+ "span_id": "3d73d392cb85d03f",
+ "start_timestamp": 1769681180.081,
+ "status": "ok",
+ "timestamp": 1769681180.138156,
+ "trace_id": "35388ec6547d13c5fbcd453514ea9f76",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "d5bf444183d29d8f",
+ "span_id": "c5f16beceaef3eae",
+ "start_timestamp": 1769681180.143,
+ "status": "ok",
+ "timestamp": 1769681180.1674297,
+ "trace_id": "35388ec6547d13c5fbcd453514ea9f76",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 37477,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 37477,
+ "url": "http://localhost:37477/openai/embeddings",
+ "url.full": "http://localhost:37477/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:37477/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c5f16beceaef3eae",
+ "span_id": "988f8cbc2286b284",
+ "start_timestamp": 1769681180.148,
+ "status": "not_found",
+ "timestamp": 1769681180.1604052,
+ "trace_id": "35388ec6547d13c5fbcd453514ea9f76",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "eb3c8a5afdddb9ff",
+ "span_id": "e9edc06abe071ab3",
+ "start_timestamp": 1769681168.863,
"status": "ok",
+ "timestamp": 1769681168.9658208,
+ "trace_id": "5e70ee9174f9f9aef8f84da4bc408660",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 39075,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 39075,
+ "url": "http://localhost:39075/openai/chat/completions",
+ "url.full": "http://localhost:39075/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:39075/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e9edc06abe071ab3",
+ "span_id": "84afcf99a86632de",
+ "start_timestamp": 1769681168.899,
+ "status": "ok",
+ "timestamp": 1769681168.9606924,
+ "trace_id": "5e70ee9174f9f9aef8f84da4bc408660",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "eb3c8a5afdddb9ff",
+ "span_id": "daca381f4d826fa2",
+ "start_timestamp": 1769681168.966,
"status": "ok",
+ "timestamp": 1769681168.992986,
+ "trace_id": "5e70ee9174f9f9aef8f84da4bc408660",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 39075,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 39075,
+ "url": "http://localhost:39075/openai/responses",
+ "url.full": "http://localhost:39075/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:39075/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "daca381f4d826fa2",
+ "span_id": "faad5791c732db1d",
+ "start_timestamp": 1769681168.969,
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "30b8f04fc26aba7f",
+ "span_id": "13579bb427aae0bf",
+ "start_timestamp": 1769681165.311,
"status": "ok",
+ "timestamp": 1769681165.4033792,
+ "trace_id": "89cba4f694da89be75652ed0b73fe6ad",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 36721,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 36721,
+ "url": "http://localhost:36721/openai/chat/completions",
+ "url.full": "http://localhost:36721/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:36721/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "13579bb427aae0bf",
+ "span_id": "10dbb40acc487d59",
+ "start_timestamp": 1769681165.344,
+ "status": "ok",
+ "timestamp": 1769681165.3985572,
+ "trace_id": "89cba4f694da89be75652ed0b73fe6ad",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "30b8f04fc26aba7f",
+ "span_id": "61fc5a764507704a",
+ "start_timestamp": 1769681165.404,
"status": "ok",
+ "timestamp": 1769681165.431266,
+ "trace_id": "89cba4f694da89be75652ed0b73fe6ad",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 36721,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 36721,
+ "url": "http://localhost:36721/openai/responses",
+ "url.full": "http://localhost:36721/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:36721/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "61fc5a764507704a",
+ "span_id": "13dd4668e9135ea3",
+ "start_timestamp": 1769681165.406,
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "86f6553b5a4d51ea",
+ "span_id": "b8d22861c043f3fa",
+ "start_timestamp": 1769681161.385,
"status": "ok",
+ "timestamp": 1769681161.4653332,
+ "trace_id": "7bdde14eb9895930f92a1e76329690e4",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 46453,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46453,
+ "url": "http://localhost:46453/openai/chat/completions",
+ "url.full": "http://localhost:46453/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:46453/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b8d22861c043f3fa",
+ "span_id": "a3c6c58a2baaa8b9",
+ "start_timestamp": 1769681161.411,
+ "status": "ok",
+ "timestamp": 1769681161.4587538,
+ "trace_id": "7bdde14eb9895930f92a1e76329690e4",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "86f6553b5a4d51ea",
+ "span_id": "44fe05c3c73c2d7c",
+ "start_timestamp": 1769681161.466,
"status": "ok",
+ "timestamp": 1769681161.4830937,
+ "trace_id": "7bdde14eb9895930f92a1e76329690e4",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 46453,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 46453,
+ "url": "http://localhost:46453/openai/responses",
+ "url.full": "http://localhost:46453/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:46453/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "44fe05c3c73c2d7c",
+ "span_id": "d67e2be18ef13ea0",
+ "start_timestamp": 1769681161.468,
+ "status": "ok",
+ "timestamp": 1769681161.481275,
+ "trace_id": "7bdde14eb9895930f92a1e76329690e4",
+ }
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "f345d50131c24f98",
+ "span_id": "b236be782862cdcb",
+ "start_timestamp": 1769681157.845,
"status": "ok",
+ "timestamp": 1769681157.9244041,
+ "trace_id": "60140d45a6c419ee55e2cb26adb9eebd",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 34345,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34345,
+ "url": "http://localhost:34345/openai/chat/completions",
+ "url.full": "http://localhost:34345/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:34345/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b236be782862cdcb",
+ "span_id": "520ea8e9c729f3b9",
+ "start_timestamp": 1769681157.873,
+ "status": "ok",
+ "timestamp": 1769681157.9192796,
+ "trace_id": "60140d45a6c419ee55e2cb26adb9eebd",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "f345d50131c24f98",
+ "span_id": "d2947f253e14fffa",
+ "start_timestamp": 1769681157.925,
"status": "ok",
+ "timestamp": 1769681157.955169,
+ "trace_id": "60140d45a6c419ee55e2cb26adb9eebd",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 34345,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 34345,
+ "url": "http://localhost:34345/openai/responses",
+ "url.full": "http://localhost:34345/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:34345/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "d2947f253e14fffa",
+ "span_id": "0fc07fd5283d38d9",
+ "start_timestamp": 1769681157.928,
+ "status": "ok",
+ "timestamp": 1769681157.95423,
+ "trace_id": "60140d45a6c419ee55e2cb26adb9eebd",
+ },
|
|
|
The logs for this run have expired and are no longer available.
Loading