fix(openai): preserve .withResponse() on create() return value #42601
Annotations
10 errors and 1 warning
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "f525ba392c80398d",
+ "span_id": "d9b453eecd46828c",
+ "start_timestamp": 1769681338.575,
"status": "ok",
+ "timestamp": 1769681338.6826,
+ "trace_id": "61d46b70b09939fba2de54e709830837",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41473,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41473,
+ "url": "http://localhost:41473/openai/chat/completions",
+ "url.full": "http://localhost:41473/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:41473/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "d9b453eecd46828c",
+ "span_id": "7e07c69709019a23",
+ "start_timestamp": 1769681338.615,
+ "status": "ok",
+ "timestamp": 1769681338.678044,
+ "trace_id": "61d46b70b09939fba2de54e709830837",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "f525ba392c80398d",
+ "span_id": "438f684f47918379",
+ "start_timestamp": 1769681338.683,
"status": "ok",
+ "timestamp": 1769681338.6999898,
+ "trace_id": "61d46b70b09939fba2de54e709830837",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41473,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41473,
+ "url": "http://localhost:41473/openai/responses",
+ "url.full": "http://localhost:41473/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:41473/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "438f684f47918379",
+ "span_id": "174b6351e4512ea6",
+ "start_timestamp": 1769681338.684,
+ "status": "ok",
+ "timestamp": 1769681338.698117,
+ "trace_id": "61d46b70b09939fba2de54e709830837",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "7166794bad5f064d",
+ "span_id": "7476ec42b309252e",
+ "start_timestamp": 1769681334.976,
"status": "ok",
+ "timestamp": 1769681335.0375686,
+ "trace_id": "50484f18acfa1ca35078a959a4bd9ab0",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 44447,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 44447,
+ "url": "http://localhost:44447/openai/chat/completions",
+ "url.full": "http://localhost:44447/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:44447/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "7476ec42b309252e",
+ "span_id": "9c7c078943cff964",
+ "start_timestamp": 1769681334.99,
+ "status": "ok",
+ "timestamp": 1769681335.0338728,
+ "trace_id": "50484f18acfa1ca35078a959a4bd9ab0",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "7166794bad5f064d",
+ "span_id": "e169c82320f2f8b1",
+ "start_timestamp": 1769681335.038,
"status": "ok",
+ "timestamp": 1769681335.0514297,
+ "trace_id": "50484f18acfa1ca35078a959a4bd9ab0",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 44447,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 44447,
+ "url": "http://localhost:44447/openai/responses",
+ "url.full": "http://localhost:44447/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 6.0.0",
+ },
+ "description": "POST http://localhost:44447/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "e169c82320f2f8b1",
+ "span_id": "62264555d3985d04",
+ "start_timestamp": 1769681335.04,
+ "status": "ok",
+ "timestamp": 1769681335.0507362,
+ "trace_id": "50484f18acfa1ca35078a959a4bd9ab0",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "bf206fc65a5c0075",
+ "span_id": "591b29ff93aa6b0d",
+ "start_timestamp": 1769681331.264,
"status": "ok",
+ "timestamp": 1769681331.357665,
+ "trace_id": "835d3accafa53eedb5d6e732a5a2677f",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 43089,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43089,
+ "url": "http://localhost:43089/openai/embeddings",
+ "url.full": "http://localhost:43089/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:43089/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "591b29ff93aa6b0d",
+ "span_id": "04b1aea57bdd9393",
+ "start_timestamp": 1769681331.298,
+ "status": "ok",
+ "timestamp": 1769681331.3548388,
+ "trace_id": "835d3accafa53eedb5d6e732a5a2677f",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "bf206fc65a5c0075",
+ "span_id": "c92053605dfade08",
+ "start_timestamp": 1769681331.358,
+ "status": "ok",
+ "timestamp": 1769681331.3758194,
+ "trace_id": "835d3accafa53eedb5d6e732a5a2677f",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 43089,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43089,
+ "url": "http://localhost:43089/openai/embeddings",
+ "url.full": "http://localhost:43089/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:43089/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c92053605dfade08",
+ "span_id": "af009de74ef9b3ac",
+ "start_timestamp": 1769681331.359,
+ "status": "not_found",
+ "timestamp": 1769681331.3685894,
+ "trac
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.embeddings.input": "Embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
@@ -18,13 +18,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "736f26ffdfb32816",
+ "span_id": "8ec224a60539993c",
+ "start_timestamp": 1769681330.449,
"status": "ok",
+ "timestamp": 1769681330.5366461,
+ "trace_id": "e5198293261416397ad54573ca30892d",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45969,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45969,
+ "url": "http://localhost:45969/openai/embeddings",
+ "url.full": "http://localhost:45969/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:45969/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "8ec224a60539993c",
+ "span_id": "8d4f5578a533d95e",
+ "start_timestamp": 1769681330.464,
+ "status": "ok",
+ "timestamp": 1769681330.5328646,
+ "trace_id": "e5198293261416397ad54573ca30892d",
+ },
+ {
"data": {
"gen_ai.embeddings.input": "Error embedding test!",
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
@@ -32,13 +67,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "736f26ffdfb32816",
+ "span_id": "bf614d7fe9a66de5",
+ "start_timestamp": 1769681330.537,
+ "status": "ok",
+ "timestamp": 1769681330.5626194,
+ "trace_id": "e5198293261416397ad54573ca30892d",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 45969,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45969,
+ "url": "http://localhost:45969/openai/embeddings",
+ "url.full": "http://localhost:45969/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:45969/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "bf614d7fe9a66de5",
+ "span_id": "8fae788ff5d63a61",
+ "start_timestamp": 1769681330.539,
+ "status": "not_found",
+ "timestamp": 1769681330.5517895,
+ "tra
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "1aa43d063327a7b3",
+ "span_id": "bd35119e0da40ec5",
+ "start_timestamp": 1769681329.364,
"status": "ok",
+ "timestamp": 1769681329.4496045,
+ "trace_id": "2bf7b00425f0f47a948a9135526a27c6",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40851,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40851,
+ "url": "http://localhost:40851/openai/embeddings",
+ "url.full": "http://localhost:40851/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:40851/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "bd35119e0da40ec5",
+ "span_id": "f6271676913b1e54",
+ "start_timestamp": 1769681329.392,
+ "status": "ok",
+ "timestamp": 1769681329.446448,
+ "trace_id": "2bf7b00425f0f47a948a9135526a27c6",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "1aa43d063327a7b3",
+ "span_id": "4043df2c1c39c002",
+ "start_timestamp": 1769681329.449,
+ "status": "ok",
+ "timestamp": 1769681329.4639688,
+ "trace_id": "2bf7b00425f0f47a948a9135526a27c6",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 40851,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40851,
+ "url": "http://localhost:40851/openai/embeddings",
+ "url.full": "http://localhost:40851/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40851/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "4043df2c1c39c002",
+ "span_id": "9e597a1d64415f66",
+ "start_timestamp": 1769681329.451,
+ "status": "not_found",
+ "timestamp": 1769681329.4597647,
+ "trace_id": "2bf7b00425f0f47a948a9135526a27c6",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.dimensions": 1536,
"gen_ai.request.encoding_format": "float",
"gen_ai.request.model": "text-embedding-3-small",
@@ -17,24 +17,147 @@
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings text-embedding-3-small",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
+ "parent_span_id": "e25854d04eac2220",
+ "span_id": "3b0714f7e2049cb8",
+ "start_timestamp": 1769681328.694,
"status": "ok",
+ "timestamp": 1769681328.7550454,
+ "trace_id": "98ce4c6899d9ba9f4c9c849bfb05e45c",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 165,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 45615,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45615,
+ "url": "http://localhost:45615/openai/embeddings",
+ "url.full": "http://localhost:45615/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
},
- ObjectContaining {
+ "description": "POST http://localhost:45615/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "3b0714f7e2049cb8",
+ "span_id": "bccdabe14d6b1906",
+ "start_timestamp": 1769681328.707,
+ "status": "ok",
+ "timestamp": 1769681328.7521486,
+ "trace_id": "98ce4c6899d9ba9f4c9c849bfb05e45c",
+ },
+ {
"data": {
"gen_ai.operation.name": "embeddings",
"gen_ai.request.model": "error-model",
"gen_ai.system": "openai",
"sentry.op": "gen_ai.embeddings",
"sentry.origin": "auto.ai.openai",
},
"description": "embeddings error-model",
"op": "gen_ai.embeddings",
"origin": "auto.ai.openai",
- "status": "internal_error",
+ "parent_span_id": "e25854d04eac2220",
+ "span_id": "c47bd3a4f270baa3",
+ "start_timestamp": 1769681328.755,
+ "status": "ok",
+ "timestamp": 1769681328.7698042,
+ "trace_id": "98ce4c6899d9ba9f4c9c849bfb05e45c",
+ },
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 15,
+ "http.response.status_code": 404,
+ "network.peer.address": "::1",
+ "network.peer.port": 45615,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 45615,
+ "url": "http://localhost:45615/openai/embeddings",
+ "url.full": "http://localhost:45615/openai/embeddings",
+ "url.path": "/openai/embeddings",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:45615/openai/embeddings",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "c47bd3a4f270baa3",
+ "span_id": "bb4c3560e8c450e3",
+ "start_timestamp": 1769681328.756,
+ "status": "not_found",
+ "timestamp": 1769681328.7647107,
+ "trace_id": "98ce4c6899d9ba9f4c9c849bfb05e45c",
+ },
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "bbd36c4b98dca178",
+ "span_id": "9fff10647b25f793",
+ "start_timestamp": 1769681317.677,
"status": "ok",
+ "timestamp": 1769681317.7563763,
+ "trace_id": "7e7032ba35e7bba64697981aa93283b4",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38005,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38005,
+ "url": "http://localhost:38005/openai/chat/completions",
+ "url.full": "http://localhost:38005/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38005/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "9fff10647b25f793",
+ "span_id": "35655cafe55c3a36",
+ "start_timestamp": 1769681317.703,
+ "status": "ok",
+ "timestamp": 1769681317.7525446,
+ "trace_id": "7e7032ba35e7bba64697981aa93283b4",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "bbd36c4b98dca178",
+ "span_id": "d4d0f60fde03804c",
+ "start_timestamp": 1769681317.757,
"status": "ok",
+ "timestamp": 1769681317.7764285,
+ "trace_id": "7e7032ba35e7bba64697981aa93283b4",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 38005,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 38005,
+ "url": "http://localhost:38005/openai/responses",
+ "url.full": "http://localhost:38005/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:38005/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "d4d0f60fde03804c",
+ "span_id": "6395b4dfaf12169f",
+ "start_timestamp": 1769681317.759,
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.input.messages": "[{\"role\":\"user\",\"content\":\"What is the capital of France?\"}]",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
@@ -26,13 +26,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "20a8362ba9b9d905",
+ "span_id": "b319b7467e493436",
+ "start_timestamp": 1769681314.297,
"status": "ok",
+ "timestamp": 1769681314.3836043,
+ "trace_id": "ae19f64ee846cbdf671c9c42bbacec69",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 43535,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43535,
+ "url": "http://localhost:43535/openai/chat/completions",
+ "url.full": "http://localhost:43535/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:43535/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b319b7467e493436",
+ "span_id": "84e50098222ac54d",
+ "start_timestamp": 1769681314.313,
+ "status": "ok",
+ "timestamp": 1769681314.3782415,
+ "trace_id": "ae19f64ee846cbdf671c9c42bbacec69",
+ },
+ {
"data": {
"gen_ai.input.messages": "Translate this to French: Hello",
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
@@ -53,13 +88,48 @@
"sentry.sdk_meta.gen_ai.input.messages.original_length": 1,
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "20a8362ba9b9d905",
+ "span_id": "9bcd3fdfe04f308d",
+ "start_timestamp": 1769681314.384,
"status": "ok",
+ "timestamp": 1769681314.401796,
+ "trace_id": "ae19f64ee846cbdf671c9c42bbacec69",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 43535,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 43535,
+ "url": "http://localhost:43535/openai/responses",
+ "url.full": "http://localhost:43535/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:43535/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "9bcd3fdfe04f308d",
+ "span_id": "670eac5e300e679a",
+ "start_timestamp": 1769681314.388,
+
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "508b1587b1a91846",
+ "span_id": "b481708d9cf33d8f",
+ "start_timestamp": 1769681310.445,
"status": "ok",
+ "timestamp": 1769681310.5555973,
+ "trace_id": "588b2f733c3149aada58477d3565d9a6",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41115,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41115,
+ "url": "http://localhost:41115/openai/chat/completions",
+ "url.full": "http://localhost:41115/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:41115/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "b481708d9cf33d8f",
+ "span_id": "2a5efa9bbfba60d8",
+ "start_timestamp": 1769681310.489,
+ "status": "ok",
+ "timestamp": 1769681310.552307,
+ "trace_id": "588b2f733c3149aada58477d3565d9a6",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "508b1587b1a91846",
+ "span_id": "739bf1878205bea4",
+ "start_timestamp": 1769681310.556,
"status": "ok",
+ "timestamp": 1769681310.5755317,
+ "trace_id": "588b2f733c3149aada58477d3565d9a6",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 41115,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 41115,
+ "url": "http://localhost:41115/openai/responses",
+ "url.full": "http://localhost:41115/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:41115/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "739bf1878205bea4",
+ "span_id": "122637a3ca0433c0",
+ "start_timestamp": 1769681310.558,
+ "status": "ok",
+ "timestamp": 1769681310.574401,
+ "trace_id": "588b2f733c3149aada58477d3565d9a6",
+ },
|
|
dev-packages/node-integration-tests/utils/assertions.ts#L35
AssertionError: expected { contexts: { …(8) }, …(14) } to match object { event_id: Any<String>, …(5) }
(90 matching properties omitted from actual)
- Expected
+ Received
@@ -1,9 +1,9 @@
{
"event_id": Any<String>,
- "spans": ArrayContaining [
- ObjectContaining {
+ "spans": [
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.request.temperature": 0.7,
"gen_ai.response.finish_reasons": "[\"stop\"]",
@@ -22,13 +22,48 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "bb621f5d1b6f00e4",
+ "span_id": "5096e2b93a6f233c",
+ "start_timestamp": 1769681306.872,
"status": "ok",
+ "timestamp": 1769681306.9304554,
+ "trace_id": "deaba6fd5039872b991c9e2f51a9b931",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 281,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40973,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40973,
+ "url": "http://localhost:40973/openai/chat/completions",
+ "url.full": "http://localhost:40973/openai/chat/completions",
+ "url.path": "/openai/chat/completions",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40973/openai/chat/completions",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "5096e2b93a6f233c",
+ "span_id": "a840d62050df9055",
+ "start_timestamp": 1769681306.882,
+ "status": "ok",
+ "timestamp": 1769681306.9275591,
+ "trace_id": "deaba6fd5039872b991c9e2f51a9b931",
+ },
+ {
"data": {
"gen_ai.operation.name": "chat",
"gen_ai.request.model": "gpt-3.5-turbo",
"gen_ai.response.finish_reasons": "[\"completed\"]",
"gen_ai.response.id": "resp_mock456",
@@ -46,26 +81,156 @@
"sentry.origin": "auto.ai.openai",
},
"description": "chat gpt-3.5-turbo",
"op": "gen_ai.chat",
"origin": "auto.ai.openai",
+ "parent_span_id": "bb621f5d1b6f00e4",
+ "span_id": "41d4fff8a26146aa",
+ "start_timestamp": 1769681306.931,
"status": "ok",
+ "timestamp": 1769681306.9419591,
+ "trace_id": "deaba6fd5039872b991c9e2f51a9b931",
},
- ObjectContaining {
+ {
+ "data": {
+ "http.request.method": "POST",
+ "http.request.method_original": "POST",
+ "http.response.header.content-length": 435,
+ "http.response.status_code": 200,
+ "network.peer.address": "::1",
+ "network.peer.port": 40973,
+ "otel.kind": "CLIENT",
+ "sentry.op": "http.client",
+ "sentry.origin": "auto.http.otel.node_fetch",
+ "server.address": "localhost",
+ "server.port": 40973,
+ "url": "http://localhost:40973/openai/responses",
+ "url.full": "http://localhost:40973/openai/responses",
+ "url.path": "/openai/responses",
+ "url.query": "",
+ "url.scheme": "http",
+ "user_agent.original": "OpenAI/JS 5.18.1",
+ },
+ "description": "POST http://localhost:40973/openai/responses",
+ "op": "http.client",
+ "origin": "auto.http.otel.node_fetch",
+ "parent_span_id": "41d4fff8a26146aa",
+ "span_id": "9e958272b6abd934",
+ "start_timestamp": 1769681306.932,
+ "status": "ok",
+ "timestamp": 1769681306.9409451,
+ "trace_id": "deaba6fd5039872b991c9e2f51a9b931",
+
|
|
|
The logs for this run have expired and are no longer available.
Loading