@@ -73,7 +73,7 @@ describe('Vercel AI integration', () => {
7373 'ai.pipeline.name' : 'generateText' ,
7474 'ai.prompt' : '{"prompt":"Where is the second span?"}' ,
7575 'ai.response.finishReason' : 'stop' ,
76- 'ai .response.text' : expect . any ( String ) ,
76+ 'gen_ai .response.text' : expect . any ( String ) ,
7777 'ai.settings.maxRetries' : 2 ,
7878 'ai.settings.maxSteps' : 1 ,
7979 'ai.streaming' : false ,
@@ -108,10 +108,10 @@ describe('Vercel AI integration', () => {
108108 'ai.response.finishReason' : 'stop' ,
109109 'ai.response.model' : 'mock-model-id' ,
110110 'ai.response.id' : expect . any ( String ) ,
111- 'ai .response.text' : expect . any ( String ) ,
111+ 'gen_ai .response.text' : expect . any ( String ) ,
112112 'ai.response.timestamp' : expect . any ( String ) ,
113113 'ai.prompt.format' : expect . any ( String ) ,
114- 'ai.prompt .messages' : expect . any ( String ) ,
114+ 'gen_ai.request .messages' : expect . any ( String ) ,
115115 'gen_ai.response.finish_reasons' : [ 'stop' ] ,
116116 'gen_ai.usage.input_tokens' : 10 ,
117117 'gen_ai.usage.output_tokens' : 20 ,
@@ -210,7 +210,7 @@ describe('Vercel AI integration', () => {
210210 'ai.pipeline.name' : 'generateText' ,
211211 'ai.prompt' : '{"prompt":"Where is the first span?"}' ,
212212 'ai.response.finishReason' : 'stop' ,
213- 'ai .response.text' : 'First span here!' ,
213+ 'gen_ai .response.text' : 'First span here!' ,
214214 'ai.settings.maxRetries' : 2 ,
215215 'ai.settings.maxSteps' : 1 ,
216216 'ai.streaming' : false ,
@@ -236,11 +236,11 @@ describe('Vercel AI integration', () => {
236236 'ai.operationId' : 'ai.generateText.doGenerate' ,
237237 'ai.pipeline.name' : 'generateText.doGenerate' ,
238238 'ai.prompt.format' : 'prompt' ,
239- 'ai.prompt .messages' : '[{"role":"user","content":[{"type":"text","text":"Where is the first span?"}]}]' ,
239+ 'gen_ai.request .messages' : '[{"role":"user","content":[{"type":"text","text":"Where is the first span?"}]}]' ,
240240 'ai.response.finishReason' : 'stop' ,
241241 'ai.response.id' : expect . any ( String ) ,
242242 'ai.response.model' : 'mock-model-id' ,
243- 'ai .response.text' : 'First span here!' ,
243+ 'gen_ai .response.text' : 'First span here!' ,
244244 'ai.response.timestamp' : expect . any ( String ) ,
245245 'ai.settings.maxRetries' : 2 ,
246246 'ai.streaming' : false ,
@@ -270,7 +270,7 @@ describe('Vercel AI integration', () => {
270270 'ai.pipeline.name' : 'generateText' ,
271271 'ai.prompt' : '{"prompt":"Where is the second span?"}' ,
272272 'ai.response.finishReason' : 'stop' ,
273- 'ai .response.text' : expect . any ( String ) ,
273+ 'gen_ai .response.text' : expect . any ( String ) ,
274274 'ai.settings.maxRetries' : 2 ,
275275 'ai.settings.maxSteps' : 1 ,
276276 'ai.streaming' : false ,
@@ -305,10 +305,10 @@ describe('Vercel AI integration', () => {
305305 'ai.response.finishReason' : 'stop' ,
306306 'ai.response.model' : 'mock-model-id' ,
307307 'ai.response.id' : expect . any ( String ) ,
308- 'ai .response.text' : expect . any ( String ) ,
308+ 'gen_ai .response.text' : expect . any ( String ) ,
309309 'ai.response.timestamp' : expect . any ( String ) ,
310310 'ai.prompt.format' : expect . any ( String ) ,
311- 'ai.prompt .messages' : expect . any ( String ) ,
311+ 'gen_ai.request .messages' : expect . any ( String ) ,
312312 'gen_ai.response.finish_reasons' : [ 'stop' ] ,
313313 'gen_ai.usage.input_tokens' : 10 ,
314314 'gen_ai.usage.output_tokens' : 20 ,
@@ -330,8 +330,8 @@ describe('Vercel AI integration', () => {
330330 'ai.pipeline.name' : 'generateText' ,
331331 'ai.prompt' : '{"prompt":"What is the weather in San Francisco?"}' ,
332332 'ai.response.finishReason' : 'tool-calls' ,
333- 'ai .response.text' : 'Tool call completed!' ,
334- 'ai .response.toolCalls ' : expect . any ( String ) ,
333+ 'gen_ai .response.text' : 'Tool call completed!' ,
334+ 'gen_ai .response.tool_calls ' : expect . any ( String ) ,
335335 'ai.settings.maxRetries' : 2 ,
336336 'ai.settings.maxSteps' : 1 ,
337337 'ai.streaming' : false ,
@@ -357,15 +357,15 @@ describe('Vercel AI integration', () => {
357357 'ai.operationId' : 'ai.generateText.doGenerate' ,
358358 'ai.pipeline.name' : 'generateText.doGenerate' ,
359359 'ai.prompt.format' : expect . any ( String ) ,
360- 'ai.prompt .messages' : expect . any ( String ) ,
360+ 'gen_ai.request .messages' : expect . any ( String ) ,
361361 'ai.prompt.toolChoice' : expect . any ( String ) ,
362- 'ai.prompt.tools ' : expect . any ( Array ) ,
362+ 'gen_ai.request.available_tools ' : expect . any ( Array ) ,
363363 'ai.response.finishReason' : 'tool-calls' ,
364364 'ai.response.id' : expect . any ( String ) ,
365365 'ai.response.model' : 'mock-model-id' ,
366- 'ai .response.text' : 'Tool call completed!' ,
366+ 'gen_ai .response.text' : 'Tool call completed!' ,
367367 'ai.response.timestamp' : expect . any ( String ) ,
368- 'ai .response.toolCalls ' : expect . any ( String ) ,
368+ 'gen_ai .response.tool_calls ' : expect . any ( String ) ,
369369 'ai.settings.maxRetries' : 2 ,
370370 'ai.streaming' : false ,
371371 'gen_ai.request.model' : 'mock-model-id' ,
0 commit comments