@@ -94,7 +94,7 @@ describe('Google GenAI integration', () => {
9494 'gen_ai.request.temperature' : 0.8 ,
9595 'gen_ai.request.top_p' : 0.9 ,
9696 'gen_ai.request.max_tokens' : 150 ,
97- 'gen_ai.request .messages' : '[{"role":"user","parts":[{"text":"Hello, how are you?"}]}]' ,
97+ 'gen_ai.input .messages' : '[{"role":"user","parts":[{"text":"Hello, how are you?"}]}]' ,
9898 } ) ,
9999 description : 'chat gemini-1.5-pro create' ,
100100 op : 'gen_ai.chat' ,
@@ -109,7 +109,7 @@ describe('Google GenAI integration', () => {
109109 'sentry.origin' : 'auto.ai.google_genai' ,
110110 'gen_ai.system' : 'google_genai' ,
111111 'gen_ai.request.model' : 'gemini-1.5-pro' ,
112- 'gen_ai.request .messages' : expect . any ( String ) , // Should include message when recordInputs: true
112+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include message when recordInputs: true
113113 'gen_ai.response.text' : expect . any ( String ) , // Should include response when recordOutputs: true
114114 'gen_ai.usage.input_tokens' : 8 ,
115115 'gen_ai.usage.output_tokens' : 12 ,
@@ -131,7 +131,7 @@ describe('Google GenAI integration', () => {
131131 'gen_ai.request.temperature' : 0.7 ,
132132 'gen_ai.request.top_p' : 0.9 ,
133133 'gen_ai.request.max_tokens' : 100 ,
134- 'gen_ai.request .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
134+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
135135 'gen_ai.response.text' : expect . any ( String ) , // Should include response when recordOutputs: true
136136 'gen_ai.usage.input_tokens' : 8 ,
137137 'gen_ai.usage.output_tokens' : 12 ,
@@ -150,7 +150,7 @@ describe('Google GenAI integration', () => {
150150 'sentry.origin' : 'auto.ai.google_genai' ,
151151 'gen_ai.system' : 'google_genai' ,
152152 'gen_ai.request.model' : 'error-model' ,
153- 'gen_ai.request .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
153+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
154154 } ) ,
155155 description : 'models error-model' ,
156156 op : 'gen_ai.models' ,
@@ -166,7 +166,7 @@ describe('Google GenAI integration', () => {
166166 // Check that custom options are respected
167167 expect . objectContaining ( {
168168 data : expect . objectContaining ( {
169- 'gen_ai.request .messages' : expect . any ( String ) , // Should include messages when recordInputs: true
169+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include messages when recordInputs: true
170170 'gen_ai.response.text' : expect . any ( String ) , // Should include response text when recordOutputs: true
171171 } ) ,
172172 description : expect . not . stringContaining ( 'stream-response' ) , // Non-streaming span
@@ -219,7 +219,7 @@ describe('Google GenAI integration', () => {
219219 'gen_ai.system' : 'google_genai' ,
220220 'gen_ai.request.model' : 'gemini-2.0-flash-001' ,
221221 'gen_ai.request.available_tools' : EXPECTED_AVAILABLE_TOOLS_JSON ,
222- 'gen_ai.request .messages' : expect . any ( String ) , // Should include contents
222+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include contents
223223 'gen_ai.response.text' : expect . any ( String ) , // Should include response text
224224 'gen_ai.response.tool_calls' : expect . any ( String ) , // Should include tool calls
225225 'gen_ai.usage.input_tokens' : 15 ,
@@ -240,7 +240,7 @@ describe('Google GenAI integration', () => {
240240 'gen_ai.system' : 'google_genai' ,
241241 'gen_ai.request.model' : 'gemini-2.0-flash-001' ,
242242 'gen_ai.request.available_tools' : EXPECTED_AVAILABLE_TOOLS_JSON ,
243- 'gen_ai.request .messages' : expect . any ( String ) , // Should include contents
243+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include contents
244244 'gen_ai.response.streaming' : true ,
245245 'gen_ai.response.text' : expect . any ( String ) , // Should include response text
246246 'gen_ai.response.tool_calls' : expect . any ( String ) , // Should include tool calls
@@ -263,7 +263,7 @@ describe('Google GenAI integration', () => {
263263 'sentry.origin' : 'auto.ai.google_genai' ,
264264 'gen_ai.system' : 'google_genai' ,
265265 'gen_ai.request.model' : 'gemini-2.0-flash-001' ,
266- 'gen_ai.request .messages' : expect . any ( String ) , // Should include contents
266+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include contents
267267 'gen_ai.response.text' : expect . any ( String ) , // Should include response text
268268 'gen_ai.usage.input_tokens' : 8 ,
269269 'gen_ai.usage.output_tokens' : 12 ,
@@ -385,7 +385,7 @@ describe('Google GenAI integration', () => {
385385 'gen_ai.request.temperature' : 0.7 ,
386386 'gen_ai.request.top_p' : 0.9 ,
387387 'gen_ai.request.max_tokens' : 100 ,
388- 'gen_ai.request .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
388+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
389389 'gen_ai.response.streaming' : true ,
390390 'gen_ai.response.id' : 'mock-response-streaming-id' ,
391391 'gen_ai.response.model' : 'gemini-1.5-pro' ,
@@ -424,7 +424,7 @@ describe('Google GenAI integration', () => {
424424 'sentry.origin' : 'auto.ai.google_genai' ,
425425 'gen_ai.system' : 'google_genai' ,
426426 'gen_ai.request.model' : 'gemini-1.5-pro' ,
427- 'gen_ai.request .messages' : expect . any ( String ) , // Should include message when recordInputs: true
427+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include message when recordInputs: true
428428 'gen_ai.response.streaming' : true ,
429429 'gen_ai.response.id' : 'mock-response-streaming-id' ,
430430 'gen_ai.response.model' : 'gemini-1.5-pro' ,
@@ -447,7 +447,7 @@ describe('Google GenAI integration', () => {
447447 'gen_ai.system' : 'google_genai' ,
448448 'gen_ai.request.model' : 'blocked-model' ,
449449 'gen_ai.request.temperature' : 0.7 ,
450- 'gen_ai.request .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
450+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
451451 'gen_ai.response.streaming' : true ,
452452 } ) ,
453453 description : 'models blocked-model stream-response' ,
@@ -464,7 +464,7 @@ describe('Google GenAI integration', () => {
464464 'gen_ai.system' : 'google_genai' ,
465465 'gen_ai.request.model' : 'error-model' ,
466466 'gen_ai.request.temperature' : 0.7 ,
467- 'gen_ai.request .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
467+ 'gen_ai.input .messages' : expect . any ( String ) , // Should include contents when recordInputs: true
468468 } ) ,
469469 description : 'models error-model stream-response' ,
470470 op : 'gen_ai.models' ,
@@ -511,7 +511,7 @@ describe('Google GenAI integration', () => {
511511 'gen_ai.system' : 'google_genai' ,
512512 'gen_ai.request.model' : 'gemini-1.5-flash' ,
513513 // Messages should be present (truncation happened) and should be a JSON array with parts
514- 'gen_ai.request .messages' : expect . stringMatching (
514+ 'gen_ai.input .messages' : expect . stringMatching (
515515 / ^ \[ \{ " r o l e " : " u s e r " , " p a r t s " : \[ \{ " t e x t " : " C + " \} \] \} \] $ / ,
516516 ) ,
517517 } ) ,
@@ -529,7 +529,7 @@ describe('Google GenAI integration', () => {
529529 'gen_ai.system' : 'google_genai' ,
530530 'gen_ai.request.model' : 'gemini-1.5-flash' ,
531531 // Small message should be kept intact
532- 'gen_ai.request .messages' : JSON . stringify ( [
532+ 'gen_ai.input .messages' : JSON . stringify ( [
533533 {
534534 role : 'user' ,
535535 parts : [ { text : 'This is a small message that fits within the limit' } ] ,
0 commit comments