@@ -100,38 +100,19 @@ describe('Plugin', () => {
100
100
expect ( span . meta ) . to . have . property ( 'span.kind' , 'client' )
101
101
102
102
expect ( span . meta ) . to . have . property ( 'vertexai.request.model' , 'gemini-1.5-flash-002' )
103
- expect ( span . meta ) . to . have . property ( 'vertexai.request.contents.0.role' , 'user' )
104
- expect ( span . meta ) . to . have . property ( 'vertexai.request.contents.0.parts.0.text' , 'Hello, how are you?' )
105
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.finish_reason' , 'STOP' )
106
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.parts.0.text' ,
107
- 'Hello! How can I assist you today?' )
108
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.role' , 'model' )
109
-
110
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.prompt_tokens' , 35 )
111
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.completion_tokens' , 2 )
112
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.total_tokens' , 37 )
113
-
114
- if ( model . systemInstruction ) {
115
- expect ( span . meta ) . to . have . property ( 'vertexai.request.system_instruction.0.text' ,
116
- 'Please provide an answer' )
117
- }
118
- expect ( span . meta ) . to . have . property ( 'vertexai.request.generation_config.max_output_tokens' , '50' )
119
- expect ( span . meta ) . to . have . property ( 'vertexai.request.generation_config.temperature' , '1' )
120
103
} )
121
104
122
105
const { response } = await model . generateContent ( {
123
106
contents : [ { role : 'user' , parts : [ { text : 'Hello, how are you?' } ] } ]
124
107
} )
125
-
126
108
expect ( response ) . to . have . property ( 'candidates' )
127
109
128
110
await checkTraces
129
111
} )
130
112
131
113
it ( 'makes a successful call with a string argument' , async ( ) => {
132
114
const checkTraces = agent . assertSomeTraces ( traces => {
133
- expect ( traces [ 0 ] [ 0 ] . meta ) . to . have . property ( 'vertexai.request.contents.0.text' ,
134
- 'Hello, how are you?' )
115
+ expect ( traces [ 0 ] [ 0 ] . meta ) . to . have . property ( 'vertexai.request.model' , 'gemini-1.5-flash-002' )
135
116
} )
136
117
137
118
const { response } = await model . generateContent ( 'Hello, how are you?' )
@@ -148,11 +129,7 @@ describe('Plugin', () => {
148
129
const checkTraces = agent . assertSomeTraces ( traces => {
149
130
const span = traces [ 0 ] [ 0 ]
150
131
151
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.parts.0.text' , 'undefined' )
152
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.parts.0.function_call.name' ,
153
- 'add' )
154
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.parts.0.function_call.args' ,
155
- JSON . stringify ( { a : 2 , b : 2 } ) )
132
+ expect ( span . meta ) . to . have . property ( 'vertexai.request.model' , 'gemini-1.5-flash-002' )
156
133
} )
157
134
158
135
await model . generateContent ( 'what is 2 + 2?' )
@@ -174,24 +151,6 @@ describe('Plugin', () => {
174
151
expect ( span . meta ) . to . have . property ( 'span.kind' , 'client' )
175
152
176
153
expect ( span . meta ) . to . have . property ( 'vertexai.request.model' , 'gemini-1.5-flash-002' )
177
- expect ( span . meta ) . to . have . property ( 'vertexai.request.contents.0.text' , 'Hello, how are you?' )
178
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.finish_reason' , 'STOP' )
179
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.parts.0.text' ,
180
- 'Hi, how are you doing today my friend?' )
181
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.role' , 'model' )
182
-
183
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.prompt_tokens' , 5 )
184
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.completion_tokens' , 10 )
185
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.total_tokens' , 15 )
186
-
187
- if ( model . systemInstruction ) {
188
- expect ( span . meta ) . to . have . property ( 'vertexai.request.system_instruction.0.text' ,
189
- 'Please provide an answer' )
190
- }
191
- expect ( span . meta ) . to . have . property ( 'vertexai.request.generation_config.max_output_tokens' , '50' )
192
- expect ( span . meta ) . to . have . property ( 'vertexai.request.generation_config.temperature' , '1' )
193
-
194
- expect ( span . metrics ) . to . have . property ( 'vertexai.request.stream' , 1 )
195
154
} )
196
155
197
156
const { stream, response } = await model . generateContentStream ( 'Hello, how are you?' )
@@ -226,28 +185,6 @@ describe('Plugin', () => {
226
185
expect ( span . meta ) . to . have . property ( 'span.kind' , 'client' )
227
186
228
187
expect ( span . meta ) . to . have . property ( 'vertexai.request.model' , 'gemini-1.5-flash-002' )
229
-
230
- expect ( span . meta ) . to . have . property ( 'vertexai.request.contents.0.role' , 'user' )
231
- expect ( span . meta ) . to . have . property ( 'vertexai.request.contents.0.parts.0.text' , 'Foobar?' )
232
- expect ( span . meta ) . to . have . property ( 'vertexai.request.contents.1.role' , 'model' )
233
- expect ( span . meta ) . to . have . property ( 'vertexai.request.contents.1.parts.0.text' , 'Foobar!' )
234
- expect ( span . meta ) . to . have . property ( 'vertexai.request.contents.2.parts.0.text' , 'Hello, how are you?' )
235
-
236
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.finish_reason' , 'STOP' )
237
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.parts.0.text' ,
238
- 'Hello! How can I assist you today?' )
239
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.role' , 'model' )
240
-
241
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.prompt_tokens' , 35 )
242
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.completion_tokens' , 2 )
243
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.total_tokens' , 37 )
244
-
245
- if ( model . systemInstruction ) {
246
- expect ( span . meta ) . to . have . property ( 'vertexai.request.system_instruction.0.text' ,
247
- 'Please provide an answer' )
248
- }
249
- expect ( span . meta ) . to . have . property ( 'vertexai.request.generation_config.max_output_tokens' , '50' )
250
- expect ( span . meta ) . to . have . property ( 'vertexai.request.generation_config.temperature' , '1' )
251
188
} )
252
189
253
190
const chat = model . startChat ( {
@@ -265,8 +202,7 @@ describe('Plugin', () => {
265
202
266
203
it ( 'tags a string input' , async ( ) => {
267
204
const checkTraces = agent . assertSomeTraces ( traces => {
268
- expect ( traces [ 0 ] [ 0 ] . meta ) . to . have . property ( 'vertexai.request.contents.0.text' ,
269
- 'Hello, how are you?' )
205
+ expect ( traces [ 0 ] [ 0 ] . meta ) . to . have . property ( 'vertexai.request.model' , 'gemini-1.5-flash-002' )
270
206
} )
271
207
272
208
const chat = model . startChat ( { } )
@@ -279,10 +215,7 @@ describe('Plugin', () => {
279
215
280
216
it ( 'tags an array of string inputs' , async ( ) => {
281
217
const checkTraces = agent . assertSomeTraces ( traces => {
282
- expect ( traces [ 0 ] [ 0 ] . meta ) . to . have . property ( 'vertexai.request.contents.0.text' ,
283
- 'Hello, how are you?' )
284
- expect ( traces [ 0 ] [ 0 ] . meta ) . to . have . property ( 'vertexai.request.contents.1.text' ,
285
- 'What should I do today?' )
218
+ expect ( traces [ 0 ] [ 0 ] . meta ) . to . have . property ( 'vertexai.request.model' , 'gemini-1.5-flash-002' )
286
219
} )
287
220
288
221
const chat = model . startChat ( { } )
@@ -306,24 +239,6 @@ describe('Plugin', () => {
306
239
expect ( span . meta ) . to . have . property ( 'span.kind' , 'client' )
307
240
308
241
expect ( span . meta ) . to . have . property ( 'vertexai.request.model' , 'gemini-1.5-flash-002' )
309
- expect ( span . meta ) . to . have . property ( 'vertexai.request.contents.0.text' , 'Hello, how are you?' )
310
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.finish_reason' , 'STOP' )
311
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.parts.0.text' ,
312
- 'Hi, how are you doing today my friend?' )
313
- expect ( span . meta ) . to . have . property ( 'vertexai.response.candidates.0.content.role' , 'model' )
314
-
315
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.prompt_tokens' , 5 )
316
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.completion_tokens' , 10 )
317
- expect ( span . metrics ) . to . have . property ( 'vertexai.response.usage.total_tokens' , 15 )
318
-
319
- if ( model . systemInstruction ) {
320
- expect ( span . meta ) . to . have . property ( 'vertexai.request.system_instruction.0.text' ,
321
- 'Please provide an answer' )
322
- }
323
- expect ( span . meta ) . to . have . property ( 'vertexai.request.generation_config.max_output_tokens' , '50' )
324
- expect ( span . meta ) . to . have . property ( 'vertexai.request.generation_config.temperature' , '1' )
325
-
326
- expect ( span . metrics ) . to . have . property ( 'vertexai.request.stream' , 1 )
327
242
} )
328
243
329
244
const chat = model . startChat ( { } )
0 commit comments