@@ -158,6 +158,8 @@ TRACE_EVENT(io_uring_queue_async_work,
158
158
__field ( unsigned int , flags )
159
159
__field ( struct io_wq_work * , work )
160
160
__field ( int , rw )
161
+
162
+ __string ( op_str , io_uring_get_opcode (opcode ) )
161
163
),
162
164
163
165
TP_fast_assign (
@@ -168,11 +170,13 @@ TRACE_EVENT(io_uring_queue_async_work,
168
170
__entry -> opcode = opcode ;
169
171
__entry -> work = work ;
170
172
__entry -> rw = rw ;
173
+
174
+ __assign_str (op_str , io_uring_get_opcode (opcode ));
171
175
),
172
176
173
177
TP_printk ("ring %p, request %p, user_data 0x%llx, opcode %s, flags 0x%x, %s queue, work %p" ,
174
178
__entry -> ctx , __entry -> req , __entry -> user_data ,
175
- io_uring_get_opcode ( __entry -> opcode ),
179
+ __get_str ( op_str ),
176
180
__entry -> flags , __entry -> rw ? "hashed" : "normal" , __entry -> work )
177
181
);
178
182
@@ -198,18 +202,22 @@ TRACE_EVENT(io_uring_defer,
198
202
__field ( void * , req )
199
203
__field ( unsigned long long, data )
200
204
__field ( u8 , opcode )
205
+
206
+ __string ( op_str , io_uring_get_opcode (opcode ) )
201
207
),
202
208
203
209
TP_fast_assign (
204
210
__entry -> ctx = ctx ;
205
211
__entry -> req = req ;
206
212
__entry -> data = user_data ;
207
213
__entry -> opcode = opcode ;
214
+
215
+ __assign_str (op_str , io_uring_get_opcode (opcode ));
208
216
),
209
217
210
218
TP_printk ("ring %p, request %p, user_data 0x%llx, opcode %s" ,
211
219
__entry -> ctx , __entry -> req , __entry -> data ,
212
- io_uring_get_opcode ( __entry -> opcode ))
220
+ __get_str ( op_str ))
213
221
);
214
222
215
223
/**
@@ -298,6 +306,8 @@ TRACE_EVENT(io_uring_fail_link,
298
306
__field ( unsigned long long, user_data )
299
307
__field ( u8 , opcode )
300
308
__field ( void * , link )
309
+
310
+ __string ( op_str , io_uring_get_opcode (opcode ) )
301
311
),
302
312
303
313
TP_fast_assign (
@@ -306,11 +316,13 @@ TRACE_EVENT(io_uring_fail_link,
306
316
__entry -> user_data = user_data ;
307
317
__entry -> opcode = opcode ;
308
318
__entry -> link = link ;
319
+
320
+ __assign_str (op_str , io_uring_get_opcode (opcode ));
309
321
),
310
322
311
323
TP_printk ("ring %p, request %p, user_data 0x%llx, opcode %s, link %p" ,
312
324
__entry -> ctx , __entry -> req , __entry -> user_data ,
313
- io_uring_get_opcode ( __entry -> opcode ), __entry -> link )
325
+ __get_str ( op_str ), __entry -> link )
314
326
);
315
327
316
328
/**
@@ -390,6 +402,8 @@ TRACE_EVENT(io_uring_submit_sqe,
390
402
__field ( u32 , flags )
391
403
__field ( bool , force_nonblock )
392
404
__field ( bool , sq_thread )
405
+
406
+ __string ( op_str , io_uring_get_opcode (opcode ) )
393
407
),
394
408
395
409
TP_fast_assign (
@@ -400,11 +414,13 @@ TRACE_EVENT(io_uring_submit_sqe,
400
414
__entry -> flags = flags ;
401
415
__entry -> force_nonblock = force_nonblock ;
402
416
__entry -> sq_thread = sq_thread ;
417
+
418
+ __assign_str (op_str , io_uring_get_opcode (opcode ));
403
419
),
404
420
405
421
TP_printk ("ring %p, req %p, user_data 0x%llx, opcode %s, flags 0x%x, "
406
422
"non block %d, sq_thread %d" , __entry -> ctx , __entry -> req ,
407
- __entry -> user_data , io_uring_get_opcode ( __entry -> opcode ),
423
+ __entry -> user_data , __get_str ( op_str ),
408
424
__entry -> flags , __entry -> force_nonblock , __entry -> sq_thread )
409
425
);
410
426
@@ -435,6 +451,8 @@ TRACE_EVENT(io_uring_poll_arm,
435
451
__field ( u8 , opcode )
436
452
__field ( int , mask )
437
453
__field ( int , events )
454
+
455
+ __string ( op_str , io_uring_get_opcode (opcode ) )
438
456
),
439
457
440
458
TP_fast_assign (
@@ -444,11 +462,13 @@ TRACE_EVENT(io_uring_poll_arm,
444
462
__entry -> opcode = opcode ;
445
463
__entry -> mask = mask ;
446
464
__entry -> events = events ;
465
+
466
+ __assign_str (op_str , io_uring_get_opcode (opcode ));
447
467
),
448
468
449
469
TP_printk ("ring %p, req %p, user_data 0x%llx, opcode %s, mask 0x%x, events 0x%x" ,
450
470
__entry -> ctx , __entry -> req , __entry -> user_data ,
451
- io_uring_get_opcode ( __entry -> opcode ),
471
+ __get_str ( op_str ),
452
472
__entry -> mask , __entry -> events )
453
473
);
454
474
@@ -474,6 +494,8 @@ TRACE_EVENT(io_uring_task_add,
474
494
__field ( unsigned long long, user_data )
475
495
__field ( u8 , opcode )
476
496
__field ( int , mask )
497
+
498
+ __string ( op_str , io_uring_get_opcode (opcode ) )
477
499
),
478
500
479
501
TP_fast_assign (
@@ -482,11 +504,13 @@ TRACE_EVENT(io_uring_task_add,
482
504
__entry -> user_data = user_data ;
483
505
__entry -> opcode = opcode ;
484
506
__entry -> mask = mask ;
507
+
508
+ __assign_str (op_str , io_uring_get_opcode (opcode ));
485
509
),
486
510
487
511
TP_printk ("ring %p, req %p, user_data 0x%llx, opcode %s, mask %x" ,
488
512
__entry -> ctx , __entry -> req , __entry -> user_data ,
489
- io_uring_get_opcode ( __entry -> opcode ),
513
+ __get_str ( op_str ),
490
514
__entry -> mask )
491
515
);
492
516
@@ -523,6 +547,8 @@ TRACE_EVENT(io_uring_req_failed,
523
547
__field ( u64 , pad1 )
524
548
__field ( u64 , addr3 )
525
549
__field ( int , error )
550
+
551
+ __string ( op_str , io_uring_get_opcode (sqe -> opcode ) )
526
552
),
527
553
528
554
TP_fast_assign (
@@ -542,6 +568,8 @@ TRACE_EVENT(io_uring_req_failed,
542
568
__entry -> pad1 = sqe -> __pad2 [0 ];
543
569
__entry -> addr3 = sqe -> addr3 ;
544
570
__entry -> error = error ;
571
+
572
+ __assign_str (op_str , io_uring_get_opcode (sqe -> opcode ));
545
573
),
546
574
547
575
TP_printk ("ring %p, req %p, user_data 0x%llx, "
@@ -550,7 +578,7 @@ TRACE_EVENT(io_uring_req_failed,
550
578
"personality=%d, file_index=%d, pad=0x%llx, addr3=%llx, "
551
579
"error=%d" ,
552
580
__entry -> ctx , __entry -> req , __entry -> user_data ,
553
- io_uring_get_opcode ( __entry -> opcode ),
581
+ __get_str ( op_str ),
554
582
__entry -> flags , __entry -> ioprio ,
555
583
(unsigned long long )__entry -> off ,
556
584
(unsigned long long ) __entry -> addr , __entry -> len ,
0 commit comments