@@ -74,7 +74,7 @@ struct ompi_part_persist_t {
74
74
unused in part_comm. */
75
75
ompi_request_t * part_comm_sreq ;
76
76
int32_t part_comm_sready ;
77
- int32_t init_comms ;
77
+ int32_t init_comms ;
78
78
int32_t init_world ;
79
79
int32_t my_world_rank ; /* Because the back end communicators use a world rank, we need to communicate ours
80
80
to set up the requests. */
@@ -138,12 +138,12 @@ __opal_attribute_always_inline__ static inline void
138
138
mca_part_persist_complete (struct mca_part_persist_request_t * request )
139
139
{
140
140
if (MCA_PART_PERSIST_REQUEST_PRECV == request -> req_type ) {
141
- request -> req_ompi .req_status .MPI_SOURCE = request -> req_peer ;
141
+ request -> req_ompi .req_status .MPI_SOURCE = request -> req_peer ;
142
142
} else {
143
143
request -> req_ompi .req_status .MPI_SOURCE = request -> req_comm -> c_my_rank ;
144
144
}
145
145
request -> req_ompi .req_complete_cb = NULL ;
146
- request -> req_ompi .req_status .MPI_TAG = request -> req_tag ;
146
+ request -> req_ompi .req_status .MPI_TAG = request -> req_tag ;
147
147
request -> req_ompi .req_status ._ucount = request -> req_bytes ;
148
148
request -> req_ompi .req_status .MPI_ERROR = OMPI_SUCCESS ;
149
149
request -> req_part_complete = true;
@@ -189,7 +189,6 @@ mca_part_persist_progress(void)
189
189
if (OMPI_SUCCESS != err ) goto end_part_progress ;
190
190
ompi_part_persist .part_comm_sready = 0 ;
191
191
ompi_part_persist .init_world = 1 ;
192
-
193
192
completed ++ ;
194
193
goto end_part_progress ;
195
194
}
@@ -214,16 +213,16 @@ mca_part_persist_progress(void)
214
213
OPAL_LIST_FOREACH (current , ompi_part_persist .progress_list , mca_part_persist_list_t ) {
215
214
mca_part_persist_request_t * req = (mca_part_persist_request_t * ) current -> item ;
216
215
217
- /* Check to see if request is initilaized */
216
+ /* Check to see if request is initialized */
218
217
if (false == req -> initialized ) {
219
- int done = 0 ;
220
-
218
+ int done = 0 ;
219
+
221
220
if (true == req -> flag_post_setup_recv ) {
222
221
err = MCA_PML_CALL (irecv (& (req -> setup_info [1 ]), sizeof (struct ompi_mca_persist_setup_t ), MPI_BYTE , OMPI_ANY_SOURCE , req -> my_recv_tag , ompi_part_persist .part_comm_setup , & req -> setup_req [1 ]));
223
222
if (OMPI_SUCCESS != err ) goto end_part_progress ;
224
223
req -> flag_post_setup_recv = false;
225
- }
226
-
224
+ }
225
+
227
226
err = ompi_request_test (& (req -> setup_req [1 ]), & done , MPI_STATUS_IGNORE );
228
227
if (OMPI_SUCCESS != err ) goto end_part_progress ;
229
228
@@ -233,7 +232,7 @@ mca_part_persist_progress(void)
233
232
234
233
if (MCA_PART_PERSIST_REQUEST_PSEND == req -> req_type ) {
235
234
/* parse message */
236
- req -> world_peer = req -> setup_info [1 ].world_rank ;
235
+ req -> world_peer = req -> setup_info [1 ].world_rank ;
237
236
238
237
err = opal_datatype_type_size (& (req -> req_datatype -> super ), & dt_size_ );
239
238
if (OMPI_SUCCESS != err ) goto end_part_progress ;
@@ -246,24 +245,22 @@ mca_part_persist_progress(void)
246
245
void * buf = ((void * ) (((char * )req -> req_addr ) + (bytes * i )));
247
246
err = MCA_PML_CALL (isend_init (buf , req -> real_count , req -> req_datatype , req -> world_peer , req -> my_send_tag + i , MCA_PML_BASE_SEND_STANDARD , ompi_part_persist .part_comm , & (req -> persist_reqs [i ])));
248
247
if (OMPI_SUCCESS != err ) goto end_part_progress ;
249
- }
248
+ }
250
249
} else {
251
250
/* parse message */
252
- req -> world_peer = req -> setup_info [1 ].world_rank ;
251
+ req -> world_peer = req -> setup_info [1 ].world_rank ;
253
252
req -> my_send_tag = req -> setup_info [1 ].start_tag ;
254
253
req -> my_recv_tag = req -> setup_info [1 ].setup_tag ;
255
254
req -> real_parts = req -> setup_info [1 ].num_parts ;
256
255
req -> real_count = req -> setup_info [1 ].count ;
257
256
req -> real_dt_size = req -> setup_info [1 ].dt_size ;
258
257
259
-
260
258
err = opal_datatype_type_size (& (req -> req_datatype -> super ), & dt_size_ );
261
259
if (OMPI_SUCCESS != err ) goto end_part_progress ;
262
260
dt_size = (dt_size_ > (size_t ) UINT_MAX ) ? MPI_UNDEFINED : (uint32_t ) dt_size_ ;
263
261
uint32_t bytes = req -> real_count * dt_size ;
264
262
265
-
266
- /* Set up persistent sends */
263
+ /* Set up persistent receives */
267
264
req -> persist_reqs = (ompi_request_t * * ) malloc (sizeof (ompi_request_t * )* (req -> real_parts ));
268
265
req -> flags = (int * ) calloc (req -> real_parts ,sizeof (int ));
269
266
if (req -> real_dt_size == dt_size ) {
@@ -293,10 +290,9 @@ mca_part_persist_progress(void)
293
290
}
294
291
} else {
295
292
if (false == req -> req_part_complete && REQUEST_COMPLETED != req -> req_ompi .req_complete && OMPI_REQUEST_ACTIVE == req -> req_ompi .req_state ) {
296
- for (i = 0 ; i < req -> real_parts ; i ++ ) {
297
-
298
- /* Check to see if partition is queued for being started. Only applicable to sends. */
299
- if (-2 == req -> flags [i ]) {
293
+ for (i = 0 ; i < req -> real_parts ; i ++ ) {
294
+ /* Check to see if partition is queued for being started. Only applicable to sends. */
295
+ if (-2 == req -> flags [i ]) {
300
296
err = req -> persist_reqs [i ]-> req_start (1 , (& (req -> persist_reqs [i ])));
301
297
if (OMPI_SUCCESS != err ) goto end_part_progress ;
302
298
req -> flags [i ] = 0 ;
@@ -343,7 +339,7 @@ mca_part_persist_progress(void)
343
339
344
340
__opal_attribute_always_inline__ static inline int
345
341
mca_part_persist_precv_init (void * buf ,
346
- size_t parts ,
342
+ size_t parts ,
347
343
size_t count ,
348
344
ompi_datatype_t * datatype ,
349
345
int src ,
@@ -374,9 +370,9 @@ mca_part_persist_precv_init(void *buf,
374
370
375
371
mca_part_persist_request_t * req = (mca_part_persist_request_t * ) recvreq ;
376
372
377
- /* Set lazy initializion flags */
373
+ /* Set lazy initialization flags */
378
374
req -> initialized = false;
379
- req -> first_send = true;
375
+ req -> first_send = true;
380
376
req -> flag_post_setup_recv = false;
381
377
req -> flags = NULL ;
382
378
/* Non-blocking receive on setup info */
@@ -398,7 +394,7 @@ mca_part_persist_precv_init(void *buf,
398
394
/* Add element to progress engine */
399
395
new_progress_elem = OBJ_NEW (mca_part_persist_list_t );
400
396
new_progress_elem -> item = req ;
401
- req -> progress_elem = new_progress_elem ;
397
+ req -> progress_elem = new_progress_elem ;
402
398
OPAL_THREAD_LOCK (& ompi_part_persist .lock );
403
399
opal_list_append (ompi_part_persist .progress_list , (opal_list_item_t * )new_progress_elem );
404
400
OPAL_THREAD_UNLOCK (& ompi_part_persist .lock );
@@ -441,7 +437,6 @@ mca_part_persist_psend_init(const void* buf,
441
437
/* Set lazy initialization variables */
442
438
req -> initialized = false;
443
439
req -> first_send = true;
444
-
445
440
446
441
/* Determine total bytes to send. */
447
442
err = opal_datatype_type_size (& (req -> req_datatype -> super ), & dt_size_ );
@@ -451,7 +446,7 @@ mca_part_persist_psend_init(const void* buf,
451
446
452
447
/* non-blocking send set-up data */
453
448
req -> setup_info [0 ].world_rank = ompi_comm_rank (& ompi_mpi_comm_world .comm );
454
- req -> setup_info [0 ].start_tag = ompi_part_persist .next_send_tag ; ompi_part_persist .next_send_tag += parts ;
449
+ req -> setup_info [0 ].start_tag = ompi_part_persist .next_send_tag ; ompi_part_persist .next_send_tag += parts ;
455
450
req -> my_send_tag = req -> setup_info [0 ].start_tag ;
456
451
req -> setup_info [0 ].setup_tag = ompi_part_persist .next_recv_tag ; ompi_part_persist .next_recv_tag ++ ;
457
452
req -> my_recv_tag = req -> setup_info [0 ].setup_tag ;
@@ -475,7 +470,7 @@ mca_part_persist_psend_init(const void* buf,
475
470
req -> flag_post_setup_recv = true;
476
471
}
477
472
478
- /* Initilaize completion variables */
473
+ /* Initialize completion variables */
479
474
sendreq -> req_base .req_ompi .req_persistent = true;
480
475
req -> req_part_complete = true;
481
476
req -> req_ompi .req_complete = REQUEST_COMPLETED ;
@@ -522,15 +517,15 @@ mca_part_persist_start(size_t count, ompi_request_t** requests)
522
517
}
523
518
} else {
524
519
req -> done_count = 0 ;
525
- }
526
- }
527
- req -> req_ompi .req_state = OMPI_REQUEST_ACTIVE ;
520
+ }
521
+ }
522
+ req -> req_ompi .req_state = OMPI_REQUEST_ACTIVE ;
528
523
req -> req_ompi .req_status .MPI_TAG = MPI_ANY_TAG ;
529
524
req -> req_ompi .req_status .MPI_ERROR = OMPI_SUCCESS ;
530
525
req -> req_ompi .req_status ._cancelled = 0 ;
531
526
req -> req_part_complete = false;
532
527
req -> req_ompi .req_complete = false;
533
- OPAL_ATOMIC_SWAP_PTR (& req -> req_ompi .req_complete , REQUEST_PENDING );
528
+ OPAL_ATOMIC_SWAP_PTR (& req -> req_ompi .req_complete , REQUEST_PENDING );
534
529
}
535
530
536
531
return err ;
@@ -564,7 +559,7 @@ mca_part_persist_pready(size_t min_part,
564
559
__opal_attribute_always_inline__ static inline int
565
560
mca_part_persist_parrived (size_t min_part ,
566
561
size_t max_part ,
567
- int * flag ,
562
+ int * flag ,
568
563
ompi_request_t * request )
569
564
{
570
565
int err = OMPI_SUCCESS ;
@@ -576,7 +571,7 @@ mca_part_persist_parrived(size_t min_part,
576
571
_flag = 1 ;
577
572
if (req -> req_parts == req -> real_parts ) {
578
573
for (i = min_part ; i <= max_part ; i ++ ) {
579
- _flag = _flag && req -> flags [i ];
574
+ _flag = _flag && req -> flags [i ];
580
575
}
581
576
} else {
582
577
float convert = ((float )req -> real_parts ) / ((float )req -> req_parts );
@@ -590,7 +585,7 @@ mca_part_persist_parrived(size_t min_part,
590
585
591
586
if (!_flag ) {
592
587
opal_progress ();
593
- }
588
+ }
594
589
* flag = _flag ;
595
590
return err ;
596
591
}
@@ -614,4 +609,4 @@ mca_part_persist_free(ompi_request_t** request)
614
609
615
610
END_C_DECLS
616
611
617
- #endif /* PART_PERSIST_H_HAS_BEEN_INCLUDED */
612
+ #endif /* PART_PERSIST_H */
0 commit comments