X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/347996b4a10c4e8579080692afa60e0afb88b60a..a5945452ef5b1114fd7c2fab0e865d776775c34d:/src/smpi/smpi_replay.cpp diff --git a/src/smpi/smpi_replay.cpp b/src/smpi/smpi_replay.cpp index 4648703d05..3929b82947 100644 --- a/src/smpi/smpi_replay.cpp +++ b/src/smpi/smpi_replay.cpp @@ -29,44 +29,38 @@ static void log_timed_action (const char *const *action, double clock){ if (XBT_LOG_ISENABLED(smpi_replay, xbt_log_priority_verbose)){ char *name = xbt_str_join_array(action, " "); XBT_VERB("%s %f", name, smpi_process_simulated_elapsed()-clock); - free(name); + xbt_free(name); } } - -static xbt_dynar_t get_reqq_self(){ - char * key; - - int size = asprintf(&key, "%d", smpi_process_index()); - if(size==-1) - xbt_die("could not allocate memory for asprintf"); - xbt_dynar_t dynar_mpi_request = (xbt_dynar_t) xbt_dict_get(reqq, key); - free(key); +static xbt_dynar_t get_reqq_self() +{ + char * key = bprintf("%d", smpi_process_index()); + xbt_dynar_t dynar_mpi_request = static_cast(xbt_dict_get(reqq, key)); + xbt_free(key); return dynar_mpi_request; } -static void set_reqq_self(xbt_dynar_t mpi_request){ - char * key; - - int size = asprintf(&key, "%d", smpi_process_index()); - if(size==-1) - xbt_die("could not allocate memory for asprintf"); +static void set_reqq_self(xbt_dynar_t mpi_request) +{ + char * key = bprintf("%d", smpi_process_index()); xbt_dict_set(reqq, key, mpi_request, free); - free(key); + xbt_free(key); } - //allocate a single buffer for all sends, growing it if needed -void* smpi_get_tmp_sendbuffer(int size){ +void* smpi_get_tmp_sendbuffer(int size) +{ if (!smpi_process_get_replaying()) - return xbt_malloc(size); + return xbt_malloc(size); if (sendbuffer_size(xbt_realloc(sendbuffer,size)); sendbuffer_size=size; } return sendbuffer; } + //allocate a single buffer for all recv void* smpi_get_tmp_recvbuffer(int size){ if (!smpi_process_get_replaying()) @@ -97,9 +91,7 @@ static double parse_double(const char *string) static MPI_Datatype decode_datatype(const char *const action) { // Declared datatypes, - - switch(atoi(action)) - { + switch(atoi(action)) { case 0: MPI_CURRENT_TYPE=MPI_DOUBLE; break; @@ -123,7 +115,6 @@ static MPI_Datatype decode_datatype(const char *const action) break; default: MPI_CURRENT_TYPE=MPI_DEFAULT_TYPE; - } return MPI_CURRENT_TYPE; } @@ -131,10 +122,10 @@ static MPI_Datatype decode_datatype(const char *const action) const char* encode_datatype(MPI_Datatype datatype, int* known) { - //default type for output is set to MPI_BYTE // MPI_DEFAULT_TYPE is not set for output, use directly MPI_BYTE - if(known)*known=1; + if(known!=NULL) + *known=1; if (datatype==MPI_BYTE){ return ""; } @@ -151,7 +142,8 @@ const char* encode_datatype(MPI_Datatype datatype, int* known) if(datatype==MPI_FLOAT) return "5"; //tell that the datatype is not handled by replay, and that its size should be measured and replayed as size*MPI_BYTE - if(known)*known=0; + if(known!=NULL) + *known=0; // default - not implemented. // do not warn here as we pass in this function even for other trace formats return "-1"; @@ -168,12 +160,12 @@ const char* encode_datatype(MPI_Datatype datatype, int* known) "Please contact the Simgrid team if support is needed", __FUNCTION__, i, mandatory, optional);\ } - static void action_init(const char *const *action) { XBT_DEBUG("Initialize the counters"); - CHECK_ACTION_PARAMS(action, 0, 1); - if(action[2]) MPI_DEFAULT_TYPE= MPI_DOUBLE; // default MPE dataype + CHECK_ACTION_PARAMS(action, 0, 1) + if(action[2]) + MPI_DEFAULT_TYPE=MPI_DOUBLE; // default MPE dataype else MPI_DEFAULT_TYPE= MPI_BYTE; // default TAU datatype /* start a simulated timer */ @@ -181,24 +173,16 @@ static void action_init(const char *const *action) /*initialize the number of active processes */ active_processes = smpi_process_count(); - if (!reqq) { + if (reqq==NULL) { reqq = xbt_dict_new(); } set_reqq_self(xbt_dynar_new(sizeof(MPI_Request),&xbt_free_ref)); - - /* - reqq=xbt_new0(xbt_dynar_t,active_processes); - - for(i=0;itype=TRACING_TEST; @@ -401,9 +390,7 @@ static void action_test(const char *const *action){ flag = smpi_mpi_test(&request, &status); XBT_DEBUG("MPI_Test result: %d", flag); - /* push back request in dynar to be caught by a subsequent wait. if the test - * did succeed, the request is now NULL. - */ + /* push back request in dynar to be caught by a subsequent wait. if the test did succeed, the request is now NULL.*/ xbt_dynar_push_as(get_reqq_self(),MPI_Request, request); TRACE_smpi_testing_out(rank); @@ -412,7 +399,7 @@ static void action_test(const char *const *action){ } static void action_wait(const char *const *action){ - CHECK_ACTION_PARAMS(action, 0, 0); + CHECK_ACTION_PARAMS(action, 0, 0) double clock = smpi_process_simulated_elapsed(); MPI_Request request; MPI_Status status; @@ -422,16 +409,12 @@ static void action_wait(const char *const *action){ xbt_str_join_array(action," ")); request = xbt_dynar_pop_as(get_reqq_self(),MPI_Request); - if (!request){ - /* Assuming that the trace is well formed, this mean the comm might have - * been caught by a MPI_test. Then just return. - */ + if (request==NULL){ + /* Assume that the trace is well formed, meaning the comm might have been caught by a MPI_test. Then just return.*/ return; } - int rank = request->comm != MPI_COMM_NULL - ? smpi_comm_rank(request->comm) - : -1; + int rank = request->comm != MPI_COMM_NULL ? smpi_comm_rank(request->comm) : -1; MPI_Group group = smpi_comm_group(request->comm); int src_traced = smpi_group_rank(group, request->src); @@ -450,7 +433,7 @@ static void action_wait(const char *const *action){ } static void action_waitall(const char *const *action){ - CHECK_ACTION_PARAMS(action, 0, 0); + CHECK_ACTION_PARAMS(action, 0, 0) double clock = smpi_process_simulated_elapsed(); int count_requests=0; unsigned int i=0; @@ -466,11 +449,10 @@ static void action_waitall(const char *const *action){ xbt_dynar_foreach(get_reqq_self(),i,requests[i]); //save information from requests - xbt_dynar_t srcs = xbt_dynar_new(sizeof(int), NULL); xbt_dynar_t dsts = xbt_dynar_new(sizeof(int), NULL); xbt_dynar_t recvs = xbt_dynar_new(sizeof(int), NULL); - for (i = 0; (int)i < count_requests; i++) { + for (i = 0; static_cast(i) < count_requests; i++) { if(requests[i]){ int *asrc = xbt_new(int, 1); int *adst = xbt_new(int, 1); @@ -500,7 +482,7 @@ static void action_waitall(const char *const *action){ smpi_mpi_waitall(count_requests, requests, status); - for (i = 0; (int)i < count_requests; i++) { + for (i = 0; static_cast(i) < count_requests; i++) { int src_traced, dst_traced, is_wait_for_receive; xbt_dynar_get_cpy(srcs, i, &src_traced); xbt_dynar_get_cpy(dsts, i, &dst_traced); @@ -514,8 +496,6 @@ static void action_waitall(const char *const *action){ xbt_dynar_free(&srcs); xbt_dynar_free(&dsts); xbt_dynar_free(&recvs); - - //TODO xbt_dynar_free_container(get_reqq_self()); set_reqq_self(xbt_dynar_new(sizeof(MPI_Request),&xbt_free_ref)); } log_timed_action (action, clock); @@ -534,17 +514,13 @@ static void action_barrier(const char *const *action){ log_timed_action (action, clock); } - static void action_bcast(const char *const *action) { - CHECK_ACTION_PARAMS(action, 1, 2); + CHECK_ACTION_PARAMS(action, 1, 2) double size = parse_double(action[2]); double clock = smpi_process_simulated_elapsed(); int root=0; - /* - * Initialize MPI_CURRENT_TYPE in order to decrease - * the number of the checks - * */ + /* Initialize MPI_CURRENT_TYPE in order to decrease the number of the checks */ MPI_CURRENT_TYPE= MPI_DEFAULT_TYPE; if(action[3]) { @@ -573,7 +549,7 @@ static void action_bcast(const char *const *action) static void action_reduce(const char *const *action) { - CHECK_ACTION_PARAMS(action, 2, 2); + CHECK_ACTION_PARAMS(action, 2, 2) double comm_size = parse_double(action[2]); double comp_size = parse_double(action[3]); double clock = smpi_process_simulated_elapsed(); @@ -586,8 +562,6 @@ static void action_reduce(const char *const *action) MPI_CURRENT_TYPE=decode_datatype(action[5]); } } - - int rank = smpi_process_index(); int root_traced = smpi_group_rank(smpi_comm_group(MPI_COMM_WORLD), root); @@ -602,20 +576,22 @@ static void action_reduce(const char *const *action) void *recvbuf = smpi_get_tmp_sendbuffer(comm_size* smpi_datatype_size(MPI_CURRENT_TYPE)); void *sendbuf = smpi_get_tmp_sendbuffer(comm_size* smpi_datatype_size(MPI_CURRENT_TYPE)); - mpi_coll_reduce_fun(sendbuf, recvbuf, comm_size, MPI_CURRENT_TYPE, MPI_OP_NULL, root, MPI_COMM_WORLD); - smpi_execute_flops(comp_size); + mpi_coll_reduce_fun(sendbuf, recvbuf, comm_size, MPI_CURRENT_TYPE, MPI_OP_NULL, root, MPI_COMM_WORLD); + smpi_execute_flops(comp_size); TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); log_timed_action (action, clock); } static void action_allReduce(const char *const *action) { - CHECK_ACTION_PARAMS(action, 2, 1); + CHECK_ACTION_PARAMS(action, 2, 1) double comm_size = parse_double(action[2]); double comp_size = parse_double(action[3]); - if(action[4]) MPI_CURRENT_TYPE=decode_datatype(action[4]); - else MPI_CURRENT_TYPE= MPI_DEFAULT_TYPE; + if(action[4]) + MPI_CURRENT_TYPE=decode_datatype(action[4]); + else + MPI_CURRENT_TYPE= MPI_DEFAULT_TYPE; double clock = smpi_process_simulated_elapsed(); int rank = smpi_process_index(); @@ -636,7 +612,7 @@ static void action_allReduce(const char *const *action) { } static void action_allToAll(const char *const *action) { - CHECK_ACTION_PARAMS(action, 2, 2); //two mandatory (send and recv volumes) + CHECK_ACTION_PARAMS(action, 2, 2) //two mandatory (send and recv volumes) //two optional (corresponding datatypes) double clock = smpi_process_simulated_elapsed(); int comm_size = smpi_comm_size(MPI_COMM_WORLD); @@ -672,11 +648,8 @@ static void action_allToAll(const char *const *action) { log_timed_action (action, clock); } - static void action_gather(const char *const *action) { - /* - The structure of the gather action for the rank 0 (total 4 processes) - is the following: + /* The structure of the gather action for the rank 0 (total 4 processes) is the following: 0 gather 68 68 0 0 0 where: @@ -686,7 +659,7 @@ static void action_gather(const char *const *action) { 4) 0 is the send datatype id, see decode_datatype() 5) 0 is the recv datatype id, see decode_datatype() */ - CHECK_ACTION_PARAMS(action, 2, 3); + CHECK_ACTION_PARAMS(action, 2, 3) double clock = smpi_process_simulated_elapsed(); int comm_size = smpi_comm_size(MPI_COMM_WORLD); int send_size = parse_double(action[2]); @@ -719,20 +692,14 @@ static void action_gather(const char *const *action) { TRACE_smpi_collective_in(smpi_process_index(), root, __FUNCTION__, extra); - mpi_coll_gather_fun(send, send_size, MPI_CURRENT_TYPE, - recv, recv_size, MPI_CURRENT_TYPE2, - root, MPI_COMM_WORLD); + mpi_coll_gather_fun(send, send_size, MPI_CURRENT_TYPE, recv, recv_size, MPI_CURRENT_TYPE2, root, MPI_COMM_WORLD); TRACE_smpi_collective_out(smpi_process_index(), -1, __FUNCTION__); log_timed_action (action, clock); } - - static void action_gatherv(const char *const *action) { - /* - The structure of the gatherv action for the rank 0 (total 4 processes) - is the following: + /* The structure of the gatherv action for the rank 0 (total 4 processes) is the following: 0 gather 68 68 10 10 10 0 0 0 where: @@ -745,7 +712,7 @@ static void action_gatherv(const char *const *action) { double clock = smpi_process_simulated_elapsed(); int comm_size = smpi_comm_size(MPI_COMM_WORLD); - CHECK_ACTION_PARAMS(action, comm_size+1, 2); + CHECK_ACTION_PARAMS(action, comm_size+1, 2) int send_size = parse_double(action[2]); int *disps = xbt_new0(int, comm_size); int *recvcounts = xbt_new0(int, comm_size); @@ -768,7 +735,7 @@ static void action_gatherv(const char *const *action) { } int root=atoi(action[3+comm_size]); - int rank = smpi_comm_rank(MPI_COMM_WORLD);; + int rank = smpi_comm_rank(MPI_COMM_WORLD); if(rank==root) recv = smpi_get_tmp_recvbuffer(recv_sum* smpi_datatype_size(MPI_CURRENT_TYPE2)); @@ -786,9 +753,7 @@ static void action_gatherv(const char *const *action) { TRACE_smpi_collective_in(smpi_process_index(), root, __FUNCTION__, extra); - smpi_mpi_gatherv(send, send_size, MPI_CURRENT_TYPE, - recv, recvcounts, disps, MPI_CURRENT_TYPE2, - root, MPI_COMM_WORLD); + smpi_mpi_gatherv(send, send_size, MPI_CURRENT_TYPE, recv, recvcounts, disps, MPI_CURRENT_TYPE2, root, MPI_COMM_WORLD); TRACE_smpi_collective_out(smpi_process_index(), -1, __FUNCTION__); log_timed_action (action, clock); @@ -797,10 +762,7 @@ static void action_gatherv(const char *const *action) { } static void action_reducescatter(const char *const *action) { - - /* - The structure of the reducescatter action for the rank 0 (total 4 processes) - is the following: + /* The structure of the reducescatter action for the rank 0 (total 4 processes) is the following: 0 reduceScatter 275427 275427 275427 204020 11346849 0 where: @@ -808,13 +770,10 @@ static void action_reducescatter(const char *const *action) { 2) The value 11346849 is the amount of instructions 3) The last value corresponds to the datatype, see decode_datatype(). - We analyze a MPI_Reduce_scatter call to one MPI_Reduce and one MPI_Scatterv. - - */ - + We analyze a MPI_Reduce_scatter call to one MPI_Reduce and one MPI_Scatterv. */ double clock = smpi_process_simulated_elapsed(); int comm_size = smpi_comm_size(MPI_COMM_WORLD); - CHECK_ACTION_PARAMS(action, comm_size+1, 1); + CHECK_ACTION_PARAMS(action, comm_size+1, 1) int comp_size = parse_double(action[2+comm_size]); int *recvcounts = xbt_new0(int, comm_size); int *disps = xbt_new0(int, comm_size); @@ -847,11 +806,9 @@ static void action_reducescatter(const char *const *action) { void *sendbuf = smpi_get_tmp_sendbuffer(size* smpi_datatype_size(MPI_CURRENT_TYPE)); void *recvbuf = smpi_get_tmp_recvbuffer(size* smpi_datatype_size(MPI_CURRENT_TYPE)); - mpi_coll_reduce_scatter_fun(sendbuf, recvbuf, recvcounts, MPI_CURRENT_TYPE, MPI_OP_NULL, - MPI_COMM_WORLD); + mpi_coll_reduce_scatter_fun(sendbuf, recvbuf, recvcounts, MPI_CURRENT_TYPE, MPI_OP_NULL, MPI_COMM_WORLD); smpi_execute_flops(comp_size); - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); xbt_free(recvcounts); xbt_free(disps); @@ -859,22 +816,16 @@ static void action_reducescatter(const char *const *action) { } static void action_allgather(const char *const *action) { - /* - The structure of the allgather action for the rank 0 (total 4 processes) - is the following: + /* The structure of the allgather action for the rank 0 (total 4 processes) is the following: 0 allGather 275427 275427 where: 1) 275427 is the sendcount 2) 275427 is the recvcount - 3) No more values mean that the datatype for sent and receive buffer - is the default one, see decode_datatype(). - - */ - + 3) No more values mean that the datatype for sent and receive buffer is the default one, see decode_datatype(). */ double clock = smpi_process_simulated_elapsed(); - CHECK_ACTION_PARAMS(action, 2, 2); + CHECK_ACTION_PARAMS(action, 2, 2) int sendcount=atoi(action[2]); int recvcount=atoi(action[3]); @@ -885,7 +836,7 @@ static void action_allgather(const char *const *action) { MPI_CURRENT_TYPE2 = decode_datatype(action[5]); } else { MPI_CURRENT_TYPE = MPI_DEFAULT_TYPE; - MPI_CURRENT_TYPE2 = MPI_DEFAULT_TYPE; + MPI_CURRENT_TYPE2 = MPI_DEFAULT_TYPE; } void *sendbuf = smpi_get_tmp_sendbuffer(sendcount* smpi_datatype_size(MPI_CURRENT_TYPE)); void *recvbuf = smpi_get_tmp_recvbuffer(recvcount* smpi_datatype_size(MPI_CURRENT_TYPE2)); @@ -908,24 +859,18 @@ static void action_allgather(const char *const *action) { } static void action_allgatherv(const char *const *action) { - - /* - The structure of the allgatherv action for the rank 0 (total 4 processes) - is the following: + /* The structure of the allgatherv action for the rank 0 (total 4 processes) is the following: 0 allGatherV 275427 275427 275427 275427 204020 where: 1) 275427 is the sendcount 2) The next four elements declare the recvcounts array 3) No more values mean that the datatype for sent and receive buffer - is the default one, see decode_datatype(). - - */ - + is the default one, see decode_datatype(). */ double clock = smpi_process_simulated_elapsed(); int comm_size = smpi_comm_size(MPI_COMM_WORLD); - CHECK_ACTION_PARAMS(action, comm_size+1, 2); + CHECK_ACTION_PARAMS(action, comm_size+1, 2) int i=0; int sendcount=atoi(action[2]); int *recvcounts = xbt_new0(int, comm_size); @@ -961,7 +906,8 @@ static void action_allgatherv(const char *const *action) { TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); - mpi_coll_allgatherv_fun(sendbuf, sendcount, MPI_CURRENT_TYPE, recvbuf, recvcounts, disps, MPI_CURRENT_TYPE2, MPI_COMM_WORLD); + mpi_coll_allgatherv_fun(sendbuf, sendcount, MPI_CURRENT_TYPE, recvbuf, recvcounts, disps, MPI_CURRENT_TYPE2, + MPI_COMM_WORLD); TRACE_smpi_collective_out(rank, -1, __FUNCTION__); log_timed_action (action, clock); @@ -970,24 +916,18 @@ static void action_allgatherv(const char *const *action) { } static void action_allToAllv(const char *const *action) { - /* - The structure of the allToAllV action for the rank 0 (total 4 processes) - is the following: + /* The structure of the allToAllV action for the rank 0 (total 4 processes) is the following: 0 allToAllV 100 1 7 10 12 100 1 70 10 5 where: 1) 100 is the size of the send buffer *sizeof(int), 2) 1 7 10 12 is the sendcounts array 3) 100*sizeof(int) is the size of the receiver buffer - 4) 1 70 10 5 is the recvcounts array - - */ - - + 4) 1 70 10 5 is the recvcounts array */ double clock = smpi_process_simulated_elapsed(); int comm_size = smpi_comm_size(MPI_COMM_WORLD); - CHECK_ACTION_PARAMS(action, 2*comm_size+2, 2); + CHECK_ACTION_PARAMS(action, 2*comm_size+2, 2) int send_buf_size=0,recv_buf_size=0,i=0; int *sendcounts = xbt_new0(int, comm_size); int *recvcounts = xbt_new0(int, comm_size); @@ -1015,7 +955,6 @@ static void action_allToAllv(const char *const *action) { recvcounts[i] = atoi(action[i+4+comm_size]); } - int rank = smpi_process_index(); instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); extra->type = TRACING_ALLTOALLV; @@ -1034,9 +973,8 @@ static void action_allToAllv(const char *const *action) { TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); - mpi_coll_alltoallv_fun(sendbuf, sendcounts, senddisps, MPI_CURRENT_TYPE, - recvbuf, recvcounts, recvdisps, MPI_CURRENT_TYPE, - MPI_COMM_WORLD); + mpi_coll_alltoallv_fun(sendbuf, sendcounts, senddisps, MPI_CURRENT_TYPE,recvbuf, recvcounts, recvdisps, + MPI_CURRENT_TYPE, MPI_COMM_WORLD); TRACE_smpi_collective_out(rank, -1, __FUNCTION__); log_timed_action (action, clock); @@ -1050,7 +988,7 @@ void smpi_replay_run(int *argc, char***argv){ /* First initializes everything */ smpi_process_init(argc, argv); smpi_process_mark_as_initialized(); - smpi_process_set_replaying(1); + smpi_process_set_replaying(true); int rank = smpi_process_index(); TRACE_smpi_init(rank); @@ -1060,9 +998,9 @@ void smpi_replay_run(int *argc, char***argv){ char *operation =bprintf("%s_init",__FUNCTION__); TRACE_smpi_collective_in(rank, -1, operation, extra); TRACE_smpi_collective_out(rank, -1, operation); - free(operation); + xbt_free(operation); - if (!_xbt_replay_action_init()) { + if (_xbt_replay_action_init()==0) { xbt_replay_action_register("init", action_init); xbt_replay_action_register("finalize", action_finalize); xbt_replay_action_register("comm_size", action_comm_size); @@ -1088,7 +1026,7 @@ void smpi_replay_run(int *argc, char***argv){ xbt_replay_action_register("reduceScatter", action_reducescatter); xbt_replay_action_register("compute", action_compute); } - + //if we have a delayed start, sleep here. if(*argc>2){ char *endptr; @@ -1102,16 +1040,15 @@ void smpi_replay_run(int *argc, char***argv){ XBT_DEBUG("Force context switch by smpi_execute_flops - Sleeping for 0.0 flops "); smpi_execute_flops(0.0); } - + /* Actually run the replay */ xbt_replay_action_runner(*argc, *argv); /* and now, finalize everything */ double sim_time= 1.; /* One active process will stop. Decrease the counter*/ - XBT_DEBUG("There are %lu elements in reqq[*]", - xbt_dynar_length(get_reqq_self())); - if (!xbt_dynar_is_empty(get_reqq_self())){ + XBT_DEBUG("There are %lu elements in reqq[*]", xbt_dynar_length(get_reqq_self())); + if (xbt_dynar_is_empty(get_reqq_self())==0){ int count_requests=xbt_dynar_length(get_reqq_self()); MPI_Request requests[count_requests]; MPI_Status status[count_requests]; @@ -1124,25 +1061,18 @@ void smpi_replay_run(int *argc, char***argv){ active_processes--; } - if(!active_processes){ + if(active_processes==0){ /* Last process alive speaking */ /* end the simulated timer */ sim_time = smpi_process_simulated_elapsed(); - } - - - //TODO xbt_dynar_free_container(get_reqq_self())); - - if(!active_processes){ XBT_INFO("Simulation time %f", sim_time); _xbt_replay_action_exit(); xbt_free(sendbuffer); xbt_free(recvbuffer); - //xbt_free(reqq); xbt_dict_free(&reqq); //not need, data have been freed ??? reqq = NULL; } - + instr_extra_data extra_fin = xbt_new0(s_instr_extra_data_t,1); extra_fin->type = TRACING_FINALIZE; operation =bprintf("%s_finalize",__FUNCTION__); @@ -1153,5 +1083,5 @@ void smpi_replay_run(int *argc, char***argv){ TRACE_smpi_collective_out(rank, -1, operation); TRACE_smpi_finalize(smpi_process_index()); smpi_process_destroy(); - free(operation); + xbt_free(operation); }