From cda303b7f7d0015dc738b9fc967aa91be8c669b0 Mon Sep 17 00:00:00 2001 From: =?utf8?q?St=C3=A9phane=20Castelli?= Date: Mon, 17 Mar 2014 09:17:03 +0100 Subject: [PATCH] First step in the implementation of topologies in SMPI --- buildtools/Cmake/AddTests.cmake | 3 + buildtools/Cmake/DefinePackages.cmake | 1 + include/smpi/smpi.h | 6 + src/smpi/private.h | 16 +- src/smpi/smpi_comm.c | 21 +- src/smpi/smpi_global.c | 4 +- src/smpi/smpi_pmpi.c | 1246 +++++++++-------- src/smpi/smpi_topo.c | 468 +++++++ teshsuite/smpi/mpich3-test/testlist | 2 +- .../smpi/mpich3-test/topo/CMakeLists.txt | 28 +- teshsuite/smpi/mpich3-test/topo/testlist | 8 +- 11 files changed, 1192 insertions(+), 611 deletions(-) create mode 100644 src/smpi/smpi_topo.c diff --git a/buildtools/Cmake/AddTests.cmake b/buildtools/Cmake/AddTests.cmake index 1d6641a11a..21e0cff84b 100644 --- a/buildtools/Cmake/AddTests.cmake +++ b/buildtools/Cmake/AddTests.cmake @@ -555,6 +555,9 @@ if(NOT enable_memcheck) ADD_TEST(smpi-mpich3-coll-ompi-thread ${CMAKE_COMMAND} -E chdir ${CMAKE_BINARY_DIR}/teshsuite/smpi/mpich3-test/coll perl ${CMAKE_HOME_DIRECTORY}/teshsuite/smpi/mpich3-test/runtests -mpiexec=${CMAKE_BINARY_DIR}/smpi_script/bin/smpirun -srcdir=${CMAKE_HOME_DIRECTORY}/teshsuite/smpi/mpich3-test/coll -tests=testlist -execarg=--cfg=contexts/factory:thread -execarg=--cfg=smpi/coll_selector:ompi -execarg=--cfg=smpi/send_is_detached_thres:0) ADD_TEST(smpi-mpich3-coll-mpich-thread ${CMAKE_COMMAND} -E chdir ${CMAKE_BINARY_DIR}/teshsuite/smpi/mpich3-test/coll perl ${CMAKE_HOME_DIRECTORY}/teshsuite/smpi/mpich3-test/runtests -mpiexec=${CMAKE_BINARY_DIR}/smpi_script/bin/smpirun -srcdir=${CMAKE_HOME_DIRECTORY}/teshsuite/smpi/mpich3-test/coll -tests=testlist -execarg=--cfg=contexts/factory:thread -execarg=--cfg=smpi/coll_selector:mpich) set_tests_properties(smpi-mpich3-coll-thread smpi-mpich3-coll-ompi-thread smpi-mpich3-coll-mpich-thread PROPERTIES PASS_REGULAR_EXPRESSION "tests passed!") + + ADD_TEST(smpi-mpich3-topo-raw ${CMAKE_COMMAND} -E chdir ${CMAKE_BINARY_DIR}/teshsuite/smpi/mpich3-test/topo perl ${CMAKE_HOME_DIRECTORY}/teshsuite/smpi/mpich3-test/runtests -mpiexec=${CMAKE_BINARY_DIR}/smpi_script/bin/smpirun -srcdir=${CMAKE_HOME_DIRECTORY}/teshsuite/smpi/mpich3-test/topo -tests=testlist -execarg=--cfg=contexts/factory:raw) + set_tests_properties(smpi-mpich3-topo-raw PROPERTIES PASS_REGULAR_EXPRESSION "tests passed!") if(CONTEXT_UCONTEXT) ADD_TEST(smpi-mpich3-coll-ompi-ucontext ${CMAKE_COMMAND} -E chdir ${CMAKE_BINARY_DIR}/teshsuite/smpi/mpich3-test/coll perl ${CMAKE_HOME_DIRECTORY}/teshsuite/smpi/mpich3-test/runtests -mpiexec=${CMAKE_BINARY_DIR}/smpi_script/bin/smpirun -srcdir=${CMAKE_HOME_DIRECTORY}/teshsuite/smpi/mpich3-test/coll -tests=testlist -execarg=--cfg=contexts/factory:ucontext -execarg=--cfg=smpi/coll_selector:ompi -execarg=--cfg=smpi/send_is_detached_thres:0) set_tests_properties(smpi-mpich3-coll-ompi-ucontext PROPERTIES PASS_REGULAR_EXPRESSION "tests passed!") diff --git a/buildtools/Cmake/DefinePackages.cmake b/buildtools/Cmake/DefinePackages.cmake index 3cda3678e7..a89ac22bc1 100644 --- a/buildtools/Cmake/DefinePackages.cmake +++ b/buildtools/Cmake/DefinePackages.cmake @@ -140,6 +140,7 @@ set(SMPI_SRC src/smpi/smpi_mpi_dt.c src/smpi/smpi_pmpi.c src/smpi/smpi_replay.c + src/smpi/smpi_topo.c src/smpi/colls/smpi_openmpi_selector.c src/smpi/colls/smpi_mpich_selector.c src/smpi/colls/colls_global.c diff --git a/include/smpi/smpi.h b/include/smpi/smpi.h index 5fd5a7c651..49c20d4dfc 100644 --- a/include/smpi/smpi.h +++ b/include/smpi/smpi.h @@ -64,6 +64,9 @@ SG_BEGIN_DECL() #define MPI_ERR_PENDING 14 #define MPI_ERR_BUFFER 15 #define MPI_ERR_NAME 16 +#define MPI_ERR_DIMS 17 +#define MPI_ERR_TOPOLOGY 18 +#define MPI_ERR_NO_MEM 19 #define MPI_ERRCODES_IGNORE (int *)0 #define MPI_IDENT 0 #define MPI_SIMILAR 1 @@ -233,6 +236,9 @@ XBT_PUBLIC_DATA( MPI_Op ) MPI_BAND; XBT_PUBLIC_DATA( MPI_Op ) MPI_BOR; XBT_PUBLIC_DATA( MPI_Op ) MPI_BXOR; +struct s_smpi_mpi_topology; +typedef struct s_smpi_mpi_topology *MPI_Topology; + struct s_smpi_mpi_group; typedef struct s_smpi_mpi_group *MPI_Group; diff --git a/src/smpi/private.h b/src/smpi/private.h index a0af7500a2..f6e09f612f 100644 --- a/src/smpi/private.h +++ b/src/smpi/private.h @@ -110,6 +110,19 @@ int smpi_process_finalized(void); int smpi_process_initialized(void); void smpi_process_mark_as_initialized(void); +void smpi_topo_destroy(MPI_Topology topo); +MPI_Topology smpi_topo_create(int ndims); +int smpi_mpi_cart_create(MPI_Comm comm_old, int ndims, int dims[], + int periodic[], int reorder, MPI_Comm *comm_cart); +int smpi_mpi_cart_shift(MPI_Comm comm, int direction, int disp, + int *rank_source, int *rank_dest); +int smpi_mpi_cart_rank(MPI_Comm comm, int* coords, int* rank); +int smpi_mpi_cart_get(MPI_Comm comm, int maxdims, int* dims, int* periods, int* coords); +int smpi_mpi_cart_coords(MPI_Comm comm, int rank, int maxdims, + int coords[]); +int smpi_mpi_cartdim_get(MPI_Comm comm, int *ndims); +int smpi_mpi_dims_create(int nnodes, int ndims, int dims[]); +int smpi_mpi_cart_sub(MPI_Comm comm, const int remain_dims[], MPI_Comm *newcomm); smpi_process_data_t smpi_process_data(void); smpi_process_data_t smpi_process_remote_data(int index); @@ -181,7 +194,8 @@ int smpi_group_unuse(MPI_Group group); int smpi_group_size(MPI_Group group); int smpi_group_compare(MPI_Group group1, MPI_Group group2); -MPI_Comm smpi_comm_new(MPI_Group group); +MPI_Topology smpi_comm_topo(MPI_Comm comm); +MPI_Comm smpi_comm_new(MPI_Group group, MPI_Topology topo); void smpi_comm_destroy(MPI_Comm comm); MPI_Group smpi_comm_group(MPI_Comm comm); int smpi_comm_size(MPI_Comm comm); diff --git a/src/smpi/smpi_comm.c b/src/smpi/smpi_comm.c index 1c36589728..a1b5939a5e 100644 --- a/src/smpi/smpi_comm.c +++ b/src/smpi/smpi_comm.c @@ -12,8 +12,16 @@ XBT_LOG_NEW_DEFAULT_SUBCATEGORY(smpi_comm, smpi, "Logging specific to SMPI (comm)"); + + +/* Support for cartesian topology was added, but there are 2 other types of + * topology, graph et dist graph. In order to support them, we have to add a + * field MPIR_Topo_type, and replace the MPI_Topology field by an union. */ + typedef struct s_smpi_mpi_communicator { MPI_Group group; + MPIR_Topo_type topoType; + MPI_Topology topo; // to be replaced by an union int refcount; } s_smpi_mpi_communicator_t; @@ -37,7 +45,7 @@ static int smpi_compare_rankmap(const void *a, const void *b) return 1; } -MPI_Comm smpi_comm_new(MPI_Group group) +MPI_Comm smpi_comm_new(MPI_Group group, MPI_Topology topo) { MPI_Comm comm; @@ -45,12 +53,14 @@ MPI_Comm smpi_comm_new(MPI_Group group) comm->group = group; smpi_group_use(comm->group); comm->refcount=1; + comm->topo = topo; return comm; } void smpi_comm_destroy(MPI_Comm comm) { smpi_group_unuse(comm->group); + smpi_topo_destroy(comm->topo); // there's no use count on topos smpi_comm_unuse(comm); } @@ -59,6 +69,12 @@ MPI_Group smpi_comm_group(MPI_Comm comm) return comm->group; } +MPI_Topology smpi_comm_topo(MPI_Comm comm) { + if (comm != MPI_COMM_NULL) + return comm->topo; + return NULL; +} + int smpi_comm_size(MPI_Comm comm) { return smpi_group_size(smpi_comm_group(comm)); @@ -157,7 +173,7 @@ MPI_Comm smpi_comm_split(MPI_Comm comm, int color, int key) } } /* otherwise, exit with group_out == NULL */ } - return group_out ? smpi_comm_new(group_out) : MPI_COMM_NULL; + return group_out ? smpi_comm_new(group_out, NULL) : MPI_COMM_NULL; } void smpi_comm_use(MPI_Comm comm){ @@ -169,3 +185,4 @@ void smpi_comm_unuse(MPI_Comm comm){ if(comm->refcount==0) xbt_free(comm); } + diff --git a/src/smpi/smpi_global.c b/src/smpi/smpi_global.c index 46a46c8de7..966ec528bf 100644 --- a/src/smpi/smpi_global.c +++ b/src/smpi/smpi_global.c @@ -290,7 +290,7 @@ MPI_Comm smpi_process_comm_self(void) smpi_process_data_t data = smpi_process_data(); if(data->comm_self==MPI_COMM_NULL){ MPI_Group group = smpi_group_new(1); - data->comm_self = smpi_comm_new(group); + data->comm_self = smpi_comm_new(group, NULL); smpi_group_set_mapping(group, smpi_process_index(), 0); } @@ -360,7 +360,7 @@ void smpi_global_init(void) process_data[i]->sampling = 0; } group = smpi_group_new(process_count); - MPI_COMM_WORLD = smpi_comm_new(group); + MPI_COMM_WORLD = smpi_comm_new(group, NULL); MPI_UNIVERSE_SIZE = smpi_comm_size(MPI_COMM_WORLD); for (i = 0; i < process_count; i++) { smpi_group_set_mapping(group, i, i); diff --git a/src/smpi/smpi_pmpi.c b/src/smpi/smpi_pmpi.c index cfd9b5bab9..cb8c79cae9 100644 --- a/src/smpi/smpi_pmpi.c +++ b/src/smpi/smpi_pmpi.c @@ -75,7 +75,7 @@ int PMPI_Get_library_version (char *version,int *len){ int retval = MPI_SUCCESS; smpi_bench_end(); snprintf(version,MPI_MAX_LIBRARY_VERSION_STRING,"SMPI Version %d.%d. Copyright The Simgrid Team 2007-2013",SIMGRID_VERSION_MAJOR, - SIMGRID_VERSION_MINOR); + SIMGRID_VERSION_MINOR); *len = strlen(version) > MPI_MAX_LIBRARY_VERSION_STRING ? MPI_MAX_LIBRARY_VERSION_STRING : strlen(version); smpi_bench_begin(); return retval; @@ -355,7 +355,7 @@ int PMPI_Group_rank(MPI_Group group, int *rank) } int PMPI_Group_translate_ranks(MPI_Group group1, int n, int *ranks1, - MPI_Group group2, int *ranks2) + MPI_Group group2, int *ranks2) { int retval, i, index; smpi_bench_end(); @@ -394,7 +394,7 @@ int PMPI_Group_compare(MPI_Group group1, MPI_Group group2, int *result) } int PMPI_Group_union(MPI_Group group1, MPI_Group group2, - MPI_Group * newgroup) + MPI_Group * newgroup) { int retval, i, proc1, proc2, size, size2; @@ -434,7 +434,7 @@ int PMPI_Group_union(MPI_Group group1, MPI_Group group2, } int PMPI_Group_intersection(MPI_Group group1, MPI_Group group2, - MPI_Group * newgroup) + MPI_Group * newgroup) { int retval, i, proc1, proc2, size; @@ -523,10 +523,10 @@ int PMPI_Group_incl(MPI_Group group, int n, int *ranks, MPI_Group * newgroup) } else if (n == smpi_group_size(group)) { *newgroup = group; if(group!= smpi_comm_group(MPI_COMM_WORLD) - && group != MPI_GROUP_NULL - && group != smpi_comm_group(MPI_COMM_SELF) - && group != MPI_GROUP_EMPTY) - smpi_group_use(group); + && group != MPI_GROUP_NULL + && group != smpi_comm_group(MPI_COMM_SELF) + && group != MPI_GROUP_EMPTY) + smpi_group_use(group); } else { *newgroup = smpi_group_new(n); for (i = 0; i < n; i++) { @@ -553,10 +553,10 @@ int PMPI_Group_excl(MPI_Group group, int n, int *ranks, MPI_Group * newgroup) if (n == 0) { *newgroup = group; if(group!= smpi_comm_group(MPI_COMM_WORLD) - && group != MPI_GROUP_NULL - && group != smpi_comm_group(MPI_COMM_SELF) - && group != MPI_GROUP_EMPTY) - smpi_group_use(group); + && group != MPI_GROUP_NULL + && group != smpi_comm_group(MPI_COMM_SELF) + && group != MPI_GROUP_EMPTY) + smpi_group_use(group); } else if (n == smpi_group_size(group)) { *newgroup = MPI_GROUP_EMPTY; } else { @@ -588,7 +588,7 @@ int PMPI_Group_excl(MPI_Group group, int n, int *ranks, MPI_Group * newgroup) } int PMPI_Group_range_incl(MPI_Group group, int n, int ranges[][3], - MPI_Group * newgroup) + MPI_Group * newgroup) { int retval, i, j, rank, size, index; @@ -605,16 +605,16 @@ int PMPI_Group_range_incl(MPI_Group group, int n, int ranges[][3], for (i = 0; i < n; i++) { for (rank = ranges[i][0]; /* First */ rank >= 0; /* Last */ - ) { + ) { size++; rank += ranges[i][2]; /* Stride */ if (ranges[i][0] ranges[i][1]) - break; + if(rank > ranges[i][1]) + break; }else{ - if(rank < ranges[i][1]) - break; + if(rank < ranges[i][1]) + break; } } } @@ -646,7 +646,7 @@ int PMPI_Group_range_incl(MPI_Group group, int n, int ranges[][3], } int PMPI_Group_range_excl(MPI_Group group, int n, int ranges[][3], - MPI_Group * newgroup) + MPI_Group * newgroup) { int retval, i, rank, newrank,oldrank, size, index, add; @@ -659,25 +659,25 @@ int PMPI_Group_range_excl(MPI_Group group, int n, int ranges[][3], if (n == 0) { *newgroup = group; if(group!= smpi_comm_group(MPI_COMM_WORLD) - && group != MPI_GROUP_NULL - && group != smpi_comm_group(MPI_COMM_SELF) - && group != MPI_GROUP_EMPTY) - smpi_group_use(group); + && group != MPI_GROUP_NULL + && group != smpi_comm_group(MPI_COMM_SELF) + && group != MPI_GROUP_EMPTY) + smpi_group_use(group); } else { size = smpi_group_size(group); for (i = 0; i < n; i++) { for (rank = ranges[i][0]; /* First */ rank >= 0; /* Last */ - ) { + ) { size--; rank += ranges[i][2]; /* Stride */ if (ranges[i][0] ranges[i][1]) - break; + if(rank > ranges[i][1]) + break; }else{ - if(rank < ranges[i][1]) - break; + if(rank < ranges[i][1]) + break; } } } @@ -692,18 +692,18 @@ int PMPI_Group_range_excl(MPI_Group group, int n, int ranges[][3], for (i = 0; i < n; i++) { for (rank = ranges[i][0];rank >= 0;){ if(rank==oldrank){ - add=0; - break; + add=0; + break; } rank += ranges[i][2]; /* Stride */ if (ranges[i][0] ranges[i][1]) - break; + if(rank > ranges[i][1]) + break; }else{ - if(rank < ranges[i][1]) - break; + if(rank < ranges[i][1]) + break; } } } @@ -786,10 +786,10 @@ int PMPI_Comm_group(MPI_Comm comm, MPI_Group * group) } else { *group = smpi_comm_group(comm); if(*group!= smpi_comm_group(MPI_COMM_WORLD) - && *group != MPI_GROUP_NULL - && *group != smpi_comm_group(MPI_COMM_SELF) - && *group != MPI_GROUP_EMPTY) - smpi_group_use(*group); + && *group != MPI_GROUP_NULL + && *group != smpi_comm_group(MPI_COMM_SELF) + && *group != MPI_GROUP_EMPTY) + smpi_group_use(*group); retval = MPI_SUCCESS; } smpi_bench_begin(); @@ -810,8 +810,8 @@ int PMPI_Comm_compare(MPI_Comm comm1, MPI_Comm comm2, int *result) *result = MPI_IDENT; } else { *result = - smpi_group_compare(smpi_comm_group(comm1), - smpi_comm_group(comm2)); + smpi_group_compare(smpi_comm_group(comm1), + smpi_comm_group(comm2)); if (*result == MPI_IDENT) { *result = MPI_CONGRUENT; } @@ -832,7 +832,7 @@ int PMPI_Comm_dup(MPI_Comm comm, MPI_Comm * newcomm) } else if (newcomm == NULL) { retval = MPI_ERR_ARG; } else { - *newcomm = smpi_comm_new(smpi_comm_group(comm)); + *newcomm = smpi_comm_new(smpi_comm_group(comm), smpi_comm_topo(comm)); retval = MPI_SUCCESS; } smpi_bench_begin(); @@ -855,7 +855,7 @@ int PMPI_Comm_create(MPI_Comm comm, MPI_Group group, MPI_Comm * newcomm) retval = MPI_SUCCESS; }else{ - *newcomm = smpi_comm_new(group); + *newcomm = smpi_comm_new(group, NULL); retval = MPI_SUCCESS; } smpi_bench_begin(); @@ -1190,7 +1190,7 @@ int PMPI_Issend(void* buf, int count, MPI_Datatype datatype, } int PMPI_Recv(void *buf, int count, MPI_Datatype datatype, int src, int tag, - MPI_Comm comm, MPI_Status * status) + MPI_Comm comm, MPI_Status * status) { int retval = 0; @@ -1213,27 +1213,27 @@ int PMPI_Recv(void *buf, int count, MPI_Datatype datatype, int src, int tag, retval = MPI_ERR_TAG; } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int src_traced = smpi_group_index(smpi_comm_group(comm), src); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_RECV; - extra->send_size = count; - extra->src = src_traced; - extra->dst = rank; - extra->datatype1 = encode_datatype(datatype); - TRACE_smpi_ptp_in(rank, src_traced, rank, __FUNCTION__, extra); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int src_traced = smpi_group_index(smpi_comm_group(comm), src); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_RECV; + extra->send_size = count; + extra->src = src_traced; + extra->dst = rank; + extra->datatype1 = encode_datatype(datatype); + TRACE_smpi_ptp_in(rank, src_traced, rank, __FUNCTION__, extra); #endif smpi_mpi_recv(buf, count, datatype, src, tag, comm, status); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - //the src may not have been known at the beginning of the recv (MPI_ANY_SOURCE) - if(status!=MPI_STATUS_IGNORE){ - src_traced = smpi_group_index(smpi_comm_group(comm), status->MPI_SOURCE); - TRACE_smpi_recv(rank, src_traced, rank); - } - TRACE_smpi_ptp_out(rank, src_traced, rank, __FUNCTION__); + //the src may not have been known at the beginning of the recv (MPI_ANY_SOURCE) + if(status!=MPI_STATUS_IGNORE){ + src_traced = smpi_group_index(smpi_comm_group(comm), status->MPI_SOURCE); + TRACE_smpi_recv(rank, src_traced, rank); + } + TRACE_smpi_ptp_out(rank, src_traced, rank, __FUNCTION__); #endif } @@ -1242,7 +1242,7 @@ int PMPI_Recv(void *buf, int count, MPI_Datatype datatype, int src, int tag, } int PMPI_Send(void *buf, int count, MPI_Datatype datatype, int dst, int tag, - MPI_Comm comm) + MPI_Comm comm) { int retval = 0; @@ -1265,23 +1265,23 @@ int PMPI_Send(void *buf, int count, MPI_Datatype datatype, int dst, int tag, } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int dst_traced = smpi_group_index(smpi_comm_group(comm), dst); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_SEND; - extra->send_size = count; - extra->src = rank; - extra->dst = dst_traced; - extra->datatype1 = encode_datatype(datatype); - TRACE_smpi_ptp_in(rank, rank, dst_traced, __FUNCTION__, extra); - TRACE_smpi_send(rank, rank, dst_traced,count*smpi_datatype_size(datatype)); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int dst_traced = smpi_group_index(smpi_comm_group(comm), dst); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_SEND; + extra->send_size = count; + extra->src = rank; + extra->dst = dst_traced; + extra->datatype1 = encode_datatype(datatype); + TRACE_smpi_ptp_in(rank, rank, dst_traced, __FUNCTION__, extra); + TRACE_smpi_send(rank, rank, dst_traced,count*smpi_datatype_size(datatype)); #endif smpi_mpi_send(buf, count, datatype, dst, tag, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_ptp_out(rank, rank, dst_traced, __FUNCTION__); + TRACE_smpi_ptp_out(rank, rank, dst_traced, __FUNCTION__); #endif } @@ -1294,52 +1294,52 @@ int PMPI_Send(void *buf, int count, MPI_Datatype datatype, int dst, int tag, int PMPI_Ssend(void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm) { int retval = 0; - smpi_bench_end(); - - if (comm == MPI_COMM_NULL) { - retval = MPI_ERR_COMM; - } else if (dst == MPI_PROC_NULL) { - retval = MPI_SUCCESS; - } else if (dst >= smpi_group_size(smpi_comm_group(comm)) || dst <0){ - retval = MPI_ERR_RANK; - } else if (count < 0) { - retval = MPI_ERR_COUNT; - } else if (buf==NULL && count > 0) { - retval = MPI_ERR_COUNT; - } else if (datatype == MPI_DATATYPE_NULL){ - retval = MPI_ERR_TYPE; - } else if(tag<0 && tag != MPI_ANY_TAG){ - retval = MPI_ERR_TAG; - } else { - - #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int dst_traced = smpi_group_index(smpi_comm_group(comm), dst); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_SSEND; - extra->send_size = count; - extra->src = rank; - extra->dst = dst_traced; - extra->datatype1 = encode_datatype(datatype); - TRACE_smpi_ptp_in(rank, rank, dst_traced, __FUNCTION__, extra); TRACE_smpi_send(rank, rank, dst_traced,count*smpi_datatype_size(datatype)); - #endif - - smpi_mpi_ssend(buf, count, datatype, dst, tag, comm); - retval = MPI_SUCCESS; - - #ifdef HAVE_TRACING - TRACE_smpi_ptp_out(rank, rank, dst_traced, __FUNCTION__); - #endif - } - - smpi_bench_begin(); - return retval;} + smpi_bench_end(); + + if (comm == MPI_COMM_NULL) { + retval = MPI_ERR_COMM; + } else if (dst == MPI_PROC_NULL) { + retval = MPI_SUCCESS; + } else if (dst >= smpi_group_size(smpi_comm_group(comm)) || dst <0){ + retval = MPI_ERR_RANK; + } else if (count < 0) { + retval = MPI_ERR_COUNT; + } else if (buf==NULL && count > 0) { + retval = MPI_ERR_COUNT; + } else if (datatype == MPI_DATATYPE_NULL){ + retval = MPI_ERR_TYPE; + } else if(tag<0 && tag != MPI_ANY_TAG){ + retval = MPI_ERR_TAG; + } else { + +#ifdef HAVE_TRACING + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int dst_traced = smpi_group_index(smpi_comm_group(comm), dst); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_SSEND; + extra->send_size = count; + extra->src = rank; + extra->dst = dst_traced; + extra->datatype1 = encode_datatype(datatype); + TRACE_smpi_ptp_in(rank, rank, dst_traced, __FUNCTION__, extra); TRACE_smpi_send(rank, rank, dst_traced,count*smpi_datatype_size(datatype)); +#endif + + smpi_mpi_ssend(buf, count, datatype, dst, tag, comm); + retval = MPI_SUCCESS; + +#ifdef HAVE_TRACING + TRACE_smpi_ptp_out(rank, rank, dst_traced, __FUNCTION__); +#endif + } + + smpi_bench_begin(); + return retval;} int PMPI_Sendrecv(void *sendbuf, int sendcount, MPI_Datatype sendtype, - int dst, int sendtag, void *recvbuf, int recvcount, - MPI_Datatype recvtype, int src, int recvtag, - MPI_Comm comm, MPI_Status * status) + int dst, int sendtag, void *recvbuf, int recvcount, + MPI_Datatype recvtype, int src, int recvtag, + MPI_Comm comm, MPI_Status * status) { int retval = 0; @@ -1351,14 +1351,14 @@ int PMPI_Sendrecv(void *sendbuf, int sendcount, MPI_Datatype sendtype, || recvtype == MPI_DATATYPE_NULL) { retval = MPI_ERR_TYPE; } else if (src == MPI_PROC_NULL || dst == MPI_PROC_NULL) { - smpi_empty_status(status); - status->MPI_SOURCE = MPI_PROC_NULL; - retval = MPI_SUCCESS; + smpi_empty_status(status); + status->MPI_SOURCE = MPI_PROC_NULL; + retval = MPI_SUCCESS; }else if (dst >= smpi_group_size(smpi_comm_group(comm)) || dst <0 || - (src!=MPI_ANY_SOURCE && (src >= smpi_group_size(smpi_comm_group(comm)) || src <0))){ + (src!=MPI_ANY_SOURCE && (src >= smpi_group_size(smpi_comm_group(comm)) || src <0))){ retval = MPI_ERR_RANK; } else if (sendcount < 0 || recvcount<0) { - retval = MPI_ERR_COUNT; + retval = MPI_ERR_COUNT; } else if ((sendbuf==NULL && sendcount > 0)||(recvbuf==NULL && recvcount>0)) { retval = MPI_ERR_COUNT; } else if((sendtag<0 && sendtag != MPI_ANY_TAG)||(recvtag<0 && recvtag != MPI_ANY_TAG)){ @@ -1366,20 +1366,20 @@ int PMPI_Sendrecv(void *sendbuf, int sendcount, MPI_Datatype sendtype, } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int dst_traced = smpi_group_index(smpi_comm_group(comm), dst); - int src_traced = smpi_group_index(smpi_comm_group(comm), src); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_SENDRECV; - extra->send_size = sendcount; - extra->recv_size = recvcount; - extra->src = src_traced; - extra->dst = dst_traced; - extra->datatype1 = encode_datatype(sendtype); - extra->datatype2 = encode_datatype(recvtype); - - TRACE_smpi_ptp_in(rank, src_traced, dst_traced, __FUNCTION__, extra); - TRACE_smpi_send(rank, rank, dst_traced,sendcount*smpi_datatype_size(sendtype)); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int dst_traced = smpi_group_index(smpi_comm_group(comm), dst); + int src_traced = smpi_group_index(smpi_comm_group(comm), src); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_SENDRECV; + extra->send_size = sendcount; + extra->recv_size = recvcount; + extra->src = src_traced; + extra->dst = dst_traced; + extra->datatype1 = encode_datatype(sendtype); + extra->datatype2 = encode_datatype(recvtype); + + TRACE_smpi_ptp_in(rank, src_traced, dst_traced, __FUNCTION__, extra); + TRACE_smpi_send(rank, rank, dst_traced,sendcount*smpi_datatype_size(sendtype)); #endif @@ -1388,8 +1388,8 @@ int PMPI_Sendrecv(void *sendbuf, int sendcount, MPI_Datatype sendtype, retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_ptp_out(rank, src_traced, dst_traced, __FUNCTION__); - TRACE_smpi_recv(rank, src_traced, rank); + TRACE_smpi_ptp_out(rank, src_traced, dst_traced, __FUNCTION__); + TRACE_smpi_recv(rank, src_traced, rank); #endif } @@ -1399,24 +1399,24 @@ int PMPI_Sendrecv(void *sendbuf, int sendcount, MPI_Datatype sendtype, } int PMPI_Sendrecv_replace(void *buf, int count, MPI_Datatype datatype, - int dst, int sendtag, int src, int recvtag, - MPI_Comm comm, MPI_Status * status) + int dst, int sendtag, int src, int recvtag, + MPI_Comm comm, MPI_Status * status) { //TODO: suboptimal implementation void *recvbuf; int retval = 0; if (datatype == MPI_DATATYPE_NULL) { - retval = MPI_ERR_TYPE; + retval = MPI_ERR_TYPE; } else if (count < 0) { - retval = MPI_ERR_COUNT; + retval = MPI_ERR_COUNT; } else { int size = smpi_datatype_get_extent(datatype) * count; recvbuf = xbt_new0(char, size); retval = - MPI_Sendrecv(buf, count, datatype, dst, sendtag, recvbuf, count, - datatype, src, recvtag, comm, status); + MPI_Sendrecv(buf, count, datatype, dst, sendtag, recvbuf, count, + datatype, src, recvtag, comm, status); if(retval==MPI_SUCCESS){ - smpi_datatype_copy(recvbuf, count, datatype, buf, count, datatype); + smpi_datatype_copy(recvbuf, count, datatype, buf, count, datatype); } xbt_free(recvbuf); @@ -1444,7 +1444,7 @@ int PMPI_Test(MPI_Request * request, int *flag, MPI_Status * status) } int PMPI_Testany(int count, MPI_Request requests[], int *index, int *flag, - MPI_Status * status) + MPI_Status * status) { int retval = 0; @@ -1607,8 +1607,8 @@ int PMPI_Waitany(int count, MPI_Request requests[], int *index, MPI_Status * sta if (is_wait_for_receive) { if(srcs[*index]==MPI_ANY_SOURCE) src_traced = (status!=MPI_STATUSES_IGNORE) ? - smpi_group_rank(smpi_comm_group(comms[*index]), status->MPI_SOURCE) : - srcs[*index]; + smpi_group_rank(smpi_comm_group(comms[*index]), status->MPI_SOURCE) : + srcs[*index]; TRACE_smpi_recv(rank_traced, src_traced, dst_traced); } TRACE_smpi_ptp_out(rank_traced, src_traced, dst_traced, __FUNCTION__); @@ -1659,16 +1659,16 @@ int PMPI_Waitall(int count, MPI_Request requests[], MPI_Status status[]) #ifdef HAVE_TRACING for (i = 0; i < count; i++) { if(valid[i]){ - //int src_traced = srcs[*index]; - //the src may not have been known at the beginning of the recv (MPI_ANY_SOURCE) + //int src_traced = srcs[*index]; + //the src may not have been known at the beginning of the recv (MPI_ANY_SOURCE) int src_traced = srcs[i]; int dst_traced = dsts[i]; int is_wait_for_receive = recvs[i]; if (is_wait_for_receive) { if(src_traced==MPI_ANY_SOURCE) - src_traced = (status!=MPI_STATUSES_IGNORE) ? - smpi_group_rank(smpi_comm_group(comms[i]), status[i].MPI_SOURCE) : - srcs[i]; + src_traced = (status!=MPI_STATUSES_IGNORE) ? + smpi_group_rank(smpi_comm_group(comms[i]), status[i].MPI_SOURCE) : + srcs[i]; TRACE_smpi_recv(rank_traced, src_traced, dst_traced); } } @@ -1686,7 +1686,7 @@ int PMPI_Waitall(int count, MPI_Request requests[], MPI_Status status[]) } int PMPI_Waitsome(int incount, MPI_Request requests[], int *outcount, - int *indices, MPI_Status status[]) + int *indices, MPI_Status status[]) { int retval = 0; @@ -1702,19 +1702,19 @@ int PMPI_Waitsome(int incount, MPI_Request requests[], int *outcount, } int PMPI_Testsome(int incount, MPI_Request requests[], int* outcount, - int* indices, MPI_Status status[]) + int* indices, MPI_Status status[]) { int retval = 0; - smpi_bench_end(); - if (outcount == NULL) { - retval = MPI_ERR_ARG; - } else { - *outcount = smpi_mpi_testsome(incount, requests, indices, status); - retval = MPI_SUCCESS; - } - smpi_bench_begin(); - return retval; + smpi_bench_end(); + if (outcount == NULL) { + retval = MPI_ERR_ARG; + } else { + *outcount = smpi_mpi_testsome(incount, requests, indices, status); + retval = MPI_SUCCESS; + } + smpi_bench_begin(); + return retval; } @@ -1728,21 +1728,21 @@ int PMPI_Bcast(void *buf, int count, MPI_Datatype datatype, int root, MPI_Comm c retval = MPI_ERR_COMM; } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int root_traced = smpi_group_index(smpi_comm_group(comm), root); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int root_traced = smpi_group_index(smpi_comm_group(comm), root); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_BCAST; - extra->send_size = count; - extra->root = root_traced; - extra->datatype1 = encode_datatype(datatype); - TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__, extra); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_BCAST; + extra->send_size = count; + extra->root = root_traced; + extra->datatype1 = encode_datatype(datatype); + TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__, extra); #endif mpi_coll_bcast_fun(buf, count, datatype, root, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); + TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); #endif } @@ -1760,15 +1760,15 @@ int PMPI_Barrier(MPI_Comm comm) retval = MPI_ERR_COMM; } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_BARRIER; - TRACE_smpi_collective_in(rank, -1, __FUNCTION__, extra); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_BARRIER; + TRACE_smpi_collective_in(rank, -1, __FUNCTION__, extra); #endif mpi_coll_barrier_fun(comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); + TRACE_smpi_collective_out(rank, -1, __FUNCTION__); #endif } @@ -1777,8 +1777,8 @@ int PMPI_Barrier(MPI_Comm comm) } int PMPI_Gather(void *sendbuf, int sendcount, MPI_Datatype sendtype, - void *recvbuf, int recvcount, MPI_Datatype recvtype, - int root, MPI_Comm comm) + void *recvbuf, int recvcount, MPI_Datatype recvtype, + int root, MPI_Comm comm) { int retval = 0; @@ -1787,10 +1787,10 @@ int PMPI_Gather(void *sendbuf, int sendcount, MPI_Datatype sendtype, if (comm == MPI_COMM_NULL) { retval = MPI_ERR_COMM; } else if ((( sendbuf != MPI_IN_PLACE) && (sendtype == MPI_DATATYPE_NULL)) || - ((smpi_comm_rank(comm) == root) && (recvtype == MPI_DATATYPE_NULL))){ + ((smpi_comm_rank(comm) == root) && (recvtype == MPI_DATATYPE_NULL))){ retval = MPI_ERR_TYPE; } else if ((( sendbuf != MPI_IN_PLACE) && (sendcount <0)) || - ((smpi_comm_rank(comm) == root) && (recvcount <0))){ + ((smpi_comm_rank(comm) == root) && (recvcount <0))){ retval = MPI_ERR_COUNT; } else { @@ -1802,25 +1802,25 @@ int PMPI_Gather(void *sendbuf, int sendcount, MPI_Datatype sendtype, sendtmptype=recvtype; } #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int root_traced = smpi_group_index(smpi_comm_group(comm), root); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_GATHER; - extra->send_size = sendtmpcount; - extra->recv_size = recvcount; - extra->root = root_traced; - extra->datatype1 = encode_datatype(sendtmptype); - extra->datatype2 = encode_datatype(recvtype); - - TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__, extra); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int root_traced = smpi_group_index(smpi_comm_group(comm), root); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_GATHER; + extra->send_size = sendtmpcount; + extra->recv_size = recvcount; + extra->root = root_traced; + extra->datatype1 = encode_datatype(sendtmptype); + extra->datatype2 = encode_datatype(recvtype); + + TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__, extra); #endif mpi_coll_gather_fun(sendtmpbuf, sendtmpcount, sendtmptype, recvbuf, recvcount, - recvtype, root, comm); + recvtype, root, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); + TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); #endif } @@ -1829,8 +1829,8 @@ int PMPI_Gather(void *sendbuf, int sendcount, MPI_Datatype sendtype, } int PMPI_Gatherv(void *sendbuf, int sendcount, MPI_Datatype sendtype, - void *recvbuf, int *recvcounts, int *displs, - MPI_Datatype recvtype, int root, MPI_Comm comm) + void *recvbuf, int *recvcounts, int *displs, + MPI_Datatype recvtype, int root, MPI_Comm comm) { int retval = 0; @@ -1839,7 +1839,7 @@ int PMPI_Gatherv(void *sendbuf, int sendcount, MPI_Datatype sendtype, if (comm == MPI_COMM_NULL) { retval = MPI_ERR_COMM; } else if ((( sendbuf != MPI_IN_PLACE) && (sendtype == MPI_DATATYPE_NULL)) || - ((smpi_comm_rank(comm) == root) && (recvtype == MPI_DATATYPE_NULL))){ + ((smpi_comm_rank(comm) == root) && (recvtype == MPI_DATATYPE_NULL))){ retval = MPI_ERR_TYPE; } else if (( sendbuf != MPI_IN_PLACE) && (sendcount <0)){ retval = MPI_ERR_COUNT; @@ -1855,28 +1855,28 @@ int PMPI_Gatherv(void *sendbuf, int sendcount, MPI_Datatype sendtype, } #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int root_traced = smpi_group_index(smpi_comm_group(comm), root); - int i=0; - int size = smpi_comm_size(comm); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_GATHERV; - extra->send_size = sendtmpcount; - extra->recvcounts= xbt_malloc(size*sizeof(int)); - for(i=0; i< size; i++)//copy data to avoid bad free - extra->recvcounts[i] = recvcounts[i]; - extra->num_processes = size; - extra->root = root_traced; - extra->datatype1 = encode_datatype(sendtmptype); - extra->datatype2 = encode_datatype(recvtype); - - TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__,extra); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int root_traced = smpi_group_index(smpi_comm_group(comm), root); + int i=0; + int size = smpi_comm_size(comm); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_GATHERV; + extra->send_size = sendtmpcount; + extra->recvcounts= xbt_malloc(size*sizeof(int)); + for(i=0; i< size; i++)//copy data to avoid bad free + extra->recvcounts[i] = recvcounts[i]; + extra->num_processes = size; + extra->root = root_traced; + extra->datatype1 = encode_datatype(sendtmptype); + extra->datatype2 = encode_datatype(recvtype); + + TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__,extra); #endif smpi_mpi_gatherv(sendtmpbuf, sendtmpcount, sendtmptype, recvbuf, recvcounts, displs, recvtype, root, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); + TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); #endif } @@ -1885,8 +1885,8 @@ int PMPI_Gatherv(void *sendbuf, int sendcount, MPI_Datatype sendtype, } int PMPI_Allgather(void *sendbuf, int sendcount, MPI_Datatype sendtype, - void *recvbuf, int recvcount, MPI_Datatype recvtype, - MPI_Comm comm) + void *recvbuf, int recvcount, MPI_Datatype recvtype, + MPI_Comm comm) { int retval = 0; @@ -1895,10 +1895,10 @@ int PMPI_Allgather(void *sendbuf, int sendcount, MPI_Datatype sendtype, if (comm == MPI_COMM_NULL) { retval = MPI_ERR_COMM; } else if ((( sendbuf != MPI_IN_PLACE) && (sendtype == MPI_DATATYPE_NULL)) || - (recvtype == MPI_DATATYPE_NULL)){ + (recvtype == MPI_DATATYPE_NULL)){ retval = MPI_ERR_TYPE; } else if ((( sendbuf != MPI_IN_PLACE) && (sendcount <0)) || - (recvcount <0)){ + (recvcount <0)){ retval = MPI_ERR_COUNT; } else { if(sendbuf == MPI_IN_PLACE) { @@ -1907,22 +1907,22 @@ int PMPI_Allgather(void *sendbuf, int sendcount, MPI_Datatype sendtype, sendtype=recvtype; } #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_ALLGATHER; - extra->send_size = sendcount; - extra->recv_size = recvcount; - extra->datatype1 = encode_datatype(sendtype); - extra->datatype2 = encode_datatype(recvtype); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_ALLGATHER; + extra->send_size = sendcount; + extra->recv_size = recvcount; + extra->datatype1 = encode_datatype(sendtype); + extra->datatype2 = encode_datatype(recvtype); - TRACE_smpi_collective_in(rank, -1, __FUNCTION__, extra); + TRACE_smpi_collective_in(rank, -1, __FUNCTION__, extra); #endif mpi_coll_allgather_fun(sendbuf, sendcount, sendtype, recvbuf, recvcount, recvtype, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); + TRACE_smpi_collective_out(rank, -1, __FUNCTION__); #endif } smpi_bench_begin(); @@ -1930,8 +1930,8 @@ int PMPI_Allgather(void *sendbuf, int sendcount, MPI_Datatype sendtype, } int PMPI_Allgatherv(void *sendbuf, int sendcount, MPI_Datatype sendtype, - void *recvbuf, int *recvcounts, int *displs, - MPI_Datatype recvtype, MPI_Comm comm) + void *recvbuf, int *recvcounts, int *displs, + MPI_Datatype recvtype, MPI_Comm comm) { int retval = 0; @@ -1940,7 +1940,7 @@ int PMPI_Allgatherv(void *sendbuf, int sendcount, MPI_Datatype sendtype, if (comm == MPI_COMM_NULL) { retval = MPI_ERR_COMM; } else if ((( sendbuf != MPI_IN_PLACE) && (sendtype == MPI_DATATYPE_NULL)) || - (recvtype == MPI_DATATYPE_NULL)){ + (recvtype == MPI_DATATYPE_NULL)){ retval = MPI_ERR_TYPE; } else if (( sendbuf != MPI_IN_PLACE) && (sendcount <0)){ retval = MPI_ERR_COUNT; @@ -1954,26 +1954,26 @@ int PMPI_Allgatherv(void *sendbuf, int sendcount, MPI_Datatype sendtype, sendtype=recvtype; } #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int i=0; - int size = smpi_comm_size(comm); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_ALLGATHERV; - extra->send_size = sendcount; - extra->recvcounts= xbt_malloc(size*sizeof(int)); - for(i=0; i< size; i++)//copy data to avoid bad free - extra->recvcounts[i] = recvcounts[i]; - extra->num_processes = size; - extra->datatype1 = encode_datatype(sendtype); - extra->datatype2 = encode_datatype(recvtype); - - TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int i=0; + int size = smpi_comm_size(comm); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_ALLGATHERV; + extra->send_size = sendcount; + extra->recvcounts= xbt_malloc(size*sizeof(int)); + for(i=0; i< size; i++)//copy data to avoid bad free + extra->recvcounts[i] = recvcounts[i]; + extra->num_processes = size; + extra->datatype1 = encode_datatype(sendtype); + extra->datatype2 = encode_datatype(recvtype); + + TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); #endif mpi_coll_allgatherv_fun(sendbuf, sendcount, sendtype, recvbuf, recvcounts, - displs, recvtype, comm); + displs, recvtype, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); + TRACE_smpi_collective_out(rank, -1, __FUNCTION__); #endif } @@ -1982,8 +1982,8 @@ int PMPI_Allgatherv(void *sendbuf, int sendcount, MPI_Datatype sendtype, } int PMPI_Scatter(void *sendbuf, int sendcount, MPI_Datatype sendtype, - void *recvbuf, int recvcount, MPI_Datatype recvtype, - int root, MPI_Comm comm) + void *recvbuf, int recvcount, MPI_Datatype recvtype, + int root, MPI_Comm comm) { int retval = 0; @@ -1997,27 +1997,27 @@ int PMPI_Scatter(void *sendbuf, int sendcount, MPI_Datatype sendtype, } else { if (recvbuf == MPI_IN_PLACE) { - recvtype=sendtype; - recvcount=sendcount; + recvtype=sendtype; + recvcount=sendcount; } #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int root_traced = smpi_group_index(smpi_comm_group(comm), root); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_SCATTER; - extra->send_size = sendcount; - extra->recv_size= recvcount; - extra->root = root_traced; - extra->datatype1 = encode_datatype(sendtype); - extra->datatype2 = encode_datatype(recvtype); - - TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__,extra); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int root_traced = smpi_group_index(smpi_comm_group(comm), root); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_SCATTER; + extra->send_size = sendcount; + extra->recv_size= recvcount; + extra->root = root_traced; + extra->datatype1 = encode_datatype(sendtype); + extra->datatype2 = encode_datatype(recvtype); + + TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__,extra); #endif mpi_coll_scatter_fun(sendbuf, sendcount, sendtype, recvbuf, recvcount, - recvtype, root, comm); + recvtype, root, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); + TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); #endif } @@ -2026,8 +2026,8 @@ int PMPI_Scatter(void *sendbuf, int sendcount, MPI_Datatype sendtype, } int PMPI_Scatterv(void *sendbuf, int *sendcounts, int *displs, - MPI_Datatype sendtype, void *recvbuf, int recvcount, - MPI_Datatype recvtype, int root, MPI_Comm comm) + MPI_Datatype sendtype, void *recvbuf, int recvcount, + MPI_Datatype recvtype, int root, MPI_Comm comm) { int retval = 0; @@ -2042,33 +2042,33 @@ int PMPI_Scatterv(void *sendbuf, int *sendcounts, int *displs, retval = MPI_ERR_TYPE; } else { if (recvbuf == MPI_IN_PLACE) { - recvtype=sendtype; - recvcount=sendcounts[smpi_comm_rank(comm)]; + recvtype=sendtype; + recvcount=sendcounts[smpi_comm_rank(comm)]; } #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int root_traced = smpi_group_index(smpi_comm_group(comm), root); - int i=0; - int size = smpi_comm_size(comm); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_SCATTERV; - extra->recv_size = recvcount; - extra->sendcounts= xbt_malloc(size*sizeof(int)); - for(i=0; i< size; i++)//copy data to avoid bad free - extra->sendcounts[i] = sendcounts[i]; - extra->num_processes = size; - extra->root = root_traced; - extra->datatype1 = encode_datatype(sendtype); - extra->datatype2 = encode_datatype(recvtype); - - TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__,extra); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int root_traced = smpi_group_index(smpi_comm_group(comm), root); + int i=0; + int size = smpi_comm_size(comm); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_SCATTERV; + extra->recv_size = recvcount; + extra->sendcounts= xbt_malloc(size*sizeof(int)); + for(i=0; i< size; i++)//copy data to avoid bad free + extra->sendcounts[i] = sendcounts[i]; + extra->num_processes = size; + extra->root = root_traced; + extra->datatype1 = encode_datatype(sendtype); + extra->datatype2 = encode_datatype(recvtype); + + TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__,extra); #endif smpi_mpi_scatterv(sendbuf, sendcounts, displs, sendtype, recvbuf, recvcount, recvtype, root, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); + TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); #endif } @@ -2077,7 +2077,7 @@ int PMPI_Scatterv(void *sendbuf, int *sendcounts, int *displs, } int PMPI_Reduce(void *sendbuf, void *recvbuf, int count, - MPI_Datatype datatype, MPI_Op op, int root, MPI_Comm comm) + MPI_Datatype datatype, MPI_Op op, int root, MPI_Comm comm) { int retval = 0; @@ -2089,21 +2089,21 @@ int PMPI_Reduce(void *sendbuf, void *recvbuf, int count, retval = MPI_ERR_ARG; } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int root_traced = smpi_group_index(smpi_comm_group(comm), root); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_REDUCE; - extra->send_size = count; - extra->datatype1 = encode_datatype(datatype); - extra->root = root_traced; + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int root_traced = smpi_group_index(smpi_comm_group(comm), root); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_REDUCE; + extra->send_size = count; + extra->datatype1 = encode_datatype(datatype); + extra->root = root_traced; - TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__,extra); + TRACE_smpi_collective_in(rank, root_traced, __FUNCTION__,extra); #endif mpi_coll_reduce_fun(sendbuf, recvbuf, count, datatype, op, root, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); + TRACE_smpi_collective_out(rank, root_traced, __FUNCTION__); #endif } @@ -2112,22 +2112,22 @@ int PMPI_Reduce(void *sendbuf, void *recvbuf, int count, } int PMPI_Reduce_local(void *inbuf, void *inoutbuf, int count, - MPI_Datatype datatype, MPI_Op op){ + MPI_Datatype datatype, MPI_Op op){ int retval = 0; - smpi_bench_end(); - if (datatype == MPI_DATATYPE_NULL || op == MPI_OP_NULL) { - retval = MPI_ERR_ARG; - } else { - smpi_op_apply(op, inbuf, inoutbuf, &count, &datatype); - retval=MPI_SUCCESS; - } - smpi_bench_begin(); - return retval; + smpi_bench_end(); + if (datatype == MPI_DATATYPE_NULL || op == MPI_OP_NULL) { + retval = MPI_ERR_ARG; + } else { + smpi_op_apply(op, inbuf, inoutbuf, &count, &datatype); + retval=MPI_SUCCESS; + } + smpi_bench_begin(); + return retval; } int PMPI_Allreduce(void *sendbuf, void *recvbuf, int count, - MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) + MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) { int retval = 0; @@ -2147,13 +2147,13 @@ int PMPI_Allreduce(void *sendbuf, void *recvbuf, int count, smpi_datatype_copy(recvbuf, count, datatype,sendtmpbuf, count, datatype); } #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_ALLREDUCE; - extra->send_size = count; - extra->datatype1 = encode_datatype(datatype); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_ALLREDUCE; + extra->send_size = count; + extra->datatype1 = encode_datatype(datatype); - TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); + TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); #endif mpi_coll_allreduce_fun(sendtmpbuf, recvbuf, count, datatype, op, comm); @@ -2163,7 +2163,7 @@ int PMPI_Allreduce(void *sendbuf, void *recvbuf, int count, retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); + TRACE_smpi_collective_out(rank, -1, __FUNCTION__); #endif } @@ -2172,7 +2172,7 @@ int PMPI_Allreduce(void *sendbuf, void *recvbuf, int count, } int PMPI_Scan(void *sendbuf, void *recvbuf, int count, - MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) + MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) { int retval = 0; @@ -2186,18 +2186,18 @@ int PMPI_Scan(void *sendbuf, void *recvbuf, int count, retval = MPI_ERR_OP; } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_SCAN; - extra->send_size = count; - extra->datatype1 = encode_datatype(datatype); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_SCAN; + extra->send_size = count; + extra->datatype1 = encode_datatype(datatype); - TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); + TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); #endif smpi_mpi_scan(sendbuf, recvbuf, count, datatype, op, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); + TRACE_smpi_collective_out(rank, -1, __FUNCTION__); #endif } @@ -2219,18 +2219,18 @@ int PMPI_Exscan(void *sendbuf, void *recvbuf, int count, MPI_Datatype datatype, retval = MPI_ERR_OP; } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_EXSCAN; - extra->send_size = count; - extra->datatype1 = encode_datatype(datatype); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_EXSCAN; + extra->send_size = count; + extra->datatype1 = encode_datatype(datatype); - TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); + TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); #endif smpi_mpi_exscan(sendbuf, recvbuf, count, datatype, op, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); + TRACE_smpi_collective_out(rank, -1, __FUNCTION__); #endif } @@ -2239,7 +2239,7 @@ int PMPI_Exscan(void *sendbuf, void *recvbuf, int count, MPI_Datatype datatype, } int PMPI_Reduce_scatter(void *sendbuf, void *recvbuf, int *recvcounts, - MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) + MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) { int retval = 0; smpi_bench_end(); @@ -2254,19 +2254,19 @@ int PMPI_Reduce_scatter(void *sendbuf, void *recvbuf, int *recvcounts, retval = MPI_ERR_ARG; } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int i=0; - int size = smpi_comm_size(comm); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_REDUCE_SCATTER; - extra->send_size = 0; - extra->recvcounts= xbt_malloc(size*sizeof(int)); - for(i=0; i< size; i++)//copy data to avoid bad free - extra->recvcounts[i] = recvcounts[i]; - extra->num_processes = size; - extra->datatype1 = encode_datatype(datatype); - - TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int i=0; + int size = smpi_comm_size(comm); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_REDUCE_SCATTER; + extra->send_size = 0; + extra->recvcounts= xbt_malloc(size*sizeof(int)); + for(i=0; i< size; i++)//copy data to avoid bad free + extra->recvcounts[i] = recvcounts[i]; + extra->num_processes = size; + extra->datatype1 = encode_datatype(datatype); + + TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); #endif void* sendtmpbuf=sendbuf; if(sendbuf==MPI_IN_PLACE){ @@ -2274,10 +2274,10 @@ int PMPI_Reduce_scatter(void *sendbuf, void *recvbuf, int *recvcounts, } mpi_coll_reduce_scatter_fun(sendtmpbuf, recvbuf, recvcounts, - datatype, op, comm); + datatype, op, comm); retval = MPI_SUCCESS; #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); + TRACE_smpi_collective_out(rank, -1, __FUNCTION__); #endif } @@ -2286,7 +2286,7 @@ int PMPI_Reduce_scatter(void *sendbuf, void *recvbuf, int *recvcounts, } int PMPI_Reduce_scatter_block(void *sendbuf, void *recvbuf, int recvcount, - MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) + MPI_Datatype datatype, MPI_Op op, MPI_Comm comm) { int retval,i; smpi_bench_end(); @@ -2303,26 +2303,26 @@ int PMPI_Reduce_scatter_block(void *sendbuf, void *recvbuf, int recvcount, int count=smpi_comm_size(comm); #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_REDUCE_SCATTER; - extra->send_size = 0; - extra->recvcounts= xbt_malloc(count*sizeof(int)); - for(i=0; i< count; i++)//copy data to avoid bad free - extra->recvcounts[i] = recvcount; - extra->num_processes = count; - extra->datatype1 = encode_datatype(datatype); - - TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_REDUCE_SCATTER; + extra->send_size = 0; + extra->recvcounts= xbt_malloc(count*sizeof(int)); + for(i=0; i< count; i++)//copy data to avoid bad free + extra->recvcounts[i] = recvcount; + extra->num_processes = count; + extra->datatype1 = encode_datatype(datatype); + + TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); #endif int* recvcounts=(int*)xbt_malloc(count); for (i=0; itype = TRACING_ALLTOALL; - extra->send_size = sendcount; - extra->recv_size = recvcount; - extra->datatype1 = encode_datatype(sendtype); - extra->datatype2 = encode_datatype(recvtype); + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_ALLTOALL; + extra->send_size = sendcount; + extra->recv_size = recvcount; + extra->datatype1 = encode_datatype(sendtype); + extra->datatype2 = encode_datatype(recvtype); - TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); + TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); #endif retval = mpi_coll_alltoall_fun(sendbuf, sendcount, sendtype, recvbuf, recvcount, recvtype, comm); #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); + TRACE_smpi_collective_out(rank, -1, __FUNCTION__); #endif } @@ -2366,8 +2366,8 @@ int PMPI_Alltoall(void *sendbuf, int sendcount, MPI_Datatype sendtype, } int PMPI_Alltoallv(void *sendbuf, int *sendcounts, int *senddisps, - MPI_Datatype sendtype, void *recvbuf, int *recvcounts, - int *recvdisps, MPI_Datatype recvtype, MPI_Comm comm) + MPI_Datatype sendtype, void *recvbuf, int *recvcounts, + int *recvdisps, MPI_Datatype recvtype, MPI_Comm comm) { int retval = 0; @@ -2383,36 +2383,36 @@ int PMPI_Alltoallv(void *sendbuf, int *sendcounts, int *senddisps, retval = MPI_ERR_ARG; } else { #ifdef HAVE_TRACING - int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; - int i=0; - int size = smpi_comm_size(comm); - instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); - extra->type = TRACING_ALLTOALLV; - extra->send_size = 0; - extra->recv_size = 0; - extra->recvcounts= xbt_malloc(size*sizeof(int)); - extra->sendcounts= xbt_malloc(size*sizeof(int)); - - for(i=0; i< size; i++){//copy data to avoid bad free - extra->send_size += sendcounts[i]; - extra->recv_size += recvcounts[i]; - - extra->sendcounts[i] = sendcounts[i]; - extra->recvcounts[i] = recvcounts[i]; - } - extra->num_processes = size; + int rank = comm != MPI_COMM_NULL ? smpi_process_index() : -1; + int i=0; + int size = smpi_comm_size(comm); + instr_extra_data extra = xbt_new0(s_instr_extra_data_t,1); + extra->type = TRACING_ALLTOALLV; + extra->send_size = 0; + extra->recv_size = 0; + extra->recvcounts= xbt_malloc(size*sizeof(int)); + extra->sendcounts= xbt_malloc(size*sizeof(int)); + + for(i=0; i< size; i++){//copy data to avoid bad free + extra->send_size += sendcounts[i]; + extra->recv_size += recvcounts[i]; + + extra->sendcounts[i] = sendcounts[i]; + extra->recvcounts[i] = recvcounts[i]; + } + extra->num_processes = size; - extra->datatype1 = encode_datatype(sendtype); - extra->datatype2 = encode_datatype(recvtype); + extra->datatype1 = encode_datatype(sendtype); + extra->datatype2 = encode_datatype(recvtype); - TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); + TRACE_smpi_collective_in(rank, -1, __FUNCTION__,extra); #endif retval = - mpi_coll_alltoallv_fun(sendbuf, sendcounts, senddisps, sendtype, - recvbuf, recvcounts, recvdisps, recvtype, - comm); + mpi_coll_alltoallv_fun(sendbuf, sendcounts, senddisps, sendtype, + recvbuf, recvcounts, recvdisps, recvtype, + comm); #ifdef HAVE_TRACING - TRACE_smpi_collective_out(rank, -1, __FUNCTION__); + TRACE_smpi_collective_out(rank, -1, __FUNCTION__); #endif } @@ -2431,8 +2431,8 @@ int PMPI_Get_processor_name(char *name, int *resultlen) strlen(SIMIX_host_get_name(SIMIX_host_self())) +1 : MPI_MAX_PROCESSOR_NAME - 1 ); *resultlen = - strlen(name) > - MPI_MAX_PROCESSOR_NAME ? MPI_MAX_PROCESSOR_NAME : strlen(name); + strlen(name) > + MPI_MAX_PROCESSOR_NAME ? MPI_MAX_PROCESSOR_NAME : strlen(name); smpi_bench_begin(); return retval; @@ -2639,476 +2639,548 @@ int PMPI_Error_class(int errorcode, int* errorclass) { int PMPI_Initialized(int* flag) { - *flag=smpi_process_initialized(); - return MPI_SUCCESS; + *flag=smpi_process_initialized(); + return MPI_SUCCESS; } -/* The following calls are not yet implemented and will fail at runtime. */ -/* Once implemented, please move them above this notice. */ - -#define NOT_YET_IMPLEMENTED {\ - XBT_WARN("Not yet implemented : %s. Please contact the Simgrid team if support is needed", __FUNCTION__);\ - return MPI_SUCCESS;\ - } +/* The topo part of MPI_COMM_WORLD should always be NULL. When other topologies + * will be implemented, not only should we check if the topology is NULL, but + * we should check if it is the good topology type (so we have to add a + * MPIR_Topo_Type field, and replace the MPI_Topology field by an union)*/ +int PMPI_Cart_create(MPI_Comm comm_old, int ndims, int* dims, int* periodic, int reorder, MPI_Comm* comm_cart) { + int retval = 0; + smpi_bench_end(); + if (comm_old == MPI_COMM_NULL){ + return MPI_ERR_COMM; + } + else if (ndims < 0 || + (ndims > 0 && (dims == NULL || + periodic == NULL)) || + comm_cart == NULL) { + return MPI_ERR_ARG; + } + retval = smpi_mpi_cart_create(comm_old, ndims, dims, periodic, reorder, comm_cart); -int PMPI_Type_dup(MPI_Datatype datatype, MPI_Datatype *newtype){ - NOT_YET_IMPLEMENTED -} + smpi_bench_begin(); -int PMPI_Type_set_name(MPI_Datatype datatype, char * name) -{ - NOT_YET_IMPLEMENTED + return retval; } -int PMPI_Type_get_name(MPI_Datatype datatype, char * name, int* len) -{ - NOT_YET_IMPLEMENTED +int PMPI_Cart_rank(MPI_Comm comm, int* coords, int* rank) { + if(comm == MPI_COMM_NULL || smpi_comm_topo(comm) == NULL) { + return MPI_ERR_TOPOLOGY; + } + if (coords == NULL) { + return MPI_ERR_ARG; + } + return smpi_mpi_cart_rank(comm, coords, rank); } -int PMPI_Pack_size(int incount, MPI_Datatype datatype, MPI_Comm comm, int* size) { - NOT_YET_IMPLEMENTED +int PMPI_Cart_shift(MPI_Comm comm, int direction, int displ, int* source, int* dest) { + if(comm == MPI_COMM_NULL || smpi_comm_topo(comm) == NULL) { + return MPI_ERR_TOPOLOGY; + } + if (source == NULL || dest == NULL || direction < 0 ) { + return MPI_ERR_ARG; + } + return smpi_mpi_cart_shift(comm, direction, displ, source, dest); } int PMPI_Cart_coords(MPI_Comm comm, int rank, int maxdims, int* coords) { - NOT_YET_IMPLEMENTED -} - -int PMPI_Cart_create(MPI_Comm comm_old, int ndims, int* dims, int* periods, int reorder, MPI_Comm* comm_cart) { - NOT_YET_IMPLEMENTED + if(comm == MPI_COMM_NULL || smpi_comm_topo(comm) == NULL) { + return MPI_ERR_TOPOLOGY; + } + if (rank < 0 || rank >= smpi_comm_size(comm)) { + return MPI_ERR_RANK; + } + if (maxdims <= 0) { + return MPI_ERR_ARG; + } + if(coords == NULL) { + return MPI_ERR_ARG; + } + return smpi_mpi_cart_coords(comm, rank, maxdims, coords); } int PMPI_Cart_get(MPI_Comm comm, int maxdims, int* dims, int* periods, int* coords) { - NOT_YET_IMPLEMENTED + if(comm == NULL || smpi_comm_topo(comm) == NULL) { + return MPI_ERR_TOPOLOGY; + } + if(maxdims <= 0 || dims == NULL || periods == NULL || coords == NULL) { + return MPI_ERR_ARG; + } + return smpi_mpi_cart_get(comm, maxdims, dims, periods, coords); } -int PMPI_Cart_map(MPI_Comm comm_old, int ndims, int* dims, int* periods, int* newrank) { - NOT_YET_IMPLEMENTED +int PMPI_Cartdim_get(MPI_Comm comm, int* ndims) { + if (comm == MPI_COMM_NULL || smpi_comm_topo(comm) == NULL) { + return MPI_ERR_TOPOLOGY; + } + if (ndims == NULL) { + return MPI_ERR_ARG; + } + return smpi_mpi_cartdim_get(comm, ndims); } -int PMPI_Cart_rank(MPI_Comm comm, int* coords, int* rank) { - NOT_YET_IMPLEMENTED -} +int PMPI_Dims_create(int nnodes, int ndims, int* dims) { + if(dims == NULL) { + return MPI_ERR_ARG; + } + if (ndims < 1 || nnodes < 1) { + return MPI_ERR_DIMS; + } -int PMPI_Cart_shift(MPI_Comm comm, int direction, int displ, int* source, int* dest) { - NOT_YET_IMPLEMENTED + return smpi_mpi_dims_create(nnodes, ndims, dims); } int PMPI_Cart_sub(MPI_Comm comm, int* remain_dims, MPI_Comm* comm_new) { - NOT_YET_IMPLEMENTED + if(comm == MPI_COMM_NULL || smpi_comm_topo(comm) == NULL) { + return MPI_ERR_TOPOLOGY; + } + if (comm_new == NULL) { + return MPI_ERR_ARG; + } + return smpi_mpi_cart_sub(comm, remain_dims, comm_new); } -int PMPI_Cartdim_get(MPI_Comm comm, int* ndims) { - NOT_YET_IMPLEMENTED -} -int PMPI_Graph_create(MPI_Comm comm_old, int nnodes, int* index, int* edges, int reorder, MPI_Comm* comm_graph) { - NOT_YET_IMPLEMENTED -} +/* The following calls are not yet implemented and will fail at runtime. */ +/* Once implemented, please move them above this notice. */ -int PMPI_Graph_get(MPI_Comm comm, int maxindex, int maxedges, int* index, int* edges) { - NOT_YET_IMPLEMENTED -} +#define NOT_YET_IMPLEMENTED { \ + XBT_WARN("Not yet implemented : %s. Please contact the Simgrid team if support is needed", __FUNCTION__); \ + return MPI_SUCCESS; \ + } + + +int PMPI_Type_dup(MPI_Datatype datatype, MPI_Datatype *newtype){ + NOT_YET_IMPLEMENTED + } + +int PMPI_Type_set_name(MPI_Datatype datatype, char * name) +{ + NOT_YET_IMPLEMENTED + } + +int PMPI_Type_get_name(MPI_Datatype datatype, char * name, int* len) +{ + NOT_YET_IMPLEMENTED + } + +int PMPI_Pack_size(int incount, MPI_Datatype datatype, MPI_Comm comm, int* size) { + NOT_YET_IMPLEMENTED + } + + +int PMPI_Cart_map(MPI_Comm comm_old, int ndims, int* dims, int* periods, int* newrank) { + NOT_YET_IMPLEMENTED + } + + +int PMPI_Graph_create(MPI_Comm comm_old, int nnodes, int* index, int* edges, int reorder, MPI_Comm* comm_graph) { + NOT_YET_IMPLEMENTED + } + +int PMPI_Graph_get(MPI_Comm comm, int maxindex, int maxedges, int* index, int* edges) { + NOT_YET_IMPLEMENTED + } int PMPI_Graph_map(MPI_Comm comm_old, int nnodes, int* index, int* edges, int* newrank) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Graph_neighbors(MPI_Comm comm, int rank, int maxneighbors, int* neighbors) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Graph_neighbors_count(MPI_Comm comm, int rank, int* nneighbors) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Graphdims_get(MPI_Comm comm, int* nnodes, int* nedges) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Topo_test(MPI_Comm comm, int* top_type) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Errhandler_create(MPI_Handler_function* function, MPI_Errhandler* errhandler) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Errhandler_free(MPI_Errhandler* errhandler) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Errhandler_get(MPI_Comm comm, MPI_Errhandler* errhandler) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Error_string(int errorcode, char* string, int* resultlen) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Errhandler_set(MPI_Comm comm, MPI_Errhandler errhandler) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_set_errhandler(MPI_Comm comm, MPI_Errhandler errhandler) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_get_errhandler(MPI_Comm comm, MPI_Errhandler* errhandler) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Cancel(MPI_Request* request) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Buffer_attach(void* buffer, int size) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Buffer_detach(void* buffer, int* size) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_test_inter(MPI_Comm comm, int* flag) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_get_attr (MPI_Comm comm, int comm_keyval, void *attribute_val, int *flag) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_set_attr (MPI_Comm comm, int comm_keyval, void *attribute_val) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_delete_attr (MPI_Comm comm, int comm_keyval) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_create_keyval(MPI_Comm_copy_attr_function* copy_fn, MPI_Comm_delete_attr_function* delete_fn, int* keyval, void* extra_state) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_free_keyval(int* keyval) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Pcontrol(const int level ) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Unpack(void* inbuf, int insize, int* position, void* outbuf, int outcount, MPI_Datatype type, MPI_Comm comm) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Type_get_attr (MPI_Datatype type, int type_keyval, void *attribute_val, int* flag) { NOT_YET_IMPLEMENTED -} + } int PMPI_Type_set_attr (MPI_Datatype type, int type_keyval, void *attribute_val) { NOT_YET_IMPLEMENTED -} + } int PMPI_Type_delete_attr (MPI_Datatype type, int comm_keyval) { NOT_YET_IMPLEMENTED -} + } int PMPI_Type_create_keyval(MPI_Type_copy_attr_function* copy_fn, MPI_Type_delete_attr_function* delete_fn, int* keyval, void* extra_state) { NOT_YET_IMPLEMENTED -} + } int PMPI_Type_free_keyval(int* keyval) { NOT_YET_IMPLEMENTED -} + } int PMPI_Intercomm_create(MPI_Comm local_comm, int local_leader, MPI_Comm peer_comm, int remote_leader, int tag, MPI_Comm* comm_out) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Intercomm_merge(MPI_Comm comm, int high, MPI_Comm* comm_out) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Bsend(void* buf, int count, MPI_Datatype datatype, int dest, int tag, MPI_Comm comm) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Bsend_init(void* buf, int count, MPI_Datatype datatype, int dest, int tag, MPI_Comm comm, MPI_Request* request) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Ibsend(void* buf, int count, MPI_Datatype datatype, int dest, int tag, MPI_Comm comm, MPI_Request* request) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_remote_group(MPI_Comm comm, MPI_Group* group) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Comm_remote_size(MPI_Comm comm, int* size) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Attr_delete(MPI_Comm comm, int keyval) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Attr_get(MPI_Comm comm, int keyval, void* attr_value, int* flag) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Attr_put(MPI_Comm comm, int keyval, void* attr_value) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Rsend(void* buf, int count, MPI_Datatype datatype, int dest, int tag, MPI_Comm comm) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Rsend_init(void* buf, int count, MPI_Datatype datatype, int dest, int tag, MPI_Comm comm, MPI_Request* request) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Irsend(void* buf, int count, MPI_Datatype datatype, int dest, int tag, MPI_Comm comm, MPI_Request* request) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Keyval_create(MPI_Copy_function* copy_fn, MPI_Delete_function* delete_fn, int* keyval, void* extra_state) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Keyval_free(int* keyval) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Test_cancelled(MPI_Status* status, int* flag) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Pack(void* inbuf, int incount, MPI_Datatype type, void* outbuf, int outcount, int* position, MPI_Comm comm) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Pack_external_size(char *datarep, int incount, MPI_Datatype datatype, MPI_Aint *size){ NOT_YET_IMPLEMENTED -} + } int PMPI_Pack_external(char *datarep, void *inbuf, int incount, MPI_Datatype datatype, void *outbuf, MPI_Aint outcount, MPI_Aint *position){ NOT_YET_IMPLEMENTED -} + } int PMPI_Unpack_external( char *datarep, void *inbuf, MPI_Aint insize, MPI_Aint *position, void *outbuf, int outcount, MPI_Datatype datatype){ NOT_YET_IMPLEMENTED -} + } int PMPI_Get_elements(MPI_Status* status, MPI_Datatype datatype, int* elements) { - NOT_YET_IMPLEMENTED -} - -int PMPI_Dims_create(int nnodes, int ndims, int* dims) { - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Win_fence( int assert, MPI_Win win){ - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Win_free( MPI_Win* win){ - NOT_YET_IMPLEMENTED -} + NOT_YET_IMPLEMENTED + } int PMPI_Win_create( void *base, MPI_Aint size, int disp_unit, MPI_Info info, MPI_Comm comm, MPI_Win *win){ NOT_YET_IMPLEMENTED -} + } int PMPI_Info_create( MPI_Info *info){ NOT_YET_IMPLEMENTED -} + } int PMPI_Info_set( MPI_Info info, char *key, char *value){ NOT_YET_IMPLEMENTED -} + } int PMPI_Info_free( MPI_Info *info){ NOT_YET_IMPLEMENTED -} + } int PMPI_Get( void *origin_addr, int origin_count, MPI_Datatype origin_datatype, int target_rank, - MPI_Aint target_disp, int target_count, MPI_Datatype target_datatype, MPI_Win win){ + MPI_Aint target_disp, int target_count, MPI_Datatype target_datatype, MPI_Win win){ NOT_YET_IMPLEMENTED -} + } int PMPI_Type_get_envelope( MPI_Datatype datatype, int *num_integers, - int *num_addresses, int *num_datatypes, int *combiner){ + int *num_addresses, int *num_datatypes, int *combiner){ NOT_YET_IMPLEMENTED -} + } int PMPI_Type_get_contents(MPI_Datatype datatype, int max_integers, int max_addresses, - int max_datatypes, int* array_of_integers, MPI_Aint* array_of_addresses, - MPI_Datatype* array_of_datatypes){ + int max_datatypes, int* array_of_integers, MPI_Aint* array_of_addresses, + MPI_Datatype* array_of_datatypes){ NOT_YET_IMPLEMENTED -} + } int PMPI_Type_create_darray(int size, int rank, int ndims, int* array_of_gsizes, int* array_of_distribs, int* array_of_dargs, int* array_of_psizes, int order, MPI_Datatype oldtype, MPI_Datatype *newtype) { NOT_YET_IMPLEMENTED -} + } int PMPI_Type_create_resized(MPI_Datatype oldtype,MPI_Aint lb, MPI_Aint extent, MPI_Datatype *newtype){ NOT_YET_IMPLEMENTED -} + } int PMPI_Type_create_subarray(int ndims,int *array_of_sizes, int *array_of_subsizes, int *array_of_starts, int order, MPI_Datatype oldtype, MPI_Datatype *newtype){ NOT_YET_IMPLEMENTED -} + } int PMPI_Type_match_size(int typeclass,int size,MPI_Datatype *datatype){ NOT_YET_IMPLEMENTED -} + } int PMPI_Alltoallw( void *sendbuf, int *sendcnts, int *sdispls, MPI_Datatype *sendtypes, - void *recvbuf, int *recvcnts, int *rdispls, MPI_Datatype *recvtypes, - MPI_Comm comm){ + void *recvbuf, int *recvcnts, int *rdispls, MPI_Datatype *recvtypes, + MPI_Comm comm){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_set_name(MPI_Comm comm, char* name){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_dup_with_info(MPI_Comm comm, MPI_Info info, MPI_Comm * newcomm){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_split_type(MPI_Comm comm, int split_type, int key, MPI_Info info, MPI_Comm *newcomm){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_set_info (MPI_Comm comm, MPI_Info info){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_get_info (MPI_Comm comm, MPI_Info* info){ NOT_YET_IMPLEMENTED -} + } int PMPI_Info_get(MPI_Info info,char *key,int valuelen, char *value, int *flag){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_create_errhandler( MPI_Comm_errhandler_fn *function, MPI_Errhandler *errhandler){ NOT_YET_IMPLEMENTED -} + } int PMPI_Add_error_class( int *errorclass){ NOT_YET_IMPLEMENTED -} + } int PMPI_Add_error_code( int errorclass, int *errorcode){ NOT_YET_IMPLEMENTED -} + } int PMPI_Add_error_string( int errorcode, char *string){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_call_errhandler(MPI_Comm comm,int errorcode){ NOT_YET_IMPLEMENTED -} + } int PMPI_Info_dup(MPI_Info info, MPI_Info *newinfo){ NOT_YET_IMPLEMENTED -} + } int PMPI_Info_delete(MPI_Info info, char *key){ NOT_YET_IMPLEMENTED -} + } int PMPI_Info_get_nkeys( MPI_Info info, int *nkeys){ NOT_YET_IMPLEMENTED -} + } int PMPI_Info_get_nthkey( MPI_Info info, int n, char *key){ NOT_YET_IMPLEMENTED -} + } int PMPI_Info_get_valuelen( MPI_Info info, char *key, int *valuelen, int *flag){ NOT_YET_IMPLEMENTED -} + } int PMPI_Request_get_status( MPI_Request request, int *flag, MPI_Status *status){ NOT_YET_IMPLEMENTED -} + } int PMPI_Grequest_start( MPI_Grequest_query_function *query_fn, MPI_Grequest_free_function *free_fn, MPI_Grequest_cancel_function *cancel_fn, void *extra_state, MPI_Request *request){ NOT_YET_IMPLEMENTED -} + } int PMPI_Grequest_complete( MPI_Request request){ NOT_YET_IMPLEMENTED -} + } int PMPI_Status_set_cancelled(MPI_Status *status,int flag){ NOT_YET_IMPLEMENTED -} + } int PMPI_Status_set_elements( MPI_Status *status, MPI_Datatype datatype, int count){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_connect( char *port_name, MPI_Info info, int root, MPI_Comm comm, MPI_Comm *newcomm){ NOT_YET_IMPLEMENTED -} + } int PMPI_Publish_name( char *service_name, MPI_Info info, char *port_name){ NOT_YET_IMPLEMENTED -} + } int PMPI_Unpublish_name( char *service_name, MPI_Info info, char *port_name){ NOT_YET_IMPLEMENTED -} + } int PMPI_Lookup_name( char *service_name, MPI_Info info, char *port_name){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_join( int fd, MPI_Comm *intercomm){ NOT_YET_IMPLEMENTED -} + } int PMPI_Open_port( MPI_Info info, char *port_name){ NOT_YET_IMPLEMENTED -} + } int PMPI_Close_port(char *port_name){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_accept( char *port_name, MPI_Info info, int root, MPI_Comm comm, MPI_Comm *newcomm){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_spawn( char *command, char **argv, int maxprocs, MPI_Info info, int root, MPI_Comm comm, MPI_Comm *intercomm, int* array_of_errcodes){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_spawn_multiple( int count, char **array_of_commands, char*** array_of_argv, - int* array_of_maxprocs, MPI_Info* array_of_info, int root, - MPI_Comm comm, MPI_Comm *intercomm, int* array_of_errcodes){ + int* array_of_maxprocs, MPI_Info* array_of_info, int root, + MPI_Comm comm, MPI_Comm *intercomm, int* array_of_errcodes){ NOT_YET_IMPLEMENTED -} + } int PMPI_Comm_get_parent( MPI_Comm *parent){ NOT_YET_IMPLEMENTED -} + } diff --git a/src/smpi/smpi_topo.c b/src/smpi/smpi_topo.c new file mode 100644 index 0000000000..7429959fa2 --- /dev/null +++ b/src/smpi/smpi_topo.c @@ -0,0 +1,468 @@ +#include "xbt/sysdep.h" +#include "smpi/smpi.h" +#include "private.h" + +#include + +typedef struct s_smpi_mpi_topology { + int nnodes; + int ndims; + int *dims; + int *periodic; + int *position; +} s_smpi_mpi_topology_t; + + + +void smpi_topo_destroy(MPI_Topology topo) { + if (topo) { + if(topo->dims) { + free(topo->dims); + } + if(topo->periodic) { + free(topo->periodic); + } + if(topo->position) { + free(topo->position); + } + free(topo); + } +} + +MPI_Topology smpi_topo_create(int ndims) { + MPI_Topology topo = xbt_malloc(sizeof(*topo)); + topo->nnodes = 0; + topo->ndims = ndims; + topo->dims = xbt_malloc(ndims * sizeof(*topo->dims)); + topo->periodic = xbt_malloc(ndims * sizeof(*topo->periodic)); + topo->position = xbt_malloc(ndims * sizeof(*topo->position)); + return topo; +} + +/* reorder is ignored, don't know what would be the consequences of a dumb + * reordering but neither do I see the point of reordering*/ +int smpi_mpi_cart_create(MPI_Comm comm_old, int ndims, int dims[], + int periods[], int reorder, MPI_Comm *comm_cart) { + int retval = MPI_SUCCESS; + int i; + MPI_Topology topo; + MPI_Group newGroup, oldGroup; + int rank, nranks, newSize; + + + + rank = smpi_comm_rank(comm_old); + + // TODO : add somewhere : if topo != NULL, free... + topo = smpi_topo_create(ndims); + newSize = 1; + if(ndims != 0) { + for (i = 0 ; i < ndims ; i++) { + newSize *= dims[i]; + } + if(rank >= newSize) { + *comm_cart = MPI_COMM_NULL; + return retval; + } + oldGroup = smpi_comm_group(comm_old); + newGroup = smpi_group_new(newSize); + for (i = 0 ; i < newSize ; i++) { + smpi_group_set_mapping(newGroup, smpi_group_index(oldGroup, i), i); + } + + topo->nnodes = newSize; + + memcpy(topo->dims, dims, ndims * sizeof(*topo->dims)); + memcpy(topo->periodic, periods, ndims * sizeof(*topo->periodic)); + + // code duplication... See smpi_mpi_cart_coords + nranks = newSize; + for (i=0; idims[i] = dims[i]; + topo->periodic[i] = periods[i]; + nranks = nranks / dims[i]; + /* FIXME: nranks could be zero (?) */ + topo->position[i] = rank / nranks; + rank = rank % nranks; + } + + *comm_cart = smpi_comm_new(newGroup, topo); + } + else { + if (rank == 0) { + *comm_cart = smpi_comm_new(smpi_comm_group(MPI_COMM_SELF), topo); + } + else { + *comm_cart = MPI_COMM_NULL; + } + } + return retval; +} + +int smpi_mpi_cart_sub(MPI_Comm comm, const int remain_dims[], MPI_Comm *newcomm) { + MPI_Topology oldTopo = smpi_comm_topo(comm); + int oldNDims = oldTopo->ndims; + int i, j = 0, newNDims, *newDims = NULL, *newPeriodic = NULL; + + if (remain_dims == NULL && oldNDims != 0) { + return MPI_ERR_ARG; + } + newNDims = 0; + for (i = 0 ; i < oldNDims ; i++) { + if (remain_dims[i]) newNDims++; + } + + if (newNDims > 0) { + newDims = malloc(newNDims * sizeof(*newDims)); + newPeriodic = malloc(newNDims * sizeof(*newPeriodic)); + + // that should not segfault + for (i = 0 ; j < newNDims ; i++) { + if(remain_dims[i]) { + newDims[j] = oldTopo->dims[i]; + newPeriodic[j] = oldTopo->periodic[i]; + j++; + } + } + } + return smpi_mpi_cart_create(comm, newNDims, newDims, newPeriodic, 0, newcomm); +} + + + + +int smpi_mpi_cart_coords(MPI_Comm comm, int rank, int maxdims, + int coords[]) { + int nnodes; + int i; + MPI_Topology topo = smpi_comm_topo(comm); + + nnodes = topo->nnodes; + for ( i=0; i < topo->ndims; i++ ) { + nnodes = nnodes / topo->dims[i]; + coords[i] = rank / nnodes; + rank = rank % nnodes; + } + return MPI_SUCCESS; +} + +int smpi_mpi_cart_get(MPI_Comm comm, int maxdims, int* dims, int* periods, int* coords) { + MPI_Topology topo = smpi_comm_topo(comm); + int i; + for(i = 0 ; i < maxdims ; i++) { + dims[i] = topo->dims[i]; + periods[i] = topo->periodic[i]; + coords[i] = topo->position[i]; + } + return MPI_SUCCESS; +} + +int smpi_mpi_cart_rank(MPI_Comm comm, int* coords, int* rank) { + MPI_Topology topo = smpi_comm_topo(comm); + int ndims = topo->ndims; + int multiplier, coord,i; + *rank = 0; + multiplier = 1; + + + + for ( i=ndims-1; i >=0; i-- ) { + coord = coords[i]; + + /* Should we check first for args correction, then process, + * or check while we work (as it is currently done) ? */ + if (coord >= topo->dims[i]) { + if ( topo->periodic[i] ) { + coord = coord % topo->dims[i]; + } + else { + // Should I do that ? + *rank = -1; + return MPI_ERR_ARG; + } + } + else if (coord < 0) { + if(topo->periodic[i]) { + coord = coord % topo->dims[i]; + if (coord) coord = topo->dims[i] + coord; + } + else { + *rank = -1; + return MPI_ERR_ARG; + } + } + + *rank += multiplier * coord; + multiplier *= topo->dims[i]; + } + return MPI_SUCCESS; +} + +int smpi_mpi_cart_shift(MPI_Comm comm, int direction, int disp, + int *rank_source, int *rank_dest) { + MPI_Topology topo = smpi_comm_topo(comm); + int position[topo->ndims]; + + + if(topo->ndims == 0) { + return MPI_ERR_ARG; + } + if (topo->ndims < direction) { + return MPI_ERR_DIMS; + } + + smpi_mpi_cart_coords(comm, smpi_comm_rank(comm), topo->ndims, position); + position[direction] += disp; + + if(position[direction] < 0 || position[direction] >= topo->dims[direction]) { + if(topo->periodic[direction]) { + position[direction] %= topo->dims[direction]; + smpi_mpi_cart_rank(comm, position, rank_dest); + } + else { + *rank_dest = MPI_PROC_NULL; + } + } + else { + smpi_mpi_cart_rank(comm, position, rank_dest); + } + + position[direction] = topo->position[direction] - disp; + if(position[direction] < 0 || position[direction] >= topo->dims[direction]) { + if(topo->periodic[direction]) { + position[direction] %= topo->dims[direction]; + smpi_mpi_cart_rank(comm, position, rank_source); + } + else { + *rank_source = MPI_PROC_NULL; + } + } + else { + smpi_mpi_cart_rank(comm, position, rank_source); + } + + return MPI_SUCCESS; +} + +int smpi_mpi_cartdim_get(MPI_Comm comm, int *ndims) { + MPI_Topology topo = smpi_comm_topo(comm); + + *ndims = topo->ndims; + return MPI_SUCCESS; +} + + + +// Everything below has been taken from ompi, but could be easily rewritten. + +/* + * Copyright (c) 2004-2007 The Trustees of Indiana University and Indiana + * University Research and Technology + * Corporation. All rights reserved. + * Copyright (c) 2004-2005 The University of Tennessee and The University + * of Tennessee Research Foundation. All rights + * reserved. + * Copyright (c) 2004-2014 High Performance Computing Center Stuttgart, + * University of Stuttgart. All rights reserved. + * Copyright (c) 2004-2005 The Regents of the University of California. + * All rights reserved. + * Copyright (c) 2012 Los Alamos National Security, LLC. All rights + * reserved. + * Copyright (c) 2014 Intel, Inc. All rights reserved + * $COPYRIGHT$ + * + * Additional copyrights may follow + * + * $HEADER$ + */ + + +/* static functions */ +static int assignnodes(int ndim, int nfactor, int *pfacts,int **pdims); +static int getfactors(int num, int *nfators, int **factors); + +/* + * This is a utility function, no need to have anything in the lower + * layer for this at all + */ +int smpi_mpi_dims_create(int nnodes, int ndims, int dims[]) +{ + int i; + int freeprocs; + int freedims; + int nfactors; + int *factors; + int *procs; + int *p; + int err; + + /* Get # of free-to-be-assigned processes and # of free dimensions */ + freeprocs = nnodes; + freedims = 0; + for (i = 0, p = dims; i < ndims; ++i,++p) { + if (*p == 0) { + ++freedims; + } else if ((*p < 0) || ((nnodes % *p) != 0)) { + return MPI_ERR_DIMS; + + } else { + freeprocs /= *p; + } + } + + if (freedims == 0) { + if (freeprocs == 1) { + return MPI_SUCCESS; + } + return MPI_ERR_DIMS; + } + + if (freeprocs == 1) { + for (i = 0; i < ndims; ++i, ++dims) { + if (*dims == 0) { + *dims = 1; + } + } + return MPI_SUCCESS; + } + + /* Factor the number of free processes */ + if (MPI_SUCCESS != (err = getfactors(freeprocs, &nfactors, &factors))) { + return err; + } + + /* Assign free processes to free dimensions */ + if (MPI_SUCCESS != (err = assignnodes(freedims, nfactors, factors, &procs))) { + return err; + } + + /* Return assignment results */ + p = procs; + for (i = 0; i < ndims; ++i, ++dims) { + if (*dims == 0) { + *dims = *p++; + } + } + + free((char *) factors); + free((char *) procs); + + /* all done */ + return MPI_SUCCESS; +} + +/* + * assignnodes + * + * Function: - assign processes to dimensions + * - get "best-balanced" grid + * - greedy bin-packing algorithm used + * - sort dimensions in decreasing order + * - dimensions array dynamically allocated + * Accepts: - # of dimensions + * - # of prime factors + * - array of prime factors + * - ptr to array of dimensions (returned value) + * Returns: - 0 or ERROR + */ +static int +assignnodes(int ndim, int nfactor, int *pfacts, int **pdims) +{ + int *bins; + int i, j; + int n; + int f; + int *p; + int *pmin; + + if (0 >= ndim) { + return MPI_ERR_DIMS; + } + + /* Allocate and initialize the bins */ + bins = (int *) malloc((unsigned) ndim * sizeof(int)); + if (NULL == bins) { + return MPI_ERR_NO_MEM; + } + *pdims = bins; + + for (i = 0, p = bins; i < ndim; ++i, ++p) { + *p = 1; + } + + /* Loop assigning factors from the highest to the lowest */ + for (j = nfactor - 1; j >= 0; --j) { + f = pfacts[j]; + /* Assign a factor to the smallest bin */ + pmin = bins; + for (i = 1, p = pmin + 1; i < ndim; ++i, ++p) { + if (*p < *pmin) { + pmin = p; + } + } + *pmin *= f; + } + + /* Sort dimensions in decreasing order (O(n^2) for now) */ + for (i = 0, pmin = bins; i < ndim - 1; ++i, ++pmin) { + for (j = i + 1, p = pmin + 1; j < ndim; ++j, ++p) { + if (*p > *pmin) { + n = *p; + *p = *pmin; + *pmin = n; + } + } + } + + return MPI_SUCCESS; +} + +/* + * getfactors + * + * Function: - factorize a number + * Accepts: - number + * - # prime factors + * - array of prime factors + * Returns: - MPI_SUCCESS or ERROR + */ +static int +getfactors(int num, int *nfactors, int **factors) { + int size; + int d; + int i; + int sqrtnum; + + if(num < 2) { + (*nfactors) = 0; + (*factors) = NULL; + return MPI_SUCCESS; + } + /* Allocate the array of prime factors which cannot exceed log_2(num) entries */ + sqrtnum = ceil(sqrt(num)); + size = ceil(log(num) / log(2)); + *factors = (int *) malloc((unsigned) size * sizeof(int)); + + i = 0; + /* determine all occurences of factor 2 */ + while((num % 2) == 0) { + num /= 2; + (*factors)[i++] = 2; + } + /* determine all occurences of uneven prime numbers up to sqrt(num) */ + d = 3; + for(d = 3; (num > 1) && (d < sqrtnum); d += 2) { + while((num % d) == 0) { + num /= d; + (*factors)[i++] = d; + } + } + /* as we looped only up to sqrt(num) one factor > sqrt(num) may be left over */ + if(num != 1) { + (*factors)[i++] = num; + } + (*nfactors) = i; + return MPI_SUCCESS; +} + diff --git a/teshsuite/smpi/mpich3-test/testlist b/teshsuite/smpi/mpich3-test/testlist index 2110a22696..900c9194ab 100644 --- a/teshsuite/smpi/mpich3-test/testlist +++ b/teshsuite/smpi/mpich3-test/testlist @@ -13,7 +13,7 @@ init pt2pt # #spawn -#topo +topo #perf #io f77 diff --git a/teshsuite/smpi/mpich3-test/topo/CMakeLists.txt b/teshsuite/smpi/mpich3-test/topo/CMakeLists.txt index 6bc6b43270..0a0bc87c9c 100644 --- a/teshsuite/smpi/mpich3-test/topo/CMakeLists.txt +++ b/teshsuite/smpi/mpich3-test/topo/CMakeLists.txt @@ -13,28 +13,28 @@ if(enable_smpi AND enable_smpi_MPICH3_testsuite) include_directories("${CMAKE_CURRENT_SOURCE_DIR}/../include/") -# add_executable(cartcreates cartcreates.c) -# add_executable(cartmap1 cartmap1.c) -# add_executable(cartshift1 cartshift1.c) -# add_executable(cartsuball cartsuball.c) -# add_executable(cartzero cartzero.c) +add_executable(cartcreates cartcreates.c) + #add_executable(cartmap1 cartmap1.c) +add_executable(cartshift1 cartshift1.c) +add_executable(cartsuball cartsuball.c) +add_executable(cartzero cartzero.c) # add_executable(dgraph_unwgt dgraph_unwgt.c) -# add_executable(dims1 dims1.c) -# add_executable(dims2 dims2.c) +#add_executable(dims1 dims1.c) +#add_executable(dims2 dims2.c) # add_executable(distgraph1 distgraph1.c) # add_executable(graphcr2 graphcr2.c) # add_executable(graphcr graphcr.c) # add_executable(graphmap1 graphmap1.c) -# add_executable(neighb_coll neighb_coll.c) +#add_executable(neighb_coll neighb_coll.c) # add_executable(topodup topodup.c) -# add_executable(topotest topotest.c) +#add_executable(topotest topotest.c) -# target_link_libraries(cartcreates simgrid mtest_c) -# target_link_libraries(cartmap1 simgrid mtest_c) -# target_link_libraries(cartshift1 simgrid mtest_c) -# target_link_libraries(cartsuball simgrid mtest_c) -# target_link_libraries(cartzero simgrid mtest_c) +target_link_libraries(cartcreates simgrid mtest_c) +# target_link_libraries(cartmap1 simgrid mtest_c) +target_link_libraries(cartshift1 simgrid mtest_c) +target_link_libraries(cartsuball simgrid mtest_c) +target_link_libraries(cartzero simgrid mtest_c) # target_link_libraries(dgraph_unwgt simgrid mtest_c) # target_link_libraries(dims1 simgrid mtest_c) # target_link_libraries(dims2 simgrid mtest_c) diff --git a/teshsuite/smpi/mpich3-test/topo/testlist b/teshsuite/smpi/mpich3-test/topo/testlist index 0ec8867d09..ba25d597b2 100644 --- a/teshsuite/smpi/mpich3-test/topo/testlist +++ b/teshsuite/smpi/mpich3-test/topo/testlist @@ -1,9 +1,9 @@ #need Cart implem #cartmap1 4 -#cartzero 4 -#cartshift1 4 -#cartsuball 4 -#cartcreates 4 +cartzero 4 +cartshift1 4 +cartsuball 4 +cartcreates 4 #need MPI_Dims_create #dims1 4 #dims2 1 -- 2.20.1