X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/7c14726fd0862a988b835b24e85c6c5b83b81900..3203ed1f4fb6f4617e94d5351661ba07e8b8f833:/src/smpi/smpi_comm.c diff --git a/src/smpi/smpi_comm.c b/src/smpi/smpi_comm.c index d71c475df1..9f38c44267 100644 --- a/src/smpi/smpi_comm.c +++ b/src/smpi/smpi_comm.c @@ -1,4 +1,4 @@ -/* Copyright (c) 2010-2014. The SimGrid Team. +/* Copyright (c) 2010-2015. The SimGrid Team. * All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it @@ -7,15 +7,18 @@ #include #include "private.h" +#include "xbt/dict.h" #include "smpi_mpi_dt_private.h" #include "limits.h" -#include "simix/smx_private.h" +#include "src/simix/smx_private.h" #include "colls/colls.h" +#include "xbt/ex.h" XBT_LOG_NEW_DEFAULT_SUBCATEGORY(smpi_comm, smpi, "Logging specific to SMPI (comm)"); - +xbt_dict_t smpi_comm_keyvals = NULL; +int comm_keyval_id = 0;//avoid collisions /* Support for cartesian topology was added, but there are 2 other types of * topology, graph et dist graph. In order to support them, we have to add a @@ -33,6 +36,7 @@ typedef struct s_smpi_mpi_communicator { int is_uniform; int* non_uniform_map; //set if smp nodes have a different number of processes allocated int is_blocked;// are ranks allocated on the same smp node contiguous ? + xbt_dict_t attributes; } s_smpi_mpi_communicator_t; static int smpi_compare_rankmap(const void *a, const void *b) @@ -71,6 +75,7 @@ MPI_Comm smpi_comm_new(MPI_Group group, MPI_Topology topo) comm->non_uniform_map = NULL; comm->leaders_map = NULL; comm->is_blocked=0; + comm->attributes=NULL; return comm; } @@ -83,6 +88,56 @@ void smpi_comm_destroy(MPI_Comm comm) smpi_comm_unuse(comm); } +int smpi_comm_dup(MPI_Comm comm, MPI_Comm* newcomm){ + if(smpi_privatize_global_variables){ //we need to switch here, as the called function may silently touch global variables + smpi_switch_data_segment(smpi_process_index()); + } + (*newcomm) = smpi_comm_new(smpi_comm_group(comm), smpi_comm_topo(comm)); + int ret = MPI_SUCCESS; + //todo: faire en sorte que ça fonctionne avec un communicator dupliqué (refaire un init_smp ?) + + /* MPI_Comm tmp=smpi_comm_get_intra_comm(comm); + if( tmp != MPI_COMM_NULL) + smpi_comm_set_intra_comm((*newcomm), smpi_comm_dup(tmp)); + tmp=smpi_comm_get_leaders_comm(comm); + if( tmp != MPI_COMM_NULL) + smpi_comm_set_leaders_comm((*newcomm), smpi_comm_dup(tmp)); + if(comm->non_uniform_map !=NULL){ + (*newcomm)->non_uniform_map= + xbt_malloc(smpi_comm_size(comm->leaders_comm)*sizeof(int)); + memcpy((*newcomm)->non_uniform_map, + comm->non_uniform_map,smpi_comm_size(comm->leaders_comm)*sizeof(int) ); + } + if(comm->leaders_map !=NULL){ + (*newcomm)->leaders_map=xbt_malloc(smpi_comm_size(comm)*sizeof(int)); + memcpy((*newcomm)->leaders_map, + comm->leaders_map,smpi_comm_size(comm)*sizeof(int) ); + }*/ + if(comm->attributes !=NULL){ + (*newcomm)->attributes=xbt_dict_new(); + xbt_dict_cursor_t cursor = NULL; + int *key; + int flag; + void* value_in; + void* value_out; + xbt_dict_foreach(comm->attributes, cursor, key, value_in){ + smpi_comm_key_elem elem = xbt_dict_get_or_null_ext(smpi_comm_keyvals, (const char*)key, sizeof(int)); + if(elem && elem->copy_fn!=MPI_NULL_COPY_FN){ + ret = elem->copy_fn(comm, *key, NULL, value_in, &value_out, &flag ); + if(ret!=MPI_SUCCESS){ + smpi_comm_destroy(*newcomm); + *newcomm=MPI_COMM_NULL; + return ret; + } + if(flag) + xbt_dict_set_ext((*newcomm)->attributes, (const char*)key, sizeof(int),value_out, NULL); + } + } + } + return ret; +} + + MPI_Group smpi_comm_group(MPI_Comm comm) { if (comm == MPI_COMM_UNINITIALIZED) @@ -125,6 +180,8 @@ void smpi_comm_get_name (MPI_Comm comm, char* name, int* len) } void smpi_comm_set_leaders_comm(MPI_Comm comm, MPI_Comm leaders){ + if (comm == MPI_COMM_UNINITIALIZED) + comm = smpi_process_comm_world(); comm->leaders_comm=leaders; } @@ -133,27 +190,38 @@ void smpi_comm_set_intra_comm(MPI_Comm comm, MPI_Comm leaders){ } int* smpi_comm_get_non_uniform_map(MPI_Comm comm){ + if (comm == MPI_COMM_UNINITIALIZED) + comm = smpi_process_comm_world(); return comm->non_uniform_map; } int* smpi_comm_get_leaders_map(MPI_Comm comm){ + if (comm == MPI_COMM_UNINITIALIZED) + comm = smpi_process_comm_world(); return comm->leaders_map; } MPI_Comm smpi_comm_get_leaders_comm(MPI_Comm comm){ + if (comm == MPI_COMM_UNINITIALIZED) + comm = smpi_process_comm_world(); return comm->leaders_comm; } MPI_Comm smpi_comm_get_intra_comm(MPI_Comm comm){ - if(comm==MPI_COMM_WORLD) return smpi_process_get_comm_intra(); + if (comm == MPI_COMM_UNINITIALIZED || comm==MPI_COMM_WORLD) + return smpi_process_get_comm_intra(); else return comm->intra_comm; } int smpi_comm_is_uniform(MPI_Comm comm){ + if (comm == MPI_COMM_UNINITIALIZED) + comm = smpi_process_comm_world(); return comm->is_uniform; } int smpi_comm_is_blocked(MPI_Comm comm){ + if (comm == MPI_COMM_UNINITIALIZED) + comm = smpi_process_comm_world(); return comm->is_blocked; } @@ -259,6 +327,17 @@ void smpi_comm_unuse(MPI_Comm comm){ xbt_free(comm->non_uniform_map); if(comm->leaders_map !=NULL) xbt_free(comm->leaders_map); + if(comm->attributes !=NULL){ + xbt_dict_cursor_t cursor = NULL; + int* key; + void * value; + int flag; + xbt_dict_foreach(comm->attributes, cursor, key, value){ + smpi_comm_key_elem elem = xbt_dict_get_or_null(smpi_comm_keyvals, (const char*)key); + if(elem && elem->delete_fn) + elem->delete_fn(comm, *key, value, &flag); + } + } xbt_free(comm); } } @@ -274,7 +353,20 @@ compare_ints (const void *a, const void *b) void smpi_comm_init_smp(MPI_Comm comm){ int leader = -1; + + if (comm == MPI_COMM_UNINITIALIZED) + comm = smpi_process_comm_world(); + int comm_size =smpi_comm_size(comm); + + // If we are in replay - perform an ugly hack + // say to SimGrid that we are not in replay for a while, because we need + // the buffers to be copied for the following calls + int replaying = 0; //cache data to set it back again after + if(smpi_process_get_replaying()){ + replaying=1; + smpi_process_set_replaying(0); + } if(smpi_privatize_global_variables){ //we need to switch here, as the called function may silently touch global variables smpi_switch_data_segment(smpi_process_index()); @@ -290,11 +382,11 @@ void smpi_comm_init_smp(MPI_Comm comm){ // smpi_process_set_comm_intra(MPI_COMM_SELF); // return; // } - XBT_DEBUG("number of processes deployed on my node : %d", intra_comm_size); + int i =0; int min_index=INT_MAX;//the minimum index will be the leader - msg_process_t process = NULL; + smx_process_t process = NULL; xbt_swag_foreach(process, process_list) { //is_in_comm=0; int index = SIMIX_process_get_PID(process) -1; @@ -307,7 +399,7 @@ void smpi_comm_init_smp(MPI_Comm comm){ i++; } } - + XBT_DEBUG("number of processes deployed on my node : %d", intra_comm_size); MPI_Group group_intra = smpi_group_new(intra_comm_size); i=0; process = NULL; @@ -337,6 +429,10 @@ void smpi_comm_init_smp(MPI_Comm comm){ smpi_coll_tuned_allgather_mpich(&leader, 1, MPI_INT , leaders_map, 1, MPI_INT, comm); + if(smpi_privatize_global_variables){ //we need to switch here, as the called function may silently touch global variables + smpi_switch_data_segment(smpi_process_index()); + } + if(!comm->leaders_map){ comm->leaders_map= leaders_map; }else{ @@ -362,7 +458,7 @@ void smpi_comm_init_smp(MPI_Comm comm){ MPI_Comm leader_comm = MPI_COMM_NULL; - if(comm!=MPI_COMM_WORLD){ + if(MPI_COMM_WORLD!=MPI_COMM_UNINITIALIZED && comm!=MPI_COMM_WORLD){ //create leader_communicator for (i=0; i< leader_group_size;i++) smpi_group_set_mapping(leaders_group, leader_list[i], i); @@ -387,7 +483,7 @@ void smpi_comm_init_smp(MPI_Comm comm){ // Are the nodes uniform ? = same number of process/node int my_local_size=smpi_comm_size(comm_intra); if(smpi_comm_rank(comm_intra)==0) { - int* non_uniform_map = xbt_malloc(sizeof(int)*leader_group_size); + int* non_uniform_map = xbt_malloc0(sizeof(int)*leader_group_size); smpi_coll_tuned_allgather_mpich(&my_local_size, 1, MPI_INT, non_uniform_map, 1, MPI_INT, leader_comm); for(i=0; i < leader_group_size; i++) { @@ -405,7 +501,9 @@ void smpi_comm_init_smp(MPI_Comm comm){ } smpi_coll_tuned_bcast_mpich(&(comm->is_uniform),1, MPI_INT, 0, comm_intra ); - + if(smpi_privatize_global_variables){ //we need to switch here, as the called function may silently touch global variables + smpi_switch_data_segment(smpi_process_index()); + } // Are the ranks blocked ? = allocated contiguously on the SMP nodes int is_blocked=1; int prev=smpi_group_rank(smpi_comm_group(comm), smpi_group_index(smpi_comm_group(comm_intra), 0)); @@ -422,7 +520,7 @@ void smpi_comm_init_smp(MPI_Comm comm){ smpi_mpi_allreduce(&is_blocked, &(global_blocked), 1, MPI_INT, MPI_LAND, comm); - if(comm==MPI_COMM_WORLD){ + if(MPI_COMM_WORLD==SMPI_UNINITIALIZED || comm==MPI_COMM_WORLD){ if(smpi_comm_rank(comm)==0){ comm->is_blocked=global_blocked; } @@ -430,5 +528,92 @@ void smpi_comm_init_smp(MPI_Comm comm){ comm->is_blocked=global_blocked; } xbt_free(leader_list); + + if(replaying==1) + smpi_process_set_replaying(1); } +int smpi_comm_attr_delete(MPI_Comm comm, int keyval){ + smpi_comm_key_elem elem = xbt_dict_get_or_null_ext(smpi_comm_keyvals, (const char*)&keyval, sizeof(int)); + if(!elem) + return MPI_ERR_ARG; + if(elem->delete_fn!=MPI_NULL_DELETE_FN){ + void * value; + int flag; + if(smpi_comm_attr_get(comm, keyval, &value, &flag)==MPI_SUCCESS){ + int ret = elem->delete_fn(comm, keyval, value, &flag); + if(ret!=MPI_SUCCESS) return ret; + } + } + if(comm->attributes==NULL) + return MPI_ERR_ARG; + + xbt_dict_remove_ext(comm->attributes, (const char*)&keyval, sizeof(int)); + return MPI_SUCCESS; +} + +int smpi_comm_attr_get(MPI_Comm comm, int keyval, void* attr_value, int* flag){ + smpi_comm_key_elem elem = xbt_dict_get_or_null_ext(smpi_comm_keyvals, (const char*)&keyval, sizeof(int)); + if(!elem) + return MPI_ERR_ARG; + xbt_ex_t ex; + if(comm->attributes==NULL){ + *flag=0; + return MPI_SUCCESS; + } + TRY { + *(void**)attr_value = xbt_dict_get_ext(comm->attributes, (const char*)&keyval, sizeof(int)); + *flag=1; + } + CATCH(ex) { + *flag=0; + xbt_ex_free(ex); + } + return MPI_SUCCESS; +} + +int smpi_comm_attr_put(MPI_Comm comm, int keyval, void* attr_value){ + if(!smpi_comm_keyvals) + smpi_comm_keyvals = xbt_dict_new(); + smpi_comm_key_elem elem = xbt_dict_get_or_null_ext(smpi_comm_keyvals, (const char*)&keyval, sizeof(int)); + if(!elem ) + return MPI_ERR_ARG; + int flag; + void* value; + smpi_comm_attr_get(comm, keyval, &value, &flag); + if(flag && elem->delete_fn!=MPI_NULL_DELETE_FN){ + int ret = elem->delete_fn(comm, keyval, value, &flag); + if(ret!=MPI_SUCCESS) return ret; + } + if(comm->attributes==NULL) + comm->attributes=xbt_dict_new(); + + xbt_dict_set_ext(comm->attributes, (const char*)&keyval, sizeof(int), attr_value, NULL); + return MPI_SUCCESS; +} + +int smpi_comm_keyval_create(MPI_Comm_copy_attr_function* copy_fn, MPI_Comm_delete_attr_function* delete_fn, int* keyval, void* extra_state){ + + if(!smpi_comm_keyvals) + smpi_comm_keyvals = xbt_dict_new(); + + smpi_comm_key_elem value = (smpi_comm_key_elem) xbt_new0(s_smpi_mpi_comm_key_elem_t,1); + + value->copy_fn=copy_fn; + value->delete_fn=delete_fn; + + *keyval = comm_keyval_id; + xbt_dict_set_ext(smpi_comm_keyvals, (const char*)keyval, sizeof(int),(void*)value, NULL); + comm_keyval_id++; + return MPI_SUCCESS; +} + +int smpi_comm_keyval_free(int* keyval){ + smpi_comm_key_elem elem = xbt_dict_get_or_null_ext(smpi_comm_keyvals, (const char*)keyval, sizeof(int)); + if(!elem){ + return MPI_ERR_ARG; + } + xbt_dict_remove_ext(smpi_comm_keyvals, (const char*)keyval, sizeof(int)); + xbt_free(elem); + return MPI_SUCCESS; +}