X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/13a775d458d6616227bfe215af742411850677dc..4c753f8d4cabd4104f3f7109823f16be2ebdcce3:/src/smpi/bindings/smpi_pmpi.cpp?ds=sidebyside diff --git a/src/smpi/bindings/smpi_pmpi.cpp b/src/smpi/bindings/smpi_pmpi.cpp index c05c46d063..684b283a86 100644 --- a/src/smpi/bindings/smpi_pmpi.cpp +++ b/src/smpi/bindings/smpi_pmpi.cpp @@ -1,15 +1,19 @@ -/* Copyright (c) 2007-2019. The SimGrid Team. All rights reserved. */ +/* Copyright (c) 2007-2022. The SimGrid Team. All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it * under the terms of the license (GNU LGPL) which comes with this package. */ #include "private.hpp" +#include "simgrid/host.h" #include "simgrid/instr.h" #include "simgrid/s4u/Engine.hpp" #include "simgrid/s4u/Host.hpp" +#include "simgrid/version.h" +#include "smpi_coll.hpp" #include "smpi_comm.hpp" #include "smpi_datatype_derived.hpp" #include "smpi_status.hpp" +#include "src/kernel/EngineImpl.hpp" #include "src/kernel/actor/ActorImpl.hpp" #include "src/smpi/include/smpi_actor.hpp" @@ -19,14 +23,13 @@ XBT_LOG_NEW_DEFAULT_SUBCATEGORY(smpi_pmpi, smpi, "Logging specific to SMPI (pmpi void TRACE_smpi_set_category(const char *category) { //need to end bench otherwise categories for execution tasks are wrong - smpi_bench_end(); + const SmpiBenchGuard suspend_bench; + if (category != nullptr) { // declare category - TRACE_category(category); + simgrid::instr::declare_tracing_category(category); smpi_process()->set_tracing_category(category); } - //begin bench after changing process's category - smpi_bench_begin(); } /* PMPI User level calls */ @@ -36,20 +39,26 @@ int PMPI_Init(int*, char***) xbt_assert(simgrid::s4u::Engine::is_initialized(), "Your MPI program was not properly initialized. The easiest is to use smpirun to start it."); - xbt_assert(not smpi_process()->initializing()); - xbt_assert(not smpi_process()->initialized()); + if(smpi_process()->initializing()){ + XBT_WARN("SMPI is already initializing - MPI_Init called twice ?"); + return MPI_ERR_OTHER; + } + if(smpi_process()->initialized()){ + XBT_WARN("SMPI already initialized once - MPI_Init called twice ?"); + return MPI_ERR_OTHER; + } + if(smpi_process()->finalized()){ + XBT_WARN("SMPI already finalized"); + return MPI_ERR_OTHER; + } simgrid::smpi::ActorExt::init(); - int rank_traced = simgrid::s4u::this_actor::get_pid(); - TRACE_smpi_init(rank_traced); - TRACE_smpi_comm_in(rank_traced, __func__, new simgrid::instr::NoOpTIData("init")); - TRACE_smpi_comm_out(rank_traced); - TRACE_smpi_computing_init(rank_traced); - TRACE_smpi_sleeping_init(rank_traced); + TRACE_smpi_init(simgrid::s4u::this_actor::get_pid(), __func__); + smpi_mpi_init(); smpi_bench_begin(); smpi_process()->mark_as_initialized(); - smpi_mpi_init(); + CHECK_COLLECTIVE(smpi_process()->comm_world(), "MPI_Init") return MPI_SUCCESS; } @@ -57,9 +66,14 @@ int PMPI_Init(int*, char***) int PMPI_Finalize() { smpi_bench_end(); - int rank_traced = simgrid::s4u::this_actor::get_pid(); + CHECK_COLLECTIVE(smpi_process()->comm_world(), "MPI_Finalize") + aid_t rank_traced = simgrid::s4u::this_actor::get_pid(); + smpi_process()->mark_as_finalizing(); TRACE_smpi_comm_in(rank_traced, __func__, new simgrid::instr::NoOpTIData("finalize")); + if (simgrid::config::get_value("smpi/barrier-finalization")) + simgrid::smpi::colls::barrier(MPI_COMM_WORLD); + smpi_process()->finalize(); TRACE_smpi_comm_out(rank_traced); @@ -79,11 +93,9 @@ int PMPI_Get_version (int *version,int *subversion){ } int PMPI_Get_library_version (char *version,int *len){ - smpi_bench_end(); - snprintf(version, MPI_MAX_LIBRARY_VERSION_STRING, "SMPI Version %d.%d. Copyright The Simgrid Team 2007-2019", + snprintf(version, MPI_MAX_LIBRARY_VERSION_STRING, "SMPI Version %d.%d. Copyright The SimGrid Team 2007-2022", SIMGRID_VERSION_MAJOR, SIMGRID_VERSION_MINOR); - *len = strlen(version) > MPI_MAX_LIBRARY_VERSION_STRING ? MPI_MAX_LIBRARY_VERSION_STRING : strlen(version); - smpi_bench_begin(); + *len = std::min(static_cast(strlen(version)), MPI_MAX_LIBRARY_VERSION_STRING); return MPI_SUCCESS; } @@ -119,12 +131,19 @@ int PMPI_Is_thread_main(int *flag) } } -int PMPI_Abort(MPI_Comm /*comm*/, int /*errorcode*/) +int PMPI_Abort(MPI_Comm comm, int /*errorcode*/) { smpi_bench_end(); - // FIXME: should kill all processes in comm instead - smx_actor_t actor = SIMIX_process_self(); - simgrid::simix::simcall([actor] { actor->exit(); }); + CHECK_COMM(1) + XBT_WARN("MPI_Abort was called, something went probably wrong in this simulation ! Killing all processes sharing the same MPI_COMM_WORLD"); + auto myself = simgrid::kernel::actor::ActorImpl::self(); + for (int i = 0; i < comm->size(); i++){ + auto actor = simgrid::kernel::EngineImpl::get_instance()->get_actor_by_pid(comm->group()->actor(i)); + if (actor != nullptr && actor != myself) + simgrid::kernel::actor::simcall_answered([actor] { actor->exit(); }); + } + // now ourself + simgrid::kernel::actor::simcall_answered([myself] { myself->exit(); }); return MPI_SUCCESS; } @@ -154,12 +173,24 @@ int PMPI_Get_address(const void *location, MPI_Aint * address) return PMPI_Address(location, address); } +MPI_Aint PMPI_Aint_add(MPI_Aint address, MPI_Aint disp) +{ + xbt_assert(address <= PTRDIFF_MAX - disp, "overflow in MPI_Aint_add"); + return address + disp; +} + +MPI_Aint PMPI_Aint_diff(MPI_Aint address, MPI_Aint disp) +{ + xbt_assert(address >= PTRDIFF_MIN + disp, "underflow in MPI_Aint_diff"); + return address - disp; +} + int PMPI_Get_processor_name(char *name, int *resultlen) { - strncpy(name, sg_host_self()->get_cname(), - strlen(sg_host_self()->get_cname()) < MPI_MAX_PROCESSOR_NAME - 1 ? strlen(sg_host_self()->get_cname()) + 1 - : MPI_MAX_PROCESSOR_NAME - 1); - *resultlen = strlen(name) > MPI_MAX_PROCESSOR_NAME ? MPI_MAX_PROCESSOR_NAME : strlen(name); + int len = std::min(static_cast(sg_host_self()->get_name().size()), MPI_MAX_PROCESSOR_NAME - 1); + sg_host_self()->get_name().copy(name, len); + name[len] = '\0'; + *resultlen = len; return MPI_SUCCESS; } @@ -171,16 +202,14 @@ int PMPI_Get_count(const MPI_Status * status, MPI_Datatype datatype, int *count) } else if (not datatype->is_valid()) { return MPI_ERR_TYPE; } else { - size_t size = datatype->size(); - if (size == 0) { + if (datatype->size() == 0) { *count = 0; - return MPI_SUCCESS; - } else if (status->count % size != 0) { - return MPI_UNDEFINED; + } else if (status->count % datatype->size() != 0) { + *count = MPI_UNDEFINED; } else { *count = simgrid::smpi::Status::get_count(status, datatype); - return MPI_SUCCESS; } + return MPI_SUCCESS; } } @@ -191,13 +220,10 @@ int PMPI_Initialized(int* flag) { int PMPI_Alloc_mem(MPI_Aint size, MPI_Info /*info*/, void* baseptr) { + CHECK_NEGATIVE(1, MPI_ERR_COUNT, size) void *ptr = xbt_malloc(size); - if(ptr==nullptr) - return MPI_ERR_NO_MEM; - else { - *static_cast(baseptr) = ptr; - return MPI_SUCCESS; - } + *static_cast(baseptr) = ptr; + return MPI_SUCCESS; } int PMPI_Free_mem(void *baseptr){ @@ -213,9 +239,8 @@ int PMPI_Error_class(int errorcode, int* errorclass) { int PMPI_Error_string(int errorcode, char* string, int* resultlen) { - static const char* smpi_error_string[] = {FOREACH_ERROR(GENERATE_STRING)}; - constexpr int nerrors = (sizeof smpi_error_string) / (sizeof smpi_error_string[0]); - if (errorcode < 0 || errorcode >= nerrors || string == nullptr) + static const std::vector smpi_error_string = {FOREACH_ERROR(GENERATE_STRING)}; + if (errorcode < 0 || static_cast(errorcode) >= smpi_error_string.size() || string == nullptr) return MPI_ERR_ARG; int len = snprintf(string, MPI_MAX_ERROR_STRING, "%s", smpi_error_string[errorcode]); @@ -230,5 +255,20 @@ int PMPI_Keyval_create(MPI_Copy_function* copy_fn, MPI_Delete_function* delete_f } int PMPI_Keyval_free(int* keyval) { + CHECK_NULL(1, MPI_ERR_ARG, keyval) + CHECK_VAL(1, MPI_KEYVAL_INVALID, MPI_ERR_KEYVAL, *keyval) return simgrid::smpi::Keyval::keyval_free(keyval); } + +int PMPI_Buffer_attach(void *buf, int size){ + if(buf==nullptr) + return MPI_ERR_BUFFER; + if(size<0) + return MPI_ERR_ARG; + return smpi_process()->set_bsend_buffer(buf, size); +} + +int PMPI_Buffer_detach(void* buffer, int* size){ + smpi_process()->bsend_buffer((void**)buffer, size); + return smpi_process()->set_bsend_buffer(nullptr, 0); +}