X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/e33beb98eb2dd42717306a77a328c9d5e9a91f05..4c753f8d4cabd4104f3f7109823f16be2ebdcce3:/src/smpi/bindings/smpi_pmpi.cpp diff --git a/src/smpi/bindings/smpi_pmpi.cpp b/src/smpi/bindings/smpi_pmpi.cpp index 76da5cd4c3..684b283a86 100644 --- a/src/smpi/bindings/smpi_pmpi.cpp +++ b/src/smpi/bindings/smpi_pmpi.cpp @@ -1,15 +1,20 @@ -/* Copyright (c) 2007-2018. The SimGrid Team. All rights reserved. */ +/* Copyright (c) 2007-2022. The SimGrid Team. All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it * under the terms of the license (GNU LGPL) which comes with this package. */ #include "private.hpp" +#include "simgrid/host.h" +#include "simgrid/instr.h" #include "simgrid/s4u/Engine.hpp" #include "simgrid/s4u/Host.hpp" +#include "simgrid/version.h" +#include "smpi_coll.hpp" #include "smpi_comm.hpp" #include "smpi_datatype_derived.hpp" #include "smpi_status.hpp" -#include "src/simix/ActorImpl.hpp" +#include "src/kernel/EngineImpl.hpp" +#include "src/kernel/actor/ActorImpl.hpp" #include "src/smpi/include/smpi_actor.hpp" XBT_LOG_NEW_DEFAULT_SUBCATEGORY(smpi_pmpi, smpi, "Logging specific to SMPI (pmpi)"); @@ -18,36 +23,42 @@ XBT_LOG_NEW_DEFAULT_SUBCATEGORY(smpi_pmpi, smpi, "Logging specific to SMPI (pmpi void TRACE_smpi_set_category(const char *category) { //need to end bench otherwise categories for execution tasks are wrong - smpi_bench_end(); - if (category != nullptr) - TRACE_internal_smpi_set_category(category); - //begin bench after changing process's category - smpi_bench_begin(); + const SmpiBenchGuard suspend_bench; + + if (category != nullptr) { + // declare category + simgrid::instr::declare_tracing_category(category); + smpi_process()->set_tracing_category(category); + } } /* PMPI User level calls */ -int PMPI_Init(int *argc, char ***argv) +int PMPI_Init(int*, char***) { xbt_assert(simgrid::s4u::Engine::is_initialized(), "Your MPI program was not properly initialized. The easiest is to use smpirun to start it."); - // Init is called only once per SMPI process - if (not smpi_process()->initializing()){ - simgrid::smpi::ActorExt::init(); + if(smpi_process()->initializing()){ + XBT_WARN("SMPI is already initializing - MPI_Init called twice ?"); + return MPI_ERR_OTHER; } - if (not smpi_process()->initialized()){ - int rank_traced = simgrid::s4u::this_actor::get_pid(); - TRACE_smpi_init(rank_traced); - TRACE_smpi_comm_in(rank_traced, __func__, new simgrid::instr::NoOpTIData("init")); - TRACE_smpi_comm_out(rank_traced); - TRACE_smpi_computing_init(rank_traced); - TRACE_smpi_sleeping_init(rank_traced); - smpi_bench_begin(); - smpi_process()->mark_as_initialized(); + if(smpi_process()->initialized()){ + XBT_WARN("SMPI already initialized once - MPI_Init called twice ?"); + return MPI_ERR_OTHER; + } + if(smpi_process()->finalized()){ + XBT_WARN("SMPI already finalized"); + return MPI_ERR_OTHER; } + simgrid::smpi::ActorExt::init(); + TRACE_smpi_init(simgrid::s4u::this_actor::get_pid(), __func__); smpi_mpi_init(); + smpi_bench_begin(); + smpi_process()->mark_as_initialized(); + + CHECK_COLLECTIVE(smpi_process()->comm_world(), "MPI_Init") return MPI_SUCCESS; } @@ -55,13 +66,17 @@ int PMPI_Init(int *argc, char ***argv) int PMPI_Finalize() { smpi_bench_end(); - int rank_traced = simgrid::s4u::this_actor::get_pid(); + CHECK_COLLECTIVE(smpi_process()->comm_world(), "MPI_Finalize") + aid_t rank_traced = simgrid::s4u::this_actor::get_pid(); + smpi_process()->mark_as_finalizing(); TRACE_smpi_comm_in(rank_traced, __func__, new simgrid::instr::NoOpTIData("finalize")); + if (simgrid::config::get_value("smpi/barrier-finalization")) + simgrid::smpi::colls::barrier(MPI_COMM_WORLD); + smpi_process()->finalize(); TRACE_smpi_comm_out(rank_traced); - TRACE_smpi_finalize(rank_traced); return MPI_SUCCESS; } @@ -78,11 +93,9 @@ int PMPI_Get_version (int *version,int *subversion){ } int PMPI_Get_library_version (char *version,int *len){ - smpi_bench_end(); - snprintf(version, MPI_MAX_LIBRARY_VERSION_STRING, "SMPI Version %d.%d. Copyright The Simgrid Team 2007-2018", + snprintf(version, MPI_MAX_LIBRARY_VERSION_STRING, "SMPI Version %d.%d. Copyright The SimGrid Team 2007-2022", SIMGRID_VERSION_MAJOR, SIMGRID_VERSION_MINOR); - *len = strlen(version) > MPI_MAX_LIBRARY_VERSION_STRING ? MPI_MAX_LIBRARY_VERSION_STRING : strlen(version); - smpi_bench_begin(); + *len = std::min(static_cast(strlen(version)), MPI_MAX_LIBRARY_VERSION_STRING); return MPI_SUCCESS; } @@ -118,12 +131,19 @@ int PMPI_Is_thread_main(int *flag) } } -int PMPI_Abort(MPI_Comm /*comm*/, int /*errorcode*/) +int PMPI_Abort(MPI_Comm comm, int /*errorcode*/) { smpi_bench_end(); - // FIXME: should kill all processes in comm instead - smx_actor_t process = SIMIX_process_self(); - simgrid::simix::simcall([process] { SIMIX_process_kill(process, process); }); + CHECK_COMM(1) + XBT_WARN("MPI_Abort was called, something went probably wrong in this simulation ! Killing all processes sharing the same MPI_COMM_WORLD"); + auto myself = simgrid::kernel::actor::ActorImpl::self(); + for (int i = 0; i < comm->size(); i++){ + auto actor = simgrid::kernel::EngineImpl::get_instance()->get_actor_by_pid(comm->group()->actor(i)); + if (actor != nullptr && actor != myself) + simgrid::kernel::actor::simcall_answered([actor] { actor->exit(); }); + } + // now ourself + simgrid::kernel::actor::simcall_answered([myself] { myself->exit(); }); return MPI_SUCCESS; } @@ -138,7 +158,7 @@ double PMPI_Wtick() return sg_maxmin_precision; } -int PMPI_Address(void *location, MPI_Aint * address) +int PMPI_Address(const void* location, MPI_Aint* address) { if (address==nullptr) { return MPI_ERR_ARG; @@ -148,38 +168,48 @@ int PMPI_Address(void *location, MPI_Aint * address) } } -int PMPI_Get_address(void *location, MPI_Aint * address) +int PMPI_Get_address(const void *location, MPI_Aint * address) { return PMPI_Address(location, address); } +MPI_Aint PMPI_Aint_add(MPI_Aint address, MPI_Aint disp) +{ + xbt_assert(address <= PTRDIFF_MAX - disp, "overflow in MPI_Aint_add"); + return address + disp; +} + +MPI_Aint PMPI_Aint_diff(MPI_Aint address, MPI_Aint disp) +{ + xbt_assert(address >= PTRDIFF_MIN + disp, "underflow in MPI_Aint_diff"); + return address - disp; +} + int PMPI_Get_processor_name(char *name, int *resultlen) { - strncpy(name, sg_host_self()->get_cname(), - strlen(sg_host_self()->get_cname()) < MPI_MAX_PROCESSOR_NAME - 1 ? strlen(sg_host_self()->get_cname()) + 1 - : MPI_MAX_PROCESSOR_NAME - 1); - *resultlen = strlen(name) > MPI_MAX_PROCESSOR_NAME ? MPI_MAX_PROCESSOR_NAME : strlen(name); + int len = std::min(static_cast(sg_host_self()->get_name().size()), MPI_MAX_PROCESSOR_NAME - 1); + sg_host_self()->get_name().copy(name, len); + name[len] = '\0'; + *resultlen = len; return MPI_SUCCESS; } -int PMPI_Get_count(MPI_Status * status, MPI_Datatype datatype, int *count) +int PMPI_Get_count(const MPI_Status * status, MPI_Datatype datatype, int *count) { if (status == nullptr || count == nullptr) { return MPI_ERR_ARG; } else if (not datatype->is_valid()) { return MPI_ERR_TYPE; } else { - size_t size = datatype->size(); - if (size == 0) { + if (datatype->size() == 0) { *count = 0; - return MPI_SUCCESS; - } else if (status->count % size != 0) { - return MPI_UNDEFINED; + } else if (status->count % datatype->size() != 0) { + *count = MPI_UNDEFINED; } else { *count = simgrid::smpi::Status::get_count(status, datatype); - return MPI_SUCCESS; } + return MPI_SUCCESS; } } @@ -190,13 +220,10 @@ int PMPI_Initialized(int* flag) { int PMPI_Alloc_mem(MPI_Aint size, MPI_Info /*info*/, void* baseptr) { + CHECK_NEGATIVE(1, MPI_ERR_COUNT, size) void *ptr = xbt_malloc(size); - if(ptr==nullptr) - return MPI_ERR_NO_MEM; - else { - *static_cast(baseptr) = ptr; - return MPI_SUCCESS; - } + *static_cast(baseptr) = ptr; + return MPI_SUCCESS; } int PMPI_Free_mem(void *baseptr){ @@ -210,17 +237,15 @@ int PMPI_Error_class(int errorcode, int* errorclass) { return MPI_SUCCESS; } -int PMPI_Error_string(int errorcode, char* string, int* resultlen){ - if (errorcode<0 || string ==nullptr){ +int PMPI_Error_string(int errorcode, char* string, int* resultlen) +{ + static const std::vector smpi_error_string = {FOREACH_ERROR(GENERATE_STRING)}; + if (errorcode < 0 || static_cast(errorcode) >= smpi_error_string.size() || string == nullptr) return MPI_ERR_ARG; - } else { - static const char *smpi_error_string[] = { - FOREACH_ERROR(GENERATE_STRING) - }; - *resultlen = strlen(smpi_error_string[errorcode]); - strncpy(string, smpi_error_string[errorcode], *resultlen); - return MPI_SUCCESS; - } + + int len = snprintf(string, MPI_MAX_ERROR_STRING, "%s", smpi_error_string[errorcode]); + *resultlen = std::min(len, MPI_MAX_ERROR_STRING - 1); + return MPI_SUCCESS; } int PMPI_Keyval_create(MPI_Copy_function* copy_fn, MPI_Delete_function* delete_fn, int* keyval, void* extra_state) { @@ -230,5 +255,20 @@ int PMPI_Keyval_create(MPI_Copy_function* copy_fn, MPI_Delete_function* delete_f } int PMPI_Keyval_free(int* keyval) { + CHECK_NULL(1, MPI_ERR_ARG, keyval) + CHECK_VAL(1, MPI_KEYVAL_INVALID, MPI_ERR_KEYVAL, *keyval) return simgrid::smpi::Keyval::keyval_free(keyval); } + +int PMPI_Buffer_attach(void *buf, int size){ + if(buf==nullptr) + return MPI_ERR_BUFFER; + if(size<0) + return MPI_ERR_ARG; + return smpi_process()->set_bsend_buffer(buf, size); +} + +int PMPI_Buffer_detach(void* buffer, int* size){ + smpi_process()->bsend_buffer((void**)buffer, size); + return smpi_process()->set_bsend_buffer(nullptr, 0); +}