X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/a470be169558c63bccfb87761d482d90421d979a..262ef6e7f7f0eadfeba8293fd4d5925c75d834ed:/src/smpi/bindings/smpi_pmpi.cpp diff --git a/src/smpi/bindings/smpi_pmpi.cpp b/src/smpi/bindings/smpi_pmpi.cpp index 6b831d51df..99501be932 100644 --- a/src/smpi/bindings/smpi_pmpi.cpp +++ b/src/smpi/bindings/smpi_pmpi.cpp @@ -1,4 +1,4 @@ -/* Copyright (c) 2007-2019. The SimGrid Team. All rights reserved. */ +/* Copyright (c) 2007-2021. The SimGrid Team. All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it * under the terms of the license (GNU LGPL) which comes with this package. */ @@ -7,6 +7,7 @@ #include "simgrid/instr.h" #include "simgrid/s4u/Engine.hpp" #include "simgrid/s4u/Host.hpp" +#include "simgrid/version.h" #include "smpi_comm.hpp" #include "smpi_datatype_derived.hpp" #include "smpi_status.hpp" @@ -31,25 +32,29 @@ void TRACE_smpi_set_category(const char *category) /* PMPI User level calls */ -int PMPI_Init(int *argc, char ***argv) +int PMPI_Init(int*, char***) { xbt_assert(simgrid::s4u::Engine::is_initialized(), "Your MPI program was not properly initialized. The easiest is to use smpirun to start it."); - // Init is called only once per SMPI process - if (not smpi_process()->initializing()){ - simgrid::smpi::ActorExt::init(); + if(smpi_process()->initializing()){ + XBT_WARN("SMPI is already initializing - MPI_Init called twice ?"); + return MPI_ERR_OTHER; } - if (not smpi_process()->initialized()){ - int rank_traced = simgrid::s4u::this_actor::get_pid(); - TRACE_smpi_init(rank_traced); - TRACE_smpi_comm_in(rank_traced, __func__, new simgrid::instr::NoOpTIData("init")); - TRACE_smpi_comm_out(rank_traced); - TRACE_smpi_computing_init(rank_traced); - TRACE_smpi_sleeping_init(rank_traced); - smpi_bench_begin(); - smpi_process()->mark_as_initialized(); + if(smpi_process()->initialized()){ + XBT_WARN("SMPI already initialized once - MPI_Init called twice ?"); + return MPI_ERR_OTHER; } + if(smpi_process()->finalized()){ + XBT_WARN("SMPI already finalized"); + return MPI_ERR_OTHER; + } + + simgrid::smpi::ActorExt::init(); + int rank_traced = simgrid::s4u::this_actor::get_pid(); + TRACE_smpi_init(rank_traced, __func__); + smpi_bench_begin(); + smpi_process()->mark_as_initialized(); smpi_mpi_init(); @@ -65,7 +70,6 @@ int PMPI_Finalize() smpi_process()->finalize(); TRACE_smpi_comm_out(rank_traced); - TRACE_smpi_finalize(rank_traced); return MPI_SUCCESS; } @@ -82,11 +86,9 @@ int PMPI_Get_version (int *version,int *subversion){ } int PMPI_Get_library_version (char *version,int *len){ - smpi_bench_end(); - snprintf(version, MPI_MAX_LIBRARY_VERSION_STRING, "SMPI Version %d.%d. Copyright The Simgrid Team 2007-2019", + snprintf(version, MPI_MAX_LIBRARY_VERSION_STRING, "SMPI Version %d.%d. Copyright The SimGrid Team 2007-2021", SIMGRID_VERSION_MAJOR, SIMGRID_VERSION_MINOR); *len = strlen(version) > MPI_MAX_LIBRARY_VERSION_STRING ? MPI_MAX_LIBRARY_VERSION_STRING : strlen(version); - smpi_bench_begin(); return MPI_SUCCESS; } @@ -122,12 +124,19 @@ int PMPI_Is_thread_main(int *flag) } } -int PMPI_Abort(MPI_Comm /*comm*/, int /*errorcode*/) +int PMPI_Abort(MPI_Comm comm, int /*errorcode*/) { smpi_bench_end(); - // FIXME: should kill all processes in comm instead + CHECK_COMM(1) + XBT_WARN("MPI_Abort was called, something went probably wrong in this simulation ! Killing all processes sharing the same MPI_COMM_WORLD"); + for (int i = 0; i < comm->size(); i++){ + smx_actor_t actor = comm->group()->actor(i)->get_impl(); + if(actor != SIMIX_process_self()) + simgrid::kernel::actor::simcall([actor] { actor->exit(); }); + } + // now ourself smx_actor_t actor = SIMIX_process_self(); - simgrid::simix::simcall([actor] { actor->exit(); }); + simgrid::kernel::actor::simcall([actor] { actor->exit(); }); return MPI_SUCCESS; } @@ -142,7 +151,7 @@ double PMPI_Wtick() return sg_maxmin_precision; } -int PMPI_Address(void *location, MPI_Aint * address) +int PMPI_Address(const void* location, MPI_Aint* address) { if (address==nullptr) { return MPI_ERR_ARG; @@ -152,22 +161,34 @@ int PMPI_Address(void *location, MPI_Aint * address) } } -int PMPI_Get_address(void *location, MPI_Aint * address) +int PMPI_Get_address(const void *location, MPI_Aint * address) { return PMPI_Address(location, address); } +MPI_Aint PMPI_Aint_add(MPI_Aint address, MPI_Aint disp) +{ + xbt_assert(address <= PTRDIFF_MAX - disp, "overflow in MPI_Aint_add"); + return address + disp; +} + +MPI_Aint PMPI_Aint_diff(MPI_Aint address, MPI_Aint disp) +{ + xbt_assert(address >= PTRDIFF_MIN + disp, "underflow in MPI_Aint_diff"); + return address - disp; +} + int PMPI_Get_processor_name(char *name, int *resultlen) { - strncpy(name, sg_host_self()->get_cname(), - strlen(sg_host_self()->get_cname()) < MPI_MAX_PROCESSOR_NAME - 1 ? strlen(sg_host_self()->get_cname()) + 1 - : MPI_MAX_PROCESSOR_NAME - 1); - *resultlen = strlen(name) > MPI_MAX_PROCESSOR_NAME ? MPI_MAX_PROCESSOR_NAME : strlen(name); + int len = std::min(sg_host_self()->get_name().size(), MPI_MAX_PROCESSOR_NAME - 1); + std::string(sg_host_self()->get_name()).copy(name, len); + name[len] = '\0'; + *resultlen = len; return MPI_SUCCESS; } -int PMPI_Get_count(MPI_Status * status, MPI_Datatype datatype, int *count) +int PMPI_Get_count(const MPI_Status * status, MPI_Datatype datatype, int *count) { if (status == nullptr || count == nullptr) { return MPI_ERR_ARG; @@ -177,13 +198,12 @@ int PMPI_Get_count(MPI_Status * status, MPI_Datatype datatype, int *count) size_t size = datatype->size(); if (size == 0) { *count = 0; - return MPI_SUCCESS; } else if (status->count % size != 0) { - return MPI_UNDEFINED; + *count = MPI_UNDEFINED; } else { *count = simgrid::smpi::Status::get_count(status, datatype); - return MPI_SUCCESS; } + return MPI_SUCCESS; } } @@ -195,12 +215,8 @@ int PMPI_Initialized(int* flag) { int PMPI_Alloc_mem(MPI_Aint size, MPI_Info /*info*/, void* baseptr) { void *ptr = xbt_malloc(size); - if(ptr==nullptr) - return MPI_ERR_NO_MEM; - else { - *static_cast(baseptr) = ptr; - return MPI_SUCCESS; - } + *static_cast(baseptr) = ptr; + return MPI_SUCCESS; } int PMPI_Free_mem(void *baseptr){ @@ -214,17 +230,15 @@ int PMPI_Error_class(int errorcode, int* errorclass) { return MPI_SUCCESS; } -int PMPI_Error_string(int errorcode, char* string, int* resultlen){ - if (errorcode<0 || errorcode>= MPI_MAX_ERROR_STRING || string ==nullptr){ +int PMPI_Error_string(int errorcode, char* string, int* resultlen) +{ + static const std::vector smpi_error_string = {FOREACH_ERROR(GENERATE_STRING)}; + if (errorcode < 0 || static_cast(errorcode) >= smpi_error_string.size() || string == nullptr) return MPI_ERR_ARG; - } else { - static const char *smpi_error_string[] = { - FOREACH_ERROR(GENERATE_STRING) - }; - *resultlen = strlen(smpi_error_string[errorcode]); - strncpy(string, smpi_error_string[errorcode], *resultlen); - return MPI_SUCCESS; - } + + int len = snprintf(string, MPI_MAX_ERROR_STRING, "%s", smpi_error_string[errorcode]); + *resultlen = std::min(len, MPI_MAX_ERROR_STRING - 1); + return MPI_SUCCESS; } int PMPI_Keyval_create(MPI_Copy_function* copy_fn, MPI_Delete_function* delete_fn, int* keyval, void* extra_state) { @@ -234,5 +248,20 @@ int PMPI_Keyval_create(MPI_Copy_function* copy_fn, MPI_Delete_function* delete_f } int PMPI_Keyval_free(int* keyval) { + CHECK_NULL(1, MPI_ERR_ARG, keyval) + CHECK_MPI_NULL(1, MPI_KEYVAL_INVALID, MPI_ERR_KEYVAL, *keyval) return simgrid::smpi::Keyval::keyval_free(keyval); } + +int PMPI_Buffer_attach(void *buf, int size){ + if(buf==nullptr) + return MPI_ERR_BUFFER; + if(size<0) + return MPI_ERR_ARG; + return smpi_process()->set_bsend_buffer(buf, size); +} + +int PMPI_Buffer_detach(void* buffer, int* size){ + smpi_process()->bsend_buffer((void**)buffer, size); + return smpi_process()->set_bsend_buffer(nullptr, 0); +}