X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/b7ed19dfcc221d7b3eca182abb5c4a3946671172..39c935d6d5ee86d153f6f7e6a10d723ae7c57f6f:/src/smpi/colls/allgather/allgather-rhv.cpp diff --git a/src/smpi/colls/allgather/allgather-rhv.cpp b/src/smpi/colls/allgather/allgather-rhv.cpp index 12a929db11..fccc0fb290 100644 --- a/src/smpi/colls/allgather/allgather-rhv.cpp +++ b/src/smpi/colls/allgather/allgather-rhv.cpp @@ -1,10 +1,10 @@ -/* Copyright (c) 2013-2014. The SimGrid Team. +/* Copyright (c) 2013-2021. The SimGrid Team. * All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it * under the terms of the license (GNU LGPL) which comes with this package. */ -#include "../colls_private.h" +#include "../colls_private.hpp" namespace simgrid{ namespace smpi{ @@ -13,10 +13,10 @@ namespace smpi{ // now only work with power of two processes int -Coll_allgather_rhv::allgather(void *sbuf, int send_count, - MPI_Datatype send_type, void *rbuf, - int recv_count, MPI_Datatype recv_type, - MPI_Comm comm) +allgather__rhv(const void *sbuf, int send_count, + MPI_Datatype send_type, void *rbuf, + int recv_count, MPI_Datatype recv_type, + MPI_Comm comm) { MPI_Status status; MPI_Aint s_extent, r_extent; @@ -28,11 +28,11 @@ Coll_allgather_rhv::allgather(void *sbuf, int send_count, unsigned int mask; int curr_count; - // get size of the communicator, followed by rank + // get size of the communicator, followed by rank unsigned int num_procs = comm->size(); if((num_procs&(num_procs-1))) - THROWF(arg_error,0, "allgather rhv algorithm can't be used with non power of two number of processes ! "); + throw std::invalid_argument("allgather rhv algorithm can't be used with non power of two number of processes!"); unsigned int rank = comm->rank(); @@ -45,10 +45,10 @@ Coll_allgather_rhv::allgather(void *sbuf, int send_count, recv_chunk = r_extent * recv_count; if (send_chunk != recv_chunk) { - XBT_WARN("MPI_allgather_rhv use default MPI_allgather."); - Coll_allgather_default::allgather(sbuf, send_count, send_type, rbuf, recv_count, - recv_type, comm); - return MPI_SUCCESS; + XBT_WARN("MPI_allgather_rhv use default MPI_allgather."); + allgather__default(sbuf, send_count, send_type, rbuf, recv_count, + recv_type, comm); + return MPI_SUCCESS; } // compute starting offset location to perform local copy @@ -94,10 +94,8 @@ Coll_allgather_rhv::allgather(void *sbuf, int send_count, // printf("node %d send to %d in phase %d s_offset = %d r_offset = %d count = %d\n",rank,dst,phase, send_base_offset, recv_base_offset, curr_count); - Request::sendrecv((char *)rbuf + send_offset, curr_count, recv_type, dst, tag, - (char *)rbuf + recv_offset, curr_count, recv_type, dst, tag, - comm, &status); - + Request::sendrecv((char*)rbuf + send_offset, curr_count, recv_type, dst, tag, (char*)rbuf + recv_offset, curr_count, + recv_type, dst, tag, comm, &status); curr_count *= 2; i *= 2;