Logo AND Algorithmique Numérique Distribuée

Public GIT Repository
factorize
[simgrid.git] / src / smpi / colls / smpi_coll.cpp
index 3606038..e13f8f9 100644 (file)
@@ -1,6 +1,6 @@
 /* smpi_coll.c -- various optimized routing for collectives                 */
 
-/* Copyright (c) 2009-2017. The SimGrid Team. All rights reserved.          */
+/* Copyright (c) 2009-2019. The SimGrid Team. All rights reserved.          */
 
 /* This program is free software; you can redistribute it and/or modify it
  * under the terms of the license (GNU LGPL) which comes with this package. */
 #include "smpi_datatype.hpp"
 #include "smpi_op.hpp"
 #include "smpi_request.hpp"
+#include "xbt/config.hpp"
 
 XBT_LOG_NEW_DEFAULT_SUBCATEGORY(smpi_coll, smpi, "Logging specific to SMPI (coll)");
 
 #define COLL_SETTER(cat, ret, args, args2)                                                                             \
   int(*Colls::cat) args;                                                                                               \
-  void Colls::set_##cat(std::string name)                                                                              \
+  void Colls::set_##cat(const std::string& name)                                                                       \
   {                                                                                                                    \
     int id = find_coll_description(mpi_coll_##cat##_description, name, #cat);                                          \
     cat    = reinterpret_cast<ret(*) args>(mpi_coll_##cat##_description[id].coll);                                     \
@@ -24,12 +25,6 @@ XBT_LOG_NEW_DEFAULT_SUBCATEGORY(smpi_coll, smpi, "Logging specific to SMPI (coll
       xbt_die("Collective " #cat " set to nullptr!");                                                                  \
   }
 
-#define SET_COLL(coll)                                                                                                 \
-  name = xbt_cfg_get_string("smpi/" #coll);                                                                            \
-  if (name.empty())                                                                                                    \
-    name = selector_name;                                                                                              \
-  set_##coll(name);
-
 namespace simgrid{
 namespace smpi{
 
@@ -37,49 +32,49 @@ void (*Colls::smpi_coll_cleanup_callback)();
 
 /* these arrays must be nullptr terminated */
 s_mpi_coll_description_t Colls::mpi_coll_gather_description[] = {
-    COLL_GATHERS(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_GATHERS(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_allgather_description[] = {
-    COLL_ALLGATHERS(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_ALLGATHERS(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_allgatherv_description[] = {
-    COLL_ALLGATHERVS(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_ALLGATHERVS(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_allreduce_description[] ={
-    COLL_ALLREDUCES(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_ALLREDUCES(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_reduce_scatter_description[] = {
-    COLL_REDUCE_SCATTERS(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_REDUCE_SCATTERS(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_scatter_description[] ={
-    COLL_SCATTERS(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_SCATTERS(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_barrier_description[] ={
-    COLL_BARRIERS(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_BARRIERS(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_alltoall_description[] = {
-    COLL_ALLTOALLS(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_ALLTOALLS(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_alltoallv_description[] = {
-    COLL_ALLTOALLVS(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_ALLTOALLVS(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_bcast_description[] = {
-    COLL_BCASTS(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_BCASTS(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 s_mpi_coll_description_t Colls::mpi_coll_reduce_description[] = {
-    COLL_REDUCES(COLL_DESCRIPTION, COLL_COMMA), {nullptr, nullptr, nullptr} };
+    COLL_REDUCES(COLL_DESCRIPTION, COLL_COMMA), {"", "", nullptr} };
 
 /** Displays the long description of all registered models, and quit */
 void Colls::coll_help(const char *category, s_mpi_coll_description_t * table)
 {
   XBT_WARN("Long description of the %s models accepted by this simulator:\n", category);
-  for (int i = 0; table[i].name; i++)
-    XBT_WARN("  %s: %s\n", table[i].name, table[i].description);
+  for (int i = 0; not table[i].name.empty(); i++)
+    XBT_WARN("  %s: %s\n", table[i].name.c_str(), table[i].description.c_str());
 }
 
-int Colls::find_coll_description(s_mpi_coll_description_t* table, std::string name, const char* desc)
+int Colls::find_coll_description(s_mpi_coll_description_t* table, const std::string& name, const char* desc)
 {
-  for (int i = 0; table[i].name; i++)
+  for (int i = 0; not table[i].name.empty(); i++)
     if (name == table[i].name) {
-      if (strcmp(table[i].name,"default"))
-        XBT_INFO("Switch to algorithm %s for collective %s",table[i].name,desc);
+      if (table[i].name != "default")
+        XBT_INFO("Switch to algorithm %s for collective %s",table[i].name.c_str(),desc);
       return i;
     }
 
-  if (not table[0].name)
+  if (table[0].name.empty())
     xbt_die("No collective is valid for '%s'! This is a bug.", name.c_str());
-  std::string name_list = std::string(table[0].name);
-  for (int i = 1; table[i].name; i++)
+  std::string name_list = table[0].name;
+  for (int i = 1; not table[i].name.empty(); i++)
     name_list = name_list + ", " + table[i].name;
 
   xbt_die("Collective '%s' is invalid! Valid collectives are: %s.", name.c_str(), name_list.c_str());
@@ -98,107 +93,58 @@ COLL_APPLY(COLL_SETTER,COLL_BCAST_SIG,"");
 COLL_APPLY(COLL_SETTER,COLL_ALLTOALL_SIG,"");
 COLL_APPLY(COLL_SETTER,COLL_ALLTOALLV_SIG,"");
 
-
 void Colls::set_collectives(){
-  std::string selector_name = xbt_cfg_get_string("smpi/coll-selector");
+  std::string selector_name = simgrid::config::get_value<std::string>("smpi/coll-selector");
   if (selector_name.empty())
     selector_name = "default";
 
-  std::string name;
-
-  SET_COLL(gather);
-  SET_COLL(allgather);
-  SET_COLL(allgatherv);
-  SET_COLL(allreduce);
-  SET_COLL(alltoall);
-  SET_COLL(alltoallv);
-  SET_COLL(reduce);
-  SET_COLL(reduce_scatter);
-  SET_COLL(scatter);
-  SET_COLL(bcast);
-  SET_COLL(barrier);
+  std::pair<std::string, std::function<void(std::string)>> setter_callbacks[] = {
+      {"gather", &Colls::set_gather},         {"allgather", &Colls::set_allgather},
+      {"allgatherv", &Colls::set_allgatherv}, {"allreduce", &Colls::set_allreduce},
+      {"alltoall", &Colls::set_alltoall},     {"alltoallv", &Colls::set_alltoallv},
+      {"reduce", &Colls::set_reduce},         {"reduce_scatter", &Colls::set_reduce_scatter},
+      {"scatter", &Colls::set_scatter},       {"bcast", &Colls::set_bcast},
+      {"barrier", &Colls::set_barrier}};
+
+  for (auto& elem : setter_callbacks) {
+    std::string name = simgrid::config::get_value<std::string>(("smpi/" + elem.first).c_str());
+    if (name.empty())
+      name = selector_name;
+
+    (elem.second)(name);
+  }
 }
 
+int Colls::finish_nbc_request(MPI_Request request){
+  MPI_Request* requests = request->get_nbc_requests();
+  int count = request->get_nbc_requests_size();
+  Request::waitall(count, requests, MPI_STATUS_IGNORE);
+  for (int i = 0; i < count; i++) {
+    if(requests[i]!=MPI_REQUEST_NULL)
+      Request::unref(&requests[i]);
+  }
+  delete[] requests;
+  Request::unref(&request);
+  return MPI_SUCCESS;
+}
 
 //Implementations of the single algorith collectives
 
 int Colls::gatherv(void *sendbuf, int sendcount, MPI_Datatype sendtype, void *recvbuf, int *recvcounts, int *displs,
                       MPI_Datatype recvtype, int root, MPI_Comm comm)
 {
-  int system_tag = COLL_TAG_GATHERV;
-  MPI_Aint lb = 0;
-  MPI_Aint recvext = 0;
-
-  int rank = comm->rank();
-  int size = comm->size();
-  if (rank != root) {
-    // Send buffer to root
-    Request::send(sendbuf, sendcount, sendtype, root, system_tag, comm);
-  } else {
-    recvtype->extent(&lb, &recvext);
-    // Local copy from root
-    Datatype::copy(sendbuf, sendcount, sendtype, static_cast<char*>(recvbuf) + displs[root] * recvext,
-                       recvcounts[root], recvtype);
-    // Receive buffers from senders
-    MPI_Request *requests = xbt_new(MPI_Request, size - 1);
-    int index = 0;
-    for (int src = 0; src < size; src++) {
-      if(src != root) {
-        requests[index] = Request::irecv_init(static_cast<char*>(recvbuf) + displs[src] * recvext,
-                          recvcounts[src], recvtype, src, system_tag, comm);
-        index++;
-      }
-    }
-    // Wait for completion of irecv's.
-    Request::startall(size - 1, requests);
-    Request::waitall(size - 1, requests, MPI_STATUS_IGNORE);
-    for (int src = 0; src < size-1; src++) {
-      Request::unref(&requests[src]);
-    }
-    xbt_free(requests);
-  }
-  return MPI_SUCCESS;
+  MPI_Request request;
+  Colls::igatherv(sendbuf, sendcount, sendtype, recvbuf, recvcounts, displs, recvtype, root, comm, &request);
+  return Colls::finish_nbc_request(request);
 }
 
 
 int Colls::scatterv(void *sendbuf, int *sendcounts, int *displs, MPI_Datatype sendtype, void *recvbuf, int recvcount,
                        MPI_Datatype recvtype, int root, MPI_Comm comm)
 {
-  int system_tag = COLL_TAG_SCATTERV;
-  MPI_Aint lb = 0;
-  MPI_Aint sendext = 0;
-
-  int rank = comm->rank();
-  int size = comm->size();
-  if(rank != root) {
-    // Recv buffer from root
-    Request::recv(recvbuf, recvcount, recvtype, root, system_tag, comm, MPI_STATUS_IGNORE);
-  } else {
-    sendtype->extent(&lb, &sendext);
-    // Local copy from root
-    if(recvbuf!=MPI_IN_PLACE){
-      Datatype::copy(static_cast<char *>(sendbuf) + displs[root] * sendext, sendcounts[root],
-                       sendtype, recvbuf, recvcount, recvtype);
-    }
-    // Send buffers to receivers
-    MPI_Request *requests = xbt_new(MPI_Request, size - 1);
-    int index = 0;
-    for (int dst = 0; dst < size; dst++) {
-      if (dst != root) {
-        requests[index] = Request::isend_init(static_cast<char *>(sendbuf) + displs[dst] * sendext, sendcounts[dst],
-                            sendtype, dst, system_tag, comm);
-        index++;
-      }
-    }
-    // Wait for completion of isend's.
-    Request::startall(size - 1, requests);
-    Request::waitall(size - 1, requests, MPI_STATUS_IGNORE);
-    for (int dst = 0; dst < size-1; dst++) {
-      Request::unref(&requests[dst]);
-    }
-    xbt_free(requests);
-  }
-  return MPI_SUCCESS;
+  MPI_Request request;
+  Colls::iscatterv(sendbuf, sendcounts, displs, sendtype, recvbuf, recvcount, recvtype, root, comm, &request);
+  return Colls::finish_nbc_request(request);
 }
 
 
@@ -330,5 +276,13 @@ int Colls::exscan(void *sendbuf, void *recvbuf, int count, MPI_Datatype datatype
   return MPI_SUCCESS;
 }
 
+int Colls::alltoallw(void *sendbuf, int *sendcounts, int *senddisps, MPI_Datatype* sendtypes,
+                              void *recvbuf, int *recvcounts, int *recvdisps, MPI_Datatype* recvtypes, MPI_Comm comm)
+{
+  MPI_Request request;
+  Colls::ialltoallw(sendbuf, sendcounts, senddisps, sendtypes, recvbuf, recvcounts, recvdisps, recvtypes, comm, &request);
+  return Colls::finish_nbc_request(request);
+}
+
 }
 }