-/* Copyright (c) 2013-2019. The SimGrid Team.
+/* Copyright (c) 2013-2022. The SimGrid Team.
* All rights reserved. */
/* This program is free software; you can redistribute it and/or modify it
// call native when MPI communication size is too small
if (size <= num_core) {
- XBT_WARN("MPI_bcast_SMP_linear use default MPI_bcast.");
+ XBT_INFO("size <= num_core : MPI_bcast_SMP_linear use default MPI_bcast.");
bcast__default(buf, count, datatype, root, comm);
return MPI_SUCCESS;
}
}
// pipeline bcast
else {
- MPI_Request* request_array = new MPI_Request[size + pipe_length];
- MPI_Status* status_array = new MPI_Status[size + pipe_length];
+ auto* request_array = new MPI_Request[size + pipe_length];
+ auto* status_array = new MPI_Status[size + pipe_length];
// case ROOT of each SMP
if (rank % num_core == 0) {
delete[] status_array;
}
- // when count is not divisible by block size, use default BCAST for the remainder
if ((remainder != 0) && (count > segment)) {
- XBT_WARN("MPI_bcast_SMP_linear use default MPI_bcast.");
- Colls::bcast((char *) buf + (pipe_length * increment), remainder, datatype,
- root, comm);
+ XBT_INFO("MPI_bcast_SMP_linear: count is not divisible by block size, use default MPI_bcast for remainder.");
+ colls::bcast((char*)buf + (pipe_length * increment), remainder, datatype, root, comm);
}
return MPI_SUCCESS;