-/* Copyright (c) 2013-2014. The SimGrid Team.
+/* Copyright (c) 2013-2020. The SimGrid Team.
* All rights reserved. */
/* This program is free software; you can redistribute it and/or modify it
* under the terms of the license (GNU LGPL) which comes with this package. */
-#include "../colls_private.h"
+#include "../colls_private.hpp"
static int bcast_NTSL_segment_size_in_byte = 8192;
namespace simgrid{
namespace smpi{
/* Non-topology-specific pipelined linear-bcast function */
-int Coll_bcast_arrival_pattern_aware::bcast(void *buf, int count,
- MPI_Datatype datatype, int root,
- MPI_Comm comm)
+int bcast__arrival_pattern_aware(void *buf, int count,
+ MPI_Datatype datatype, int root,
+ MPI_Comm comm)
{
int tag = -COLL_TAG_BCAST;
MPI_Status status;
MPI_Request request;
- MPI_Request *send_request_array;
- MPI_Request *recv_request_array;
- MPI_Status *send_status_array;
- MPI_Status *recv_status_array;
MPI_Status temp_status_array[MAX_NODE];
/* segment is segment size in number of elements (not bytes) */
int segment = bcast_NTSL_segment_size_in_byte / extent;
- segment = segment == 0 ? 1 :segment;
+ segment = segment == 0 ? 1 :segment;
/* pipeline length */
int pipe_length = count / segment;
/* use for buffer offset for sending and receiving data = segment size in byte */
int increment = segment * extent;
- /* if the input size is not divisible by segment size =>
+ /* if the input size is not divisible by segment size =>
the small remainder will be done with native implementation */
int remainder = count % segment;
}
/* pipeline bcast */
else {
- send_request_array =
- (MPI_Request *) xbt_malloc((size + pipe_length) * sizeof(MPI_Request));
- recv_request_array =
- (MPI_Request *) xbt_malloc((size + pipe_length) * sizeof(MPI_Request));
- send_status_array =
- (MPI_Status *) xbt_malloc((size + pipe_length) * sizeof(MPI_Status));
- recv_status_array =
- (MPI_Status *) xbt_malloc((size + pipe_length) * sizeof(MPI_Status));
+ auto* send_request_array = new MPI_Request[size + pipe_length];
+ auto* recv_request_array = new MPI_Request[size + pipe_length];
+ auto* send_status_array = new MPI_Status[size + pipe_length];
+ auto* recv_status_array = new MPI_Status[size + pipe_length];
if (rank == 0) {
//double start2 = MPI_Wtime();
}else{
Request::waitall(pipe_length, recv_request_array, recv_status_array);
}
-
+
}
- free(send_request_array);
- free(recv_request_array);
- free(send_status_array);
- free(recv_status_array);
+ delete[] send_request_array;
+ delete[] recv_request_array;
+ delete[] send_status_array;
+ delete[] recv_status_array;
} /* end pipeline */
/* when count is not divisible by block size, use default BCAST for the remainder */
if ((remainder != 0) && (count > segment)) {
- XBT_WARN("MPI_bcast_arrival_pattern_aware use default MPI_bcast.");
- Colls::bcast((char *)buf + (pipe_length * increment), remainder, datatype, root, comm);
+ XBT_WARN("MPI_bcast_arrival_pattern_aware use default MPI_bcast.");
+ colls::bcast((char*)buf + (pipe_length * increment), remainder, datatype, root, comm);
}
return MPI_SUCCESS;