--- /dev/null
+#include "colls.h"
+
+/*****************************************************************************
+
+Copyright (c) 2006, Ahmad Faraj & Xin Yuan,
+All rights reserved.
+
+Redistribution and use in source and binary forms, with or without
+modification, are permitted provided that the following conditions are met:
+
+ * Redistributions of source code must retain the above copyright notice,
+ this list of conditions and the following disclaimer.
+
+ * Redistributions in binary form must reproduce the above copyright notice,
+ this list of conditions and the following disclaimer in the documentation
+ and/or other materials provided with the distribution.
+
+ * Neither the name of the Florida State University nor the names of its
+ contributors may be used to endorse or promote products derived from this
+ software without specific prior written permission.
+
+THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
+ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
+WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
+DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR
+ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
+(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
+LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON
+ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
+SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+ *************************************************************************
+ * Any results obtained from executing this software require the *
+ * acknowledgment and citation of the software and its owners. *
+ * The full citation is given below: *
+ * *
+ * A. Faraj and X. Yuan. "Automatic Generation and Tuning of MPI *
+ * Collective Communication Routines." The 19th ACM International *
+ * Conference on Supercomputing (ICS), Cambridge, Massachusetts, *
+ * June 20-22, 2005. *
+ *************************************************************************
+
+*****************************************************************************/
+
+
+/*****************************************************************************
+ * Function: allgather_bruck
+ * return: int
+ * inputs:
+ * send_buff: send input buffer
+ * send_count: number of elements to send
+ * send_type: data type of elements being sent
+ * recv_buff: receive output buffer
+ * recv_count: number of elements to received
+ * recv_type: data type of elements being received
+ * comm: communication
+ * Descrp: Function realizes the allgather operation using the bruck
+ * algorithm.
+ * Auther: MPICH
+ * Comment: Original bruck algorithm from MPICH is slightly modified by
+ * Ahmad Faraj.
+ ****************************************************************************/
+int smpi_coll_tuned_allgather_bruck(void *send_buff, int send_count,
+ MPI_Datatype send_type, void *recv_buff,
+ int recv_count, MPI_Datatype recv_type,
+ MPI_Comm comm)
+{
+ // MPI variables
+ MPI_Status status;
+ MPI_Aint recv_extent;
+
+ // local int variables
+ int i, src, dst, rank, num_procs, count, remainder;
+ int tag = 1;
+ int pof2 = 1;
+ int success = 0;
+
+ // local string variables
+ char *tmp_buff;
+ char *send_ptr = (char *) send_buff;
+ char *recv_ptr = (char *) recv_buff;
+
+ // get size of the communicator, followed by rank
+ MPI_Comm_size(comm, &num_procs);
+ MPI_Comm_rank(comm, &rank);
+
+ // get size of single element's type for recv buffer
+ MPI_Type_extent(recv_type, &recv_extent);
+
+ count = recv_count;
+
+ tmp_buff = (char *) malloc(num_procs * recv_count * recv_extent);
+ if (!tmp_buff) {
+ printf("allgather-bruck:54: cannot allocate memory\n");
+ MPI_Finalize();
+ exit(0);
+ }
+ // perform a local copy
+ MPIR_Localcopy(send_ptr, send_count, send_type, tmp_buff, recv_count,
+ recv_type);
+
+ while (pof2 <= (num_procs / 2)) {
+ src = (rank + pof2) % num_procs;
+ dst = (rank - pof2 + num_procs) % num_procs;
+
+ MPIC_Sendrecv(tmp_buff, count, recv_type, dst, tag,
+ tmp_buff + count * recv_extent, count, recv_type,
+ src, tag, comm, &status);
+ count *= 2;
+ pof2 *= 2;
+ }
+
+ remainder = num_procs - pof2;
+ if (remainder) {
+ src = (rank + pof2) % num_procs;
+ dst = (rank - pof2 + num_procs) % num_procs;
+
+ MPIC_Sendrecv(tmp_buff, remainder * recv_count, recv_type, dst, tag,
+ tmp_buff + count * recv_extent, remainder * recv_count,
+ recv_type, src, tag, comm, &status);
+ }
+
+ MPIC_Sendrecv(tmp_buff, (num_procs - rank) * recv_count, recv_type, rank,
+ tag, recv_ptr + rank * recv_count * recv_extent,
+ (num_procs - rank) * recv_count, recv_type, rank, tag, comm,
+ &status);
+
+ if (rank)
+ MPIC_Sendrecv(tmp_buff + (num_procs - rank) * recv_count * recv_extent,
+ rank * recv_count, recv_type, rank, tag, recv_ptr,
+ rank * recv_count, recv_type, rank, tag, comm, &status);
+ free(tmp_buff);
+ return success;
+}
+
+/*#include "ompi_bindings.h"
+
+int ompi_coll_tuned_alltoall_intra_pairwise(void *sbuf, int scount,
+ MPI_Datatype sdtype,
+ void* rbuf, int rcount,
+ MPI_Datatype rdtype,
+ MPI_Comm comm)
+{
+ int line = -1, err = 0;
+ int rank, size, step;
+ int sendto, recvfrom;
+ void * tmpsend, *tmprecv;
+ ptrdiff_t lb, sext, rext;
+
+ size = ompi_comm_size(comm);
+ rank = ompi_comm_rank(comm);
+
+ OPAL_OUTPUT((ompi_coll_tuned_stream,
+ "coll:tuned:alltoall_intra_pairwise rank %d", rank));
+
+ err = ompi_datatype_get_extent (sdtype, &lb, &sext);
+ if (err != MPI_SUCCESS) { line = __LINE__; goto err_hndl; }
+ err = ompi_datatype_get_extent (rdtype, &lb, &rext);
+ if (err != MPI_SUCCESS) { line = __LINE__; goto err_hndl; }
+
+
+ // Perform pairwise exchange - starting from 1 so the local copy is last
+ for (step = 1; step < size + 1; step++) {
+
+ // Determine sender and receiver for this step.
+ sendto = (rank + step) % size;
+ recvfrom = (rank + size - step) % size;
+
+ // Determine sending and receiving locations
+ tmpsend = (char*)sbuf + sendto * sext * scount;
+ tmprecv = (char*)rbuf + recvfrom * rext * rcount;
+
+ // send and receive
+ err = ompi_coll_tuned_sendrecv( tmpsend, scount, sdtype, sendto,
+ MCA_COLL_BASE_TAG_ALLTOALL,
+ tmprecv, rcount, rdtype, recvfrom,
+ MCA_COLL_BASE_TAG_ALLTOALL,
+ comm, MPI_STATUS_IGNORE, rank);
+ if (err != MPI_SUCCESS) { line = __LINE__; goto err_hndl; }
+ }
+
+ return MPI_SUCCESS;
+
+ err_hndl:
+ OPAL_OUTPUT((ompi_coll_tuned_stream,
+ "%s:%4d\tError occurred %d, rank %2d", __FILE__, line,
+ err, rank));
+ return err;
+}
+*/