A
lgorithmique
N
umérique
D
istribuée
Public GIT Repository
projects
/
simgrid.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Update the SMPI documentation, mainly to add the collective algorithms
[simgrid.git]
/
src
/
smpi
/
colls
/
allgather-spreading-simple.c
diff --git
a/src/smpi/colls/allgather-spreading-simple.c
b/src/smpi/colls/allgather-spreading-simple.c
index
3b109e4
..
c7dfbc9
100644
(file)
--- a/
src/smpi/colls/allgather-spreading-simple.c
+++ b/
src/smpi/colls/allgather-spreading-simple.c
@@
-1,4
+1,4
@@
-#include "colls.h"
+#include "colls
_private
.h"
/*****************************************************************************
/*****************************************************************************
@@
-70,16
+70,16
@@
smpi_coll_tuned_allgather_spreading_simple(void *send_buff, int send_count,
MPI_Request *reqs, *req_ptr;
MPI_Aint extent;
int i, src, dst, rank, num_procs, num_reqs;
MPI_Request *reqs, *req_ptr;
MPI_Aint extent;
int i, src, dst, rank, num_procs, num_reqs;
- int tag =
1
;
+ int tag =
COLL_TAG_ALLGATHER
;
MPI_Status status;
char *recv_ptr = (char *) recv_buff;
MPI_Status status;
char *recv_ptr = (char *) recv_buff;
-
MPI_Comm_rank(comm, &rank
);
-
MPI_Comm_size(comm, &num_procs
);
-
MPI_Type_extent(send_type, &extent
);
+
rank = smpi_comm_rank(comm
);
+
num_procs = smpi_comm_size(comm
);
+
extent = smpi_datatype_get_extent(send_type
);
num_reqs = (2 * num_procs) - 2;
num_reqs = (2 * num_procs) - 2;
- reqs = (MPI_Request *) malloc(num_reqs * sizeof(MPI_Request));
+ reqs = (MPI_Request *)
xbt_
malloc(num_reqs * sizeof(MPI_Request));
if (!reqs) {
printf("allgather-spreading-simple.c:40: cannot allocate memory\n");
MPI_Finalize();
if (!reqs) {
printf("allgather-spreading-simple.c:40: cannot allocate memory\n");
MPI_Finalize();
@@
-87,7
+87,7
@@
smpi_coll_tuned_allgather_spreading_simple(void *send_buff, int send_count,
}
req_ptr = reqs;
}
req_ptr = reqs;
-
MPI_S
endrecv(send_buff, send_count, send_type, rank, tag,
+
smpi_mpi_s
endrecv(send_buff, send_count, send_type, rank, tag,
(char *) recv_buff + rank * recv_count * extent, recv_count,
recv_type, rank, tag, comm, &status);
(char *) recv_buff + rank * recv_count * extent, recv_count,
recv_type, rank, tag, comm, &status);
@@
-95,18
+95,18
@@
smpi_coll_tuned_allgather_spreading_simple(void *send_buff, int send_count,
src = (rank + i) % num_procs;
if (src == rank)
continue;
src = (rank + i) % num_procs;
if (src == rank)
continue;
-
MPI_I
recv(recv_ptr + src * recv_count * extent, recv_count, recv_type,
- src, tag, comm
, req_ptr++
);
+
*(req_ptr++) = smpi_mpi_i
recv(recv_ptr + src * recv_count * extent, recv_count, recv_type,
+ src, tag, comm);
}
for (i = 0; i < num_procs; i++) {
dst = (rank + i) % num_procs;
if (dst == rank)
continue;
}
for (i = 0; i < num_procs; i++) {
dst = (rank + i) % num_procs;
if (dst == rank)
continue;
-
MPI_Isend(send_buff, send_count, send_type, dst, tag, comm, req_ptr++
);
+
*(req_ptr++) = smpi_mpi_isend(send_buff, send_count, send_type, dst, tag, comm
);
}
}
-
MPI_W
aitall(num_reqs, reqs, MPI_STATUSES_IGNORE);
+
smpi_mpi_w
aitall(num_reqs, reqs, MPI_STATUSES_IGNORE);
free(reqs);
return MPI_SUCCESS;
free(reqs);
return MPI_SUCCESS;