Logo AND Algorithmique Numérique Distribuée

Public GIT Repository
speeling mistakes in teshsuite (even mpich tests because why not)
[simgrid.git] / teshsuite / smpi / mpich3-test / rma / accpscw1.c
index 4b4976e..dd9162c 100644 (file)
 static char MTEST_Descrip[] = "Accumulate/replace with Post/Start/Complete/Wait";
 */
 
-int main( int argc, char *argv[] )
+int main(int argc, char *argv[])
 {
     int errs = 0, err;
     int rank, size, source, dest;
-    int minsize = 2, count; 
-    MPI_Comm      comm;
-    MPI_Win       win;
-    MPI_Aint      extent;
-    MPI_Group     wingroup, neighbors;
+    int minsize = 2, count;
+    MPI_Comm comm;
+    MPI_Win win;
+    MPI_Aint extent, lb;
+    MPI_Group wingroup, neighbors;
     MTestDatatype sendtype, recvtype;
 
-    MTest_Init( &argc, &argv );
+    MTest_Init(&argc, &argv);
 
-    /* The following illustrates the use of the routines to 
-       run through a selection of communicators and datatypes.
-       Use subsets of these for tests that do not involve combinations 
-       of communicators, datatypes, and counts of datatypes */
-    while (MTestGetIntracommGeneral( &comm, minsize, 1 )) {
-       if (comm == MPI_COMM_NULL) continue;
-       /* Determine the sender and receiver */
-       MPI_Comm_rank( comm, &rank );
-       MPI_Comm_size( comm, &size );
-       source = 0;
-       dest   = size - 1;
-       
-       for (count = 1; count < 65000; count = count * 2) {
-           while (MTestGetDatatypes( &sendtype, &recvtype, count )) {
-               /* Make sure that everyone has a recv buffer */
-               recvtype.InitBuf( &recvtype );
+    /* The following illustrates the use of the routines to
+     * run through a selection of communicators and datatypes.
+     * Use subsets of these for tests that do not involve combinations
+     * of communicators, datatypes, and counts of datatypes */
+    while (MTestGetIntracommGeneral(&comm, minsize, 1)) {
+        if (comm == MPI_COMM_NULL)
+            continue;
+        /* Determine the sender and receiver */
+        MPI_Comm_rank(comm, &rank);
+        MPI_Comm_size(comm, &size);
+        source = 0;
+        dest = size - 1;
 
-               MPI_Type_extent( recvtype.datatype, &extent );
-               MPI_Win_create( recvtype.buf, recvtype.count * extent, 
-                               (int)extent, MPI_INFO_NULL, comm, &win );
-               MPI_Win_get_group( win, &wingroup );
-               if (rank == source) {
-                   /* To improve reporting of problems about operations, we
-                      change the error handler to errors return */
-                   MPI_Win_set_errhandler( win, MPI_ERRORS_RETURN );
-                   sendtype.InitBuf( &sendtype );
-                   
-                   /* Neighbor is dest only */
-                   MPI_Group_incl( wingroup, 1, &dest, &neighbors );
-                   err = MPI_Win_start( neighbors, 0, win );
-                   if (err) {
-                       errs++;
-                       if (errs < 10) {
-                           MTestPrintError( err );
-                       }
-                   }
-                   MPI_Group_free( &neighbors );
-                   err = MPI_Accumulate( sendtype.buf, sendtype.count, 
-                                         sendtype.datatype, dest, 0, 
-                                         recvtype.count, recvtype.datatype, 
-                                         MPI_REPLACE, win );
-                   if (err) {
-                       errs++;
-                       MTestPrintError( err );
-                   }
-                   err = MPI_Win_complete( win );
-                   if (err) {
-                       errs++;
-                       if (errs < 10) {
-                           MTestPrintError( err );
-                       }
-                   }
-               }
-               else if (rank == dest) {
-                   MPI_Group_incl( wingroup, 1, &source, &neighbors );
-                   MPI_Win_post( neighbors, 0, win );
-                   MPI_Group_free( &neighbors );
-                   MPI_Win_wait( win );
-                   /* This should have the same effect, in terms of
-                      transfering data, as a send/recv pair */
-                   err = MTestCheckRecv( 0, &recvtype );
-                   if (err) {
-                       errs += errs;
-                   }
-               }
-               else {
-                   /* Nothing; the other processes need not call any 
-                      MPI routines */
-                   ;
-               }
-               MPI_Win_free( &win );
-               MTestFreeDatatype( &sendtype );
-               MTestFreeDatatype( &recvtype );
-               MPI_Group_free( &wingroup );
-           }
-       }
-       MTestFreeComm( &comm );
+        MTEST_DATATYPE_FOR_EACH_COUNT(count) {
+            while (MTestGetDatatypes(&sendtype, &recvtype, count)) {
+                /* Make sure that everyone has a recv buffer */
+                recvtype.InitBuf(&recvtype);
+
+                MPI_Type_extent(recvtype.datatype, &extent);
+                MPI_Type_lb(recvtype.datatype, &lb);
+                MPI_Win_create(recvtype.buf, recvtype.count * extent + lb,
+                               (int) extent, MPI_INFO_NULL, comm, &win);
+                MPI_Win_get_group(win, &wingroup);
+                if (rank == source) {
+                    /* To improve reporting of problems about operations, we
+                     * change the error handler to errors return */
+                    MPI_Win_set_errhandler(win, MPI_ERRORS_RETURN);
+                    sendtype.InitBuf(&sendtype);
+
+                    /* Neighbor is dest only */
+                    MPI_Group_incl(wingroup, 1, &dest, &neighbors);
+                    err = MPI_Win_start(neighbors, 0, win);
+                    if (err) {
+                        errs++;
+                        if (errs < 10) {
+                            MTestPrintError(err);
+                        }
+                    }
+                    MPI_Group_free(&neighbors);
+                    err = MPI_Accumulate(sendtype.buf, sendtype.count,
+                                         sendtype.datatype, dest, 0,
+                                         recvtype.count, recvtype.datatype, MPI_REPLACE, win);
+                    if (err) {
+                        errs++;
+                        MTestPrintError(err);
+                    }
+                    err = MPI_Win_complete(win);
+                    if (err) {
+                        errs++;
+                        if (errs < 10) {
+                            MTestPrintError(err);
+                        }
+                    }
+                }
+                else if (rank == dest) {
+                    MPI_Group_incl(wingroup, 1, &source, &neighbors);
+                    MPI_Win_post(neighbors, 0, win);
+                    MPI_Group_free(&neighbors);
+                    MPI_Win_wait(win);
+                    /* This should have the same effect, in terms of
+                     * transferring data, as a send/recv pair */
+                    err = MTestCheckRecv(0, &recvtype);
+                    if (err) {
+                        errs += errs;
+                    }
+                }
+                else {
+                    /* Nothing; the other processes need not call any
+                     * MPI routines */
+                    ;
+                }
+                MPI_Win_free(&win);
+                MTestFreeDatatype(&sendtype);
+                MTestFreeDatatype(&recvtype);
+                MPI_Group_free(&wingroup);
+            }
+        }
+        MTestFreeComm(&comm);
     }
 
-    MTest_Finalize( errs );
+    MTest_Finalize(errs);
     MPI_Finalize();
     return 0;
 }