X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/b3685de3f6639ce25de105539ef95dd5b753ccae..9982cd9fee0903300e2323bc196a5262f4e5b969:/teshsuite/smpi/coll-allreduce/coll-allreduce-automatic.tesh diff --git a/teshsuite/smpi/coll-allreduce/coll-allreduce-automatic.tesh b/teshsuite/smpi/coll-allreduce/coll-allreduce-automatic.tesh index b34fd9bdf9..b8b77622c1 100644 --- a/teshsuite/smpi/coll-allreduce/coll-allreduce-automatic.tesh +++ b/teshsuite/smpi/coll-allreduce/coll-allreduce-automatic.tesh @@ -1,9 +1,8 @@ # Smpi Allreduce collectives tests -! setenv LD_LIBRARY_PATH=../../lib p Test allreduce ! output sort -$ ${bindir:=.}/../../../bin/smpirun -map -hostfile ../hostfile_coll -platform ../../../examples/platforms/small_platform.xml -np 16 --log=xbt_cfg.thres:critical ${bindir:=.}/coll-allreduce --log=smpi_kernel.thres:warning --log=smpi_coll.thres:error --cfg=smpi/allreduce:automatic --cfg=smpi/async-small-thresh:65536 --cfg=smpi/send-is-detached-thresh:128000 --cfg=smpi/simulate-computation:no "--log=root.fmt:[%10.6r]%e(%i:%P@%h)%e%m%n" +$ ${bindir:=.}/../../../smpi_script/bin/smpirun -map -hostfile ../hostfile_coll -platform ../../../examples/platforms/small_platform.xml -np 16 --log=xbt_cfg.thres:critical ${bindir:=.}/coll-allreduce --log=smpi_kernel.thres:warning --log=smpi_coll.thres:error --cfg=smpi/allreduce:automatic --cfg=smpi/async-small-thresh:65536 --cfg=smpi/send-is-detached-thresh:128000 --cfg=smpi/simulate-computation:no "--log=root.fmt:[%10.6r]%e(%i:%P@%h)%e%m%n" > [rank 0] -> Tremblay > [rank 1] -> Tremblay > [rank 2] -> Tremblay @@ -20,22 +19,22 @@ $ ${bindir:=.}/../../../bin/smpirun -map -hostfile ../hostfile_coll -platform .. > [rank 13] -> Ginette > [rank 14] -> Ginette > [rank 15] -> Ginette -> [ 0.420171] (8:7@Jupiter) The quickest allreduce was redbcast on rank 7 and took 0.007545 -> [ 0.420171] (5:4@Jupiter) The quickest allreduce was redbcast on rank 4 and took 0.007484 -> [ 0.420171] (7:6@Jupiter) The quickest allreduce was redbcast on rank 6 and took 0.007514 -> [ 0.420171] (6:5@Jupiter) The quickest allreduce was redbcast on rank 5 and took 0.007514 -> [ 0.420383] (14:13@Ginette) The quickest allreduce was mvapich2_two_level on rank 13 and took 0.007278 -> [ 0.420383] (13:12@Ginette) The quickest allreduce was mvapich2_two_level on rank 12 and took 0.007247 -> [ 0.420383] (16:15@Ginette) The quickest allreduce was ompi on rank 15 and took 0.007263 -> [ 0.420383] (15:14@Ginette) The quickest allreduce was mvapich2_two_level on rank 14 and took 0.007278 -> [ 0.421775] (2:1@Tremblay) The quickest allreduce was redbcast on rank 1 and took 0.006005 -> [ 0.421775] (3:2@Tremblay) The quickest allreduce was redbcast on rank 2 and took 0.006005 -> [ 0.421775] (4:3@Tremblay) The quickest allreduce was redbcast on rank 3 and took 0.006036 -> [ 0.422926] (12:11@Fafard) The quickest allreduce was mvapich2_two_level on rank 11 and took 0.006523 -> [ 0.422926] (10:9@Fafard) The quickest allreduce was mvapich2_two_level on rank 9 and took 0.006492 -> [ 0.422926] (9:8@Fafard) The quickest allreduce was mvapich2_two_level on rank 8 and took 0.006462 -> [ 0.422926] (11:10@Fafard) The quickest allreduce was mvapich2_two_level on rank 10 and took 0.006492 -> [ 0.426911] (1:0@Tremblay) For rank 0, the quickest was redbcast : 0.005990 , but global was mvapich2_two_level : 0.008672 at max +> [ 0.427764] (8:7@Jupiter) The quickest allreduce was redbcast on rank 7 and took 0.007546 +> [ 0.427764] (5:4@Jupiter) The quickest allreduce was redbcast on rank 4 and took 0.007485 +> [ 0.427764] (7:6@Jupiter) The quickest allreduce was redbcast on rank 6 and took 0.007515 +> [ 0.427764] (6:5@Jupiter) The quickest allreduce was redbcast on rank 5 and took 0.007515 +> [ 0.427976] (14:13@Ginette) The quickest allreduce was mvapich2_two_level on rank 13 and took 0.007278 +> [ 0.427976] (13:12@Ginette) The quickest allreduce was mvapich2_two_level on rank 12 and took 0.007247 +> [ 0.427976] (16:15@Ginette) The quickest allreduce was ompi on rank 15 and took 0.007263 +> [ 0.427976] (15:14@Ginette) The quickest allreduce was mvapich2_two_level on rank 14 and took 0.007278 +> [ 0.429367] (2:1@Tremblay) The quickest allreduce was redbcast on rank 1 and took 0.006006 +> [ 0.429367] (3:2@Tremblay) The quickest allreduce was redbcast on rank 2 and took 0.006006 +> [ 0.429367] (4:3@Tremblay) The quickest allreduce was redbcast on rank 3 and took 0.006037 +> [ 0.430519] (12:11@Fafard) The quickest allreduce was mvapich2_two_level on rank 11 and took 0.006523 +> [ 0.430519] (10:9@Fafard) The quickest allreduce was mvapich2_two_level on rank 9 and took 0.006492 +> [ 0.430519] (9:8@Fafard) The quickest allreduce was mvapich2_two_level on rank 8 and took 0.006462 +> [ 0.430519] (11:10@Fafard) The quickest allreduce was mvapich2_two_level on rank 10 and took 0.006492 +> [ 0.434504] (1:0@Tremblay) For rank 0, the quickest was redbcast : 0.005991 , but global was mvapich2_two_level : 0.008672 at max > [0] sndbuf=[0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 ] > [1] sndbuf=[16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 ] > [2] sndbuf=[32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 ]