-> [ 0.420130] (5:4@Jupiter) The quickest allreduce was redbcast on rank 4 and took 0.007482
-> [ 0.420130] (6:5@Jupiter) The quickest allreduce was redbcast on rank 5 and took 0.007513
-> [ 0.420130] (7:6@Jupiter) The quickest allreduce was redbcast on rank 6 and took 0.007513
-> [ 0.420130] (8:7@Jupiter) The quickest allreduce was redbcast on rank 7 and took 0.007543
-> [ 0.420342] (13:12@Ginette) The quickest allreduce was mvapich2_two_level on rank 12 and took 0.007247
-> [ 0.420342] (14:13@Ginette) The quickest allreduce was mvapich2_two_level on rank 13 and took 0.007278
-> [ 0.420342] (15:14@Ginette) The quickest allreduce was mvapich2_two_level on rank 14 and took 0.007278
-> [ 0.420342] (16:15@Ginette) The quickest allreduce was ompi on rank 15 and took 0.007263
-> [ 0.421734] (2:1@Tremblay) The quickest allreduce was redbcast on rank 1 and took 0.006004
-> [ 0.421734] (3:2@Tremblay) The quickest allreduce was redbcast on rank 2 and took 0.006004
-> [ 0.421734] (4:3@Tremblay) The quickest allreduce was redbcast on rank 3 and took 0.006034
-> [ 0.422884] (10:9@Fafard) The quickest allreduce was mvapich2_two_level on rank 9 and took 0.006492
-> [ 0.422884] (11:10@Fafard) The quickest allreduce was mvapich2_two_level on rank 10 and took 0.006492
-> [ 0.422884] (12:11@Fafard) The quickest allreduce was mvapich2_two_level on rank 11 and took 0.006523
-> [ 0.422884] (9:8@Fafard) The quickest allreduce was mvapich2_two_level on rank 8 and took 0.006462
-> [ 0.426869] (1:0@Tremblay) For rank 0, the quickest was redbcast : 0.005988 , but global was mvapich2_two_level : 0.008672 at max
-> [0] rcvbuf=[1920 1936 1952 1968 1984 2000 2016 2032 2048 2064 2080 2096 2112 2128 2144 2160 ]
+> [rank 0] -> Tremblay
+> [rank 1] -> Tremblay
+> [rank 2] -> Tremblay
+> [rank 3] -> Tremblay
+> [rank 4] -> Jupiter
+> [rank 5] -> Jupiter
+> [rank 6] -> Jupiter
+> [rank 7] -> Jupiter
+> [rank 8] -> Fafard
+> [rank 9] -> Fafard
+> [rank 10] -> Fafard
+> [rank 11] -> Fafard
+> [rank 12] -> Ginette
+> [rank 13] -> Ginette
+> [rank 14] -> Ginette
+> [rank 15] -> Ginette
+> [ 0.420156] (8:7@Jupiter) The quickest allreduce was redbcast on rank 7 and took 0.007545
+> [ 0.420156] (5:4@Jupiter) The quickest allreduce was redbcast on rank 4 and took 0.007484
+> [ 0.420156] (7:6@Jupiter) The quickest allreduce was redbcast on rank 6 and took 0.007514
+> [ 0.420156] (6:5@Jupiter) The quickest allreduce was redbcast on rank 5 and took 0.007514
+> [ 0.420368] (14:13@Ginette) The quickest allreduce was mvapich2_two_level on rank 13 and took 0.007278
+> [ 0.420368] (13:12@Ginette) The quickest allreduce was mvapich2_two_level on rank 12 and took 0.007247
+> [ 0.420368] (16:15@Ginette) The quickest allreduce was ompi on rank 15 and took 0.007263
+> [ 0.420368] (15:14@Ginette) The quickest allreduce was mvapich2_two_level on rank 14 and took 0.007278
+> [ 0.421760] (2:1@Tremblay) The quickest allreduce was redbcast on rank 1 and took 0.006005
+> [ 0.421760] (3:2@Tremblay) The quickest allreduce was redbcast on rank 2 and took 0.006005
+> [ 0.421760] (4:3@Tremblay) The quickest allreduce was redbcast on rank 3 and took 0.006036
+> [ 0.422911] (12:11@Fafard) The quickest allreduce was mvapich2_two_level on rank 11 and took 0.006523
+> [ 0.422911] (10:9@Fafard) The quickest allreduce was mvapich2_two_level on rank 9 and took 0.006492
+> [ 0.422911] (9:8@Fafard) The quickest allreduce was mvapich2_two_level on rank 8 and took 0.006462
+> [ 0.422911] (11:10@Fafard) The quickest allreduce was mvapich2_two_level on rank 10 and took 0.006492
+> [ 0.426896] (1:0@Tremblay) For rank 0, the quickest was redbcast : 0.005990 , but global was mvapich2_two_level : 0.008672 at max