!-------------------------------------------------------------------------! ! ! ! N A S P A R A L L E L B E N C H M A R K S 3.3 ! ! ! ! S P ! ! ! !-------------------------------------------------------------------------! ! ! ! This benchmark is part of the NAS Parallel Benchmark 3.3 suite. ! ! It is described in NAS Technical Reports 95-020 and 02-007 ! ! ! ! Permission to use, copy, distribute and modify this software ! ! for any purpose with or without fee is hereby granted. We ! ! request, however, that all derived work reference the NAS ! ! Parallel Benchmarks 3.3. This software is provided "as is" ! ! without express or implied warranty. ! ! ! ! Information on NPB 3.3, including the technical report, the ! ! original specifications, source code, results and information ! ! on how to submit new results, is available at: ! ! ! ! http://www.nas.nasa.gov/Software/NPB/ ! ! ! ! Send comments or suggestions to npb@nas.nasa.gov ! ! ! ! NAS Parallel Benchmarks Group ! ! NASA Ames Research Center ! ! Mail Stop: T27A-1 ! ! Moffett Field, CA 94035-1000 ! ! ! ! E-mail: npb@nas.nasa.gov ! ! Fax: (650) 604-3957 ! ! ! !-------------------------------------------------------------------------! c--------------------------------------------------------------------- c c Authors: R. F. Van der Wijngaart c W. Saphir c--------------------------------------------------------------------- c--------------------------------------------------------------------- program MPSP c--------------------------------------------------------------------- include 'header.h' include 'mpinpb.h' integer i, niter, step, c, error, fstatus external timer_read double precision mflops, t, tmax, timer_read logical verified character class call setup_mpi if (.not. active) goto 999 c--------------------------------------------------------------------- c Root node reads input file (if it exists) else takes c defaults from parameters c--------------------------------------------------------------------- if (node .eq. root) then write(*, 1000) open (unit=2,file='inputsp.data',status='old', iostat=fstatus) c if (fstatus .eq. 0) then write(*,233) 233 format(' Reading from input file inputsp.data') read (2,*) niter read (2,*) dt read (2,*) grid_points(1), grid_points(2), grid_points(3) close(2) else write(*,234) niter = niter_default dt = dt_default grid_points(1) = problem_size grid_points(2) = problem_size grid_points(3) = problem_size endif 234 format(' No input file inputsp.data. Using compiled defaults') write(*, 1001) grid_points(1), grid_points(2), grid_points(3) write(*, 1002) niter, dt if (no_nodes .ne. total_nodes) write(*, 1004) total_nodes if (no_nodes .ne. maxcells*maxcells) > write(*, 1005) maxcells*maxcells write(*, 1003) no_nodes 1000 format(//,' NAS Parallel Benchmarks 3.3 -- SP Benchmark',/) 1001 format(' Size: ', i4, 'x', i4, 'x', i4) 1002 format(' Iterations: ', i4, ' dt: ', F11.7) 1004 format(' Total number of processes: ', i5) 1005 format(' WARNING: compiled for ', i5, ' processes ') 1003 format(' Number of active processes: ', i5, /) endif call mpi_bcast(niter, 1, MPI_INTEGER, > root, comm_setup, error) call mpi_bcast(dt, 1, dp_type, > root, comm_setup, error) call mpi_bcast(grid_points(1), 3, MPI_INTEGER, > root, comm_setup, error) call make_set do c = 1, ncells if ( (cell_size(1,c) .gt. IMAX) .or. > (cell_size(2,c) .gt. JMAX) .or. > (cell_size(3,c) .gt. KMAX) ) then print *,node, c, (cell_size(i,c),i=1,3) print *,' Problem size too big for compiled array sizes' goto 999 endif end do call set_constants call initialize c call mpi_finalize(error) c stop call lhsinit call exact_rhs call compute_buffer_size(5) c--------------------------------------------------------------------- c do one time step to touch all code, and reinitialize c--------------------------------------------------------------------- call adi call initialize c--------------------------------------------------------------------- c Synchronize before placing time stamp c--------------------------------------------------------------------- call mpi_barrier(comm_setup, error) call timer_clear(1) call timer_start(1) do step = 1, niter if (node .eq. root) then if (mod(step, 20) .eq. 0 .or. > step .eq. 1) then write(*, 200) step 200 format(' Time step ', i4) endif endif call adi end do call timer_stop(1) t = timer_read(1) call verify(niter, class, verified) call mpi_reduce(t, tmax, 1, > dp_type, MPI_MAX, > root, comm_setup, error) if( node .eq. root ) then if( tmax .ne. 0. ) then mflops = (881.174*float( problem_size )**3 > -4683.91*float( problem_size )**2 > +11484.5*float( problem_size ) > -19272.4) * float( niter ) / (tmax*1000000.0d0) else mflops = 0.0 endif call print_results('SP', class, grid_points(1), > grid_points(2), grid_points(3), niter, maxcells*maxcells, > total_nodes, tmax, mflops, ' floating point', > verified, npbversion,compiletime, cs1, cs2, cs3, cs4, cs5, > cs6, '(none)') endif 999 continue call mpi_barrier(MPI_COMM_WORLD, error) call mpi_finalize(error) end