[mvapich-discuss] MPI_Cart_Create
amith rajith mamidala
mamidala at cse.ohio-state.edu
Wed May 7 13:51:20 EDT 2008
Hi,
Can you try these two options?
1. use MVAPICH-1.0 and see if you are seeing the issue.
2.Can you run the code by passing the environment variable:
VIADEV_USE_SHMEM_COLL=0 with mvapich
Thanks,
Amith
On Wed, 7 May 2008 rcbord at wm.edu wrote:
> Hi,
> I am having an issue with the MVAPICH-0.9.9 compiled with PGI-7.0 with
> ofed 1.2 for infinaband. We have be able to use it for 6 months without
> any problem. I tried to port a users fortran code that runs under
> solaris-sparc without any problems. The code is fairly vanilla fotran-90
> with mpi I know because I wrote it. When I tried to port it to the Linux
> Cluster it hangs a mpisendrecv function. I included a simple test code
> that shows that the new communicator (comm2d) returned by the
> MPI_Cart_create function has multiple values. Same code run on the
> sun sparc cluster returns a single value. I am guessing it could be how
> I compiled MVAPICH, but have tried a few more flags without any success.
> I tried this with MVAPICH2 also and got the same error, but I compiled it
> in a similar fashion. Could it be a PGI problem?
>
>
> Has anyone else seen this? Thanks for any help!
>
> Program mpicart
> ! With input files
> ! Conversion of efit2d.f90 to 3d JPB 11-2007
>
> Implicit none
>
> include "mpif.h"
>
> Interface
> Subroutine read_model(Pdim0,Pdim1)
> Integer, Intent(out) ::Pdim0,Pdim1
> end Subroutine read_model
> end Interface
>
> Double Precision :: t1,t2
> !
> ! define MPI variables
> !
> Integer :: Pid, N_proc, ierr
> Integer :: comm2d = 0
> Integer :: status
> Integer :: period(2),Pdim_size(2),coords(2)
> Integer :: nbrleft,nbrright,nbrtop,nbrbottom
> Integer :: Pdim0, Pdim1
> Integer :: i, t, numt
>
> call MPI_INIT(ierr)
> call MPI_COMM_RANK(MPI_COMM_WORLD,Pid,ierr)
> call MPI_COMM_SIZE(MPI_COMM_WORLD,N_proc,ierr)
>
> if (Pid .EQ. 0) then
> call read_model(Pdim0,Pdim1)
>
> end if
> !
> ! Broad cast input parameters values to all processors
> !
> call MPI_Bcast(Pdim0,1,MPI_INTEGER,0, &
> & MPI_COMM_WORLD,ierr)
> call MPI_Bcast(Pdim1,1,MPI_INTEGER,0, &
> & MPI_COMM_WORLD,ierr)
>
> call MPI_BARRIER(MPI_COMM_WORLD,ierr)
>
> Pdim_size(1) = Pdim0
> Pdim_size(2) = Pdim1
>
> period(1) = 0
> period(2) = 0
>
> call MPI_Dims_create(N_proc,2,Pdim_size,ierr)
>
> call MPI_Cart_create(MPI_COMM_WORLD,2,Pdim_size,period,.true.,comm2d,ierr)
> do i = 0,N_proc-1
> if (Pid == i) then
> write(*,*) 'pid ',Pid,' mpi_comm_2d ',comm2d
> end if
> end do
>
> call MPI_BARRIER(MPI_COMM_WORLD,ierr)
> call MPI_Comm_free(comm2d,ierr)
> call MPI_FINALIZE(ierr)
>
> end program mpicart
>
> ###################### OUTPUT #########################
>
> Model parameters for mpicart test
>
> Processor Topology is ( 3 by 4 )
>
> pid 0 mpi_comm_2d 140
> pid 11 mpi_comm_2d 138
> pid 2 mpi_comm_2d 140
> pid 4 mpi_comm_2d 140
> pid 6 mpi_comm_2d 140
> pid 8 mpi_comm_2d 140
> pid 3 mpi_comm_2d 138
> pid 10 mpi_comm_2d 140
> pid 1 mpi_comm_2d 138
> pid 5 mpi_comm_2d 138
> pid 7 mpi_comm_2d 138
> pid 9 mpi_comm_2d 138
>
>
> Chris Bording
> Application Analyst
> High Performance Computing Group
> Information Technology
> The College of William and Mary
> (757)-221-3488
> rcbord at wm.edu
> _______________________________________________
> mvapich-discuss mailing list
> mvapich-discuss at cse.ohio-state.edu
> http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
>
More information about the mvapich-discuss
mailing list