[mvapich-discuss] charm++ / namd
Les Faby
les at pathscale.com
Fri Jun 2 12:50:06 EDT 2006
Dear Roland Fehrenbacher & Matthew Koop,
Instructions for building NAMD with Pathscale compilers:
http://www.pathscale.com/building_code/namd.html
Instructions for building other famous packages:
http://www.pathscale.com/building_code/
--
Regards ....
Les Faby
Tech Support Engineer
QLogic Corp.
System Interconnect Group
2071 Stierlin Ct, STE 200
Mountain View, CA 94043
les at pathscale.com
Please CC: your support mail to support at pathscale.com
----- Forwarded message from Roland Fehrenbacher
<Roland.Fehrenbacher at transtec.de> -----
From: Roland Fehrenbacher <Roland.Fehrenbacher at transtec.de>
Date: Fri, 2 Jun 2006 13:32:31 +0200
To: Matthew Koop <koop at cse.ohio-state.edu>
Subject: Re: [mvapich-discuss] charm++ / namd
X-Mailer: VM 7.17 under 21.4 (patch 17) "Jumbo Shrimp" XEmacs Lucid
Cc: mvapich-discuss at cse.ohio-state.edu
Reply-To: rfehrenb at transtec.de
>>>>> "Matthew" == Matthew Koop <koop at cse.ohio-state.edu> writes:
Matthew,
many thanks for your quick reply.
Matthew> Roland, Are you able to run the OSU benchmarks, such as
Matthew> osu_bw and osu_latency, or do you see the problem there
Matthew> as well? If the problem exists there a setup issue needs
Matthew> to be worked out first.
Yes, everything works fine for other prograns.
Matthew> If you are able to run basic MPI programs then we'll need
Matthew> to make sure everything was configured properly for
Matthew> charm.
Matthew> Here's what I've done in the past to prepare charm++ for
Matthew> NAMD:
Matthew> cd charm-5.9 cd ./src/arch
Matthew> cp -r mpi-linux-amd64 mpi-linux-amd64-mvapich cd
Matthew> mpi-linux-amd64-mvapich
Matthew> * edit conv-mach.h and change:
Matthew> #define CMK_MALLOC_USE_GNU_MALLOC 1
Matthew> #define CMK_MALLOC_USE_OS_BUILTIN 0
Matthew> to
Matthew> #define CMK_MALLOC_USE_GNU_MALLOC 0
Matthew> #define CMK_MALLOC_USE_OS_BUILTIN 1
This indeed solved the problem. My hello (simplearrayhello) test works
now. However, I still have a segmentation fault when trying to run the
pgm test:
$ mpiexec -comm mpich-ib ./pgm 12 10
Megatest is running on 8 processors.
test 0: initiated [bitvector (jbooth)]
mpiexec: Warning: task 0 died with signal 11 (Segmentation fault).
mpiexec: Warning: tasks 1-7 died with signal 15 (Terminated).
I compiled everything with Pathscale 2.3 (including mvapich). Do you
have any experience with Pathscale compilers and charm++? The mvapich
I built works fine with other stuff.
Thanks again,
Roland
Matthew> * make sure the MVAPICH mpicc and mpiCC are first in your
Matthew> path. Otherwise, add the full path to the mpicc and mpiCC
Matthew> commands in conv_mach.sh
Matthew> cd ../../..
Matthew> ./build charm++ mpi-linux-amd64-mvapich --no-build-shared
Matthew> cd tests/charm++/simplearrayhello make
Matthew> [koop at bm1 simplearrayhello]$ mpirun_rsh -np 2 bm1 bm2
Matthew> ./hello Running Hello on 2 processors for 5 elements ...
Matthew> Please let us know if this works for you or if you have
Matthew> any other problems.
Matthew> Thanks,
Matthew> Matthew Koop - Network-Based Computing Lab Ohio State
Matthew> University
Matthew> On Thu, 1 Jun 2006, Roland Fehrenbacher wrote:
>> Hi,
>>
>> I'm trying to get charm++ (http://charm.cs.uiuc.edu/ needed for
>> namd) to work with mvapich 0.9.7. While I can compile
>> everything fine, when executing a test job like hello (from
>> charm++) I get:
>>
>> $ mpiexec -verbose -n 2 -comm mpich-ib ./hello mpiexec:
>> resolve_exe: using absolute exe "./hello". mpiexec:
>> concurrent_init: old master died, reusing his fifo as master.
>> mpiexec: wait_task_start: start evt 2 task 0 on beo-15.
>> mpiexec: wait_task_start: start evt 3 task 1 on beo-15.
>> mpiexec: All 2 tasks started. read_ib_startup_ports: waiting
>> for checkins read_ib_startup_ports: version 3 startup
>> read_ib_startup_ports: rank 0 checked in, 1 left
>> read_ib_startup_ports: rank 1 checked in, 0 left
>> read_ib_startup_ports: barrier start mpiexec: Error: read_full:
>> EOF, only 0 of 4 bytes.
>>
>> , i.e. the job never starts.
>>
>> $ mpirun_rsh -np 2 -hostfile /var/spool/pbs/aux/84.beosrv-c
>> ./hello
>>
>> doesn't give any output.
>>
>> A run with a single CPU works fine:
>>
>> $ mpiexec -n 1 -comm mpich-ib ./hello Running Hello on 1
>> processors for 5 elements Hello 0 created Hello 1 created Hello
>> 2 created Hello 3 created Hello 4 created Hi[17] from element 0
>> Hi[18] from element 1 Hi[19] from element 2 Hi[20] from element
>> 3 Hi[21] from element 4 All done End of program
>>
>> Any ideas?
>>
>> Thanks,
>>
>> Roland
>>
>> _______________________________________________ mvapich-discuss
>> mailing list mvapich-discuss at cse.ohio-state.edu
>> http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
>>
Matthew> _______________________________________________
Matthew> mvapich-discuss mailing list
Matthew> mvapich-discuss at cse.ohio-state.edu
Matthew> http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-
discuss
_______________________________________________
mvapich-discuss mailing list
mvapich-discuss at cse.ohio-state.edu
http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
----- End forwarded message -----
More information about the mvapich-discuss
mailing list