[mvapich-discuss] Announcing the Release of MVAPICH2 1.5RC1

Sayantan Sur surs at cse.ohio-state.edu
Tue May 25 11:39:13 EDT 2010


Hi Eric,

Thanks for bringing this to our attention. This will be applied to
MVAPICH2 as well.

On Tue, May 25, 2010 at 10:45 AM, Eric A. Borisch <eborisch at ieee.org> wrote:
> There is a typo in the current src/include/mpimem.h at line 331:
> (trunk and 1.5RC1)
>
> It is only evident with --enable-alloca
>
> == ORIGINAL ==
>
> #define MPIU_CHKLMEM_MALLOC_ORSTMT(pointer_,type_,nbytes_,rc_,name_,stmt_) \
> {pointer_ = (type_)alloca(nbytes_); \
>    if (!(pointer_) && (nbytes > 0)) {     \
>    MPIU_CHKMEM_SETERR(rc_,nbytes_,name_); \
>    stmt_;\
> }}
>
> == FIXED ==
>
> #define MPIU_CHKLMEM_MALLOC_ORSTMT(pointer_,type_,nbytes_,rc_,name_,stmt_) \
> {pointer_ = (type_)alloca(nbytes_); \
>    if (!(pointer_) && (nbytes_ > 0)) {    \
>    MPIU_CHKMEM_SETERR(rc_,nbytes_,name_); \
>    stmt_;\
> }}
>
> == ==
>
> It's missing the underscore after nbytes in the if() clause.
>
> It was fixed in mpich2 a few months ago:
> http://trac.mcs.anl.gov/projects/mpich2/ticket/960
>
> Thanks,
>  Eric
>
> On Wed, May 5, 2010 at 12:08 AM, Dhabaleswar Panda
> <panda at cse.ohio-state.edu> wrote:
>> The MVAPICH team is pleased to announce the release of MVAPICH2 1.5RC1
>> with the following NEW features:
>>
>> - MPI 2.2 standard compliant
>> - Based on MPICH2 1.2.1p1
>> - OFA-IB-Nemesis interface design
>>    - OpenFabrics InfiniBand network module support for
>>      MPICH2 Nemesis modular design
>>    - Support for high-performance intra-node shared memory
>>      communication provided by the Nemesis design
>>    - Adaptive RDMA Fastpath with Polling Set for high-performance
>>      inter-node communication
>>    - Shared Receive Queue (SRQ) support with flow control,
>>      uses significantly less memory for MPI library
>>    - Header caching
>>    - Advanced AVL tree-based Resource-aware registration cache
>>    - Memory Hook Support provided by integration with ptmalloc2
>>      library. This provides safe release of memory to the
>>      Operating System and is expected to benefit the memory
>>      usage of applications that heavily use malloc and free operations.
>>    - Support for TotalView debugger
>>    - Shared Library Support for existing binary MPI application
>>      programs to run ROMIO Support for MPI-IO
>>    - Support for additional features (such as hwloc,
>>      hierarchical collectives, one-sided, multithreading, etc.),
>>      as included in the MPICH2 1.2.1p1 Nemesis channel
>> - Flexible process manager support
>>    - mpirun_rsh to work with any of the eight interfaces
>>      (CH3 and Nemesis channel-based) including OFA-IB-Nemesis,
>>      TCP/IP-CH3 and TCP/IP-Nemesis
>>    - Hydra process manager to work with any of the eight interfaces
>>      (CH3 and Nemesis channel-based) including OFA-IB-CH3,
>>      OFA-iWARP-CH3, OFA-RoCE-CH3 and TCP/IP-CH3
>> - MPIEXEC_TIMEOUT is honored by mpirun_rsh
>>
>> This release also contains multiple bug fixes since MVAPICH2-1.4.1 A
>> summary of the major fixes are as follows:
>>
>> - Fix compilation error when configured with
>>  `--enable-thread-funneled'
>> - Fix MPE functionality, thanks to Anthony Chan <chan at mcs.anl.gov> for
>>  reporting and providing the resolving patch
>> - Cleanup after a failure in the init phase is handled better by
>>  mpirun_rsh
>> - Path determination is correctly handled by mpirun_rsh when DPM is
>>  used
>> - Shared libraries are correctly built (again)
>>
>> For downloading MVAPICH2 1.5RC1, associated user guide and accessing
>> the SVN, please visit the following URL:
>>
>> http://mvapich.cse.ohio-state.edu
>>
>> All feedbacks, including bug reports and hints for performance tuning,
>> patches and enhancements are welcome. Please post it to the
>> mvapich-discuss mailing list.
>>
>> We are also happy to inform that the number of organizations using
>> MVAPICH/MVAPICH2 (and registered at the MVAPICH site) has crossed
>> 1,100 world-wide (in 58 countries). The MVAPICH team extends
>> thanks to all these organizations.
>>
>> Thanks,
>>
>> The MVAPICH Team
>>
>>
>> _______________________________________________
>> mvapich-discuss mailing list
>> mvapich-discuss at cse.ohio-state.edu
>> http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
>>
>
>
>
> --
> Eric A. Borisch
> eborisch at ieee.org
>
> Howard Roark laughed.
>
> _______________________________________________
> mvapich-discuss mailing list
> mvapich-discuss at cse.ohio-state.edu
> http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
>
>



-- 
Sayantan Sur

Research Scientist
Department of Computer Science
The Ohio State University.



More information about the mvapich-discuss mailing list