===============================================================================
Changes in 1.2.1
===============================================================================
# OVERALL: Improved support for fine-grained multithreading.
# OVERALL: Improved integration with Valgrind for debugging builds of MPICH2.
# PM/PMI: Initial support for hwloc process-core binding library in
Hydra.
# PM/PMI: Updates to the PMI-2 code to match the PMI-2 API and
wire-protocol draft.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r5425:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.2.1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.2.1?action=follow_copy&rev=HEAD&stop_rev=5425&mode=follow_copy
===============================================================================
Changes in 1.2
===============================================================================
# OVERALL: Support for MPI-2.2
# OVERALL: Several fixes to Nemesis/MX.
# WINDOWS: Performance improvements to Nemesis/windows.
# PM/PMI: Scalability and performance improvements to Hydra using
PMI-1.1 process-mapping features.
# PM/PMI: Support for process-binding for hyperthreading enabled
systems in Hydra.
# PM/PMI: Initial support for PBS as a resource management kernel in
Hydra.
# PM/PMI: PMI2 client code is now officially included in the release.
# TEST SUITE: Support to run the MPICH2 test suite through valgrind.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r5025:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.2
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.2?action=follow_copy&rev=HEAD&stop_rev=5025&mode=follow_copy
===============================================================================
Changes in 1.1.1p1
===============================================================================
- OVERALL: Fixed an invalid read in the dataloop code for zero count types.
- OVERALL: Fixed several bugs in ch3:nemesis:mx (tickets #744,#760;
also change r5126).
- BUILD SYSTEM: Several fixes for functionality broken in 1.1.1 release,
including MPICH2LIB_xFLAGS and extra libraries living in $LIBS instead of
$LDFLAGS. Also, '-lpthread' should no longer be duplicated in link lines.
- BUILD SYSTEM: MPICH2 shared libraries are now compatible with glibc versioned
symbols on Linux, such as those present in the MX shared libraries.
- BUILD SYSTEM: Minor tweaks to improve compilation under the nvcc CUDA
compiler.
- PM/PMI: Fix mpd incompatibility with python2.3 introduced in mpich2-1.1.1.
- PM/PMI: Several fixes to hydra, including memory leak fixes and process
binding issues.
- TEST SUITE: Correct invalid arguments in the coll2 and coll3 tests.
- Several other minor bug fixes, memory leak fixes, and code cleanup. A full
list of changes is available using:
svn log -r5032:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.1.1p1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.1.1p1?action=follow_copy&rev=HEAD&stop_rev=5032&mode=follow_copy
===============================================================================
Changes in 1.1.1
===============================================================================
# OVERALL: Improved support for Boost MPI.
# PM/PMI: Significantly improved time taken by MPI_Init with Nemesis and MPD on
large numbers of processes.
# PM/PMI: Improved support for hybrid MPI-UPC program launching with
Hydra.
# PM/PMI: Improved support for process-core binding with Hydra.
# PM/PMI: Preliminary support for PMI-2. Currently supported only
with Hydra.
# Many other bug fixes, memory leak fixes and code cleanup. A full
list of changes is available using:
svn log -r4655:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.1.1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.1.1?action=follow_copy&rev=HEAD&stop_rev=4655&mode=follow_copy
===============================================================================
Changes in 1.1
===============================================================================
- OVERALL: Added MPI 2.1 support.
- OVERALL: Nemesis is now the default configuration channel with a
completely new TCP communication module.
- OVERALL: Windows support for nemesis.
- OVERALL: Added a new Myrinet MX network module for nemesis.
- OVERALL: Initial support for shared-memory aware collective
communication operations. Currently MPI_Bcast, MPI_Reduce, MPI_Allreduce,
and MPI_Scan.
- OVERALL: Improved handling of MPI Attributes.
- OVERALL: Support for BlueGene/P through the DCMF library (thanks to
IBM for the patch).
- OVERALL: Experimental support for fine-grained multithreading
- OVERALL: Added dynamic processes support for Nemesis.
- OVERALL: Added automatic as well as statically runtime configurable
receive timeout variation for MPD (thanks to OSU for the patch).
- OVERALL: Improved performance for MPI_Allgatherv, MPI_Gatherv, and MPI_Alltoall.
- PM/PMI: Initial support for the new Hydra process management
framework (current support is for ssh, rsh, fork and a preliminary
version of slurm).
- ROMIO: Added support for MPI_Type_create_resized and
MPI_Type_create_indexed_block datatypes in ROMIO.
- ROMIO: Optimized Lustre ADIO driver (thanks to Weikuan Yu for
initial work and Sun for further improvements).
- Many other bug fixes, memory leak fixes and code cleanup. A full
list of changes is available using:
svn log -r813:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.1?action=follow_copy&rev=HEAD&stop_rev=813&mode=follow_copy
Changes since 1.0.7:
- Added support for MPI 2.1
- Added support for MPI_Type_create_resized and
MPI_Type_create_indexed_block datatypes in ROMIO.
- Bug fixes, memory leak fixes and code cleanup.
MPICH2 is an all-new implementation of MPI from the group at Argonne
National Laboratory. It shares many goals with the original MPICH but
no actual code. It is a portable, high-performance implementation of
the entire MPI-2 standard. This release has all MPI-2 functions and
features required by the standard with the exception of support for the
"external32" portable I/O format.
An extensive list of changes since the last release is available online
at http://www-unix.mcs.anl.gov/mpi/mpich/r1_2_5changes.html
(It's difficult to destill the important changes from that file, which
is why I'm not pasting the entire collection in here.)
From the packages README:
Version 1.2.5 is primarily a bug-fix release. New features include
Significantly better collective algorithms are available.
Major changes for the globus2 device and the ch_nt device.
Information on previous versions
--------------------------------
Features new or improved in Version 1.2.4 included:
SMP support has been inproved through the use of sched_yield. SMP Systems that
are not oversubscribed can consider configuring with --disable-yeild.
A number of bugs in the ch_p4, ch_p4mpd, and ch_shmem device hae been fixed.
These could cause programs to hang or generate error messages.
Features new or improved in Version 1.2.3 included:
Bug fixes and improved-portability,
particularly in the area of Linux and Mac OS X support. Other new features
include better support for Intel compilers, particularly for IA64 Linux,
better support for various Fortran compilers, and some fixes for rare race
conditions in the ch_p4 device. The documentation on the ch_p4mpd device and
the MPD system has been improved. Version 1.2.3 also includes all of the
source for both the Unix and Windows versions. Windows users should still
use the self-installing version available through the web page. The globus
device now supports version 2 of Globus.
Features new or improved in Version 1.2.2 included:
There are a number of improvements to the ch_p4 device that improve both
performance and reliability, particularly on Linux platforms.
Significant upgrades have been made to the MPD system of daemons that provide
fast startup of MPICH jobs, management of stdio, and a crude parallel debugger
based on gdb.
packages.
mpi-ch:
a free implementation of the Message Passing Interface.
pvm3:
Paralell Virtual Machine libraries and environment
clusterit:
A clustering package for unix