* AI_V4MAPPED is not defined on NetBSD, so conditionalize its use.
* Force the use of pkgsrc GCC (here we require >= gcc 10) so that
<ISO_Fortran_binding.h> is made available to the C compiler.
* Sort PLIST.
The result builds and installs on amd64 and i386, but not on powerpc
due to Fortran's "real128" not being available there (CPU limitation).
This also removes the MPI_PREFIX business. No parallel installation of multiple
MPIs intended right now. Also, the f90 option is default now to support upcoming
parallel-computing packages in the default setup.
===============================================================================
Changes in 3.4.3
===============================================================================
# Fix bugs in GPFS ROMIO driver
# Fix bugs in DAOS ROMIO driver
# Fix nemesis shm bug on systems with weak memory consistency
# Fix progress issue in ch4:ucx finalize
# Fix nonblocking collective issue during finalize
# Fix UBSan warnings in 32-bit builds
# Fix predefined datatype handle leaks
# Fix bug in topology-aware communicator creation
# Fix Slurm nodelist parsing
# Fix testcase for out-of-tree configuration
# Fix truncation testcase for possible disconnection error
# Hydra help message improvements
# Add support for DAOS pool and container labels
===============================================================================
Changes in 3.4.2
===============================================================================
# Update configure options to support --disable-ze and --disable-cuda
to explicitly disable GPU-awareness
# Update embedded hwloc to version 2.4.1. Fixes a build issue on
systems with OpenCL.
# Fixes for Argobots thread package support.
# Fixes for the DAOS ROMIO driver (thanks to Intel for contributing)
# Fix additional bugs in GPU-aware nonblocking op collectives
# Fix 'make check' errors with embedded MPL and json-c libraries
# Fix support for GNI provider in ch4:ofi netmod
# Fix bug in dynamic window memory registration with ch4:ofi netmod
# Fix bug in persistent pt2pt when used with MPI_PROC_NULL
# Fix bug in hydra nameserver when freeing published entries
# Fix bugs in GPFS ROMIO driver (thanks to IBM for contributing)
# Fix bug in darray type creation (thanks to IBM for contributing)
# Fix bugs in XPMEM support
# Fix bug in MPI_Comm_spawn_multiple where non-root args were not ignored
===============================================================================
Changes in 3.4.1
===============================================================================
# Fix bug in GPU memory hooks that caused failures with NCCL
# Fix bug in GPU-aware nonblocking op collectives
# Hide symbols from embedded json-c module
# Removed anonymous struct/union usage in MPL
# Disable Java support in embedded UCX module
===============================================================================
Changes in 3.4
===============================================================================
# ch4 replaces ch3 as the default device configuration. If no network
module is specified at configuration-time, MPICH will search the
user environment in order to select one to build. The user will be
prompted to choose if no preferred network library is detected.
# Add support for Yaksa datatype engine (default in ch4).
# Add support for GPU buffers (CUDA, Level Zero) in pt2pt,
collectives, and one-sided communication.
# Add support for XPMEM.
# Add support for multiple virtual communication interfaces for more
efficient MPI_THREAD_MULTIPLE (experimental).
# Add DAOS ADIO driver to ROMIO (contributed by Intel).
# Add Quobyte ADIO driver to ROMIO (contributed by Quobyte).
# Add support for Arm compiler toolchain
# Add support for NVIDIA HPC compilers
# Add support for flang/f18 Fortran compiler
# Add support for AddressSanitizer and UndefinedBehaviorSanitizer to
debug configuration
# Remove mxm, llc, and portals4 netmods from ch3.
# Remove support for logical reduction operations on floating point
types.
# Remove MPIX_Mutex interfaces.
# Further improvements to ch4 business card exchange: extra
long address support and fixes for PMIx integration.
# Un-inline non-critical ch4 code for improved build times.
# Fix several test program bugs.
# Fix several static analysis and compiler warnings.
# Change the signature of MPID_Init to include requested and provided
thread levels.
===============================================================================
Changes in 3.3.2
===============================================================================
# Add support for struct sockaddr in MPICH, Hydra, and PMI socket
code. Works with both IPv4 and IPv6 addresses.
# Fix localhost detection on FreeBSD and macOS, avoiding long delay
during startup.
# Fix thread-local storage detection.
# Fix several test program bugs.
# Fix several static analysis and compiler warnings.
===============================================================================
Changes in 3.3.1
===============================================================================
# Fix bug in MPI_Testany/MPI_Waitany that could cause deadlock
# Add missing functionality in Argobots library support
# Fix configure-time detection for thread local storage support
# Better support for reproducible builds. Thanks to Bernhard
Wiedemann for the report and fixes
# Fix support for XL compiler toolchain
# Add support for -static-intel linking option
# Fix building on systems without weak symbols
# Fix several static analysis and compiler warnings
===============================================================================
Changes in 3.3
===============================================================================
# CH4 Device: A new device layer implementation designed for low software
overheads. CH4 has experimental support for OFI and UCX network libraries,
and POSIX shared memory. Thanks to Intel, Mellanox, and RIKEN AICS for
participating in the CH4 coding effort.
# Fixed SLURM integration in Hydra for new node list format.
# Added support for PMIx (https://pmix.github.io/pmix/) client
library in CH4 netmods. Note that you must use a compatible PMIx
server in this configuration.
# Better organization of collectives in the MPI layer. The new
scheme, which de-couples implementation from selection logic,
enables easier integration of additional algorithms.
# TSP collectives framework: A C++-template style framework for
collective algorithms is added to allow single collective
implementation to move data over generic or device-specific
transport functions.
# Improvements to derived datatype testing (DTPools -
https://wiki.mpich.org/mpich/index.php/DTPools).
# Added new "non-catastrophic" error codes to expose internal
resource exhaustion.
# Added info hints to MPI_Comm_split_type to support splitting
communicators by machine topology. Both on-node (socket, core,
etc.) and off-node (switch-level) hints are defined.
# Improvements to MPI_THREAD_MULTIPLE in CH4 through new thread safety
models at the Virtual Network Interface (VNI) level. This introduces two
new models that leverage work-queues to offload operations and improve
scalability under contention.
# Message Driven Thread Activation (MDTA). An alternative locking
model is defined for MPI_THREAD_MULTIPLE in CH4.
# Added PMI usage optimizations for business card exchange in CH4
netmods.
# Improvements on MPI_Abort. MPI_Abort invoked on subcommunicators will
only abort the connected processes within that communicator.
`
# Cleanup of whitespace (ch3 excluded) using the
maint/code-cleanup.sh script. For instructions on how to update
PRs/branches based on MPICH before the cleanup, see
https://github.com/pmodels/mpich/wiki/Code-Cleanup-Procedure.
# Removed the PAMI device and poe PMI client.
# C99 compiler support is now required to build MPICH.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available at the following link:
http://git.mpich.org/mpich.git/shortlog/v3.2..v3.3
A list of bugs that have been fixed is available at the following
link:
https://github.com/pmodels/mpich/milestone/25?closed=1
The binaries have the WRKDIR in strings, printed via the -info
argument to show how it was built. (I don't understand how this is
buidling for anyone else.)
pkglint -r --network --only "migrate"
As a side-effect of migrating the homepages, pkglint also fixed a few
indentations in unrelated lines. These and the new homepages have been
checked manually.
Patch from Eric A. Borisch via pull request.
ClosesNetBSD/pkgsrc#38.
Remove patches no longer needed; migrate to configure option for rpath.
===============================================================================
Changes in 3.2.1
===============================================================================
# Fixes for platforms with strict memory alignment requirements.
# Fixes for MPI_Win info management.
# Fixed a progress bug with MPI generalized requests.
# Fixed multiple integer overflow bugs in CH3 and ROMIO.
# Improved detection for Fortran 2008 binding support.
# Enhanced support for libfabric (OFI) netmod.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available at the following link:
http://git.mpich.org/mpich.git/shortlog/v3.2..v3.2.1
* Added support for MPI-3.1 features including nonblocking collective I/O,
address manipulation routines, thread-safety for MPI initialization,
pre-init functionality, and new MPI_T routines to look up variables
by name.
* Fortran 2008 bindings are enabled by default and fully supported.
* Added support for the Mellanox MXM InfiniBand interface. (thanks
to Mellanox for the code contribution).
* Added support for the Mellanox HCOLL interface for collectives.
(thanks to Mellanox for the code contribution).
* Significant stability improvements to the MPICH/portals4
implementation.
* Completely revamped RMA infrastructure including several
scalability improvements, performance improvements, and bug fixes.
* Added experimental support for Open Fabrics Interfaces (OFI) version 1.0.0.
https://github.com/ofiwg/libfabric (thanks to Intel for code contribution)
* The Myrinet MX network module, which had a life cyle from 1.1 till
3.1.2, has now been deleted.
* Several other minor bug fixes, memory leak fixes, and code cleanup.
Do it for all packages that
* mention perl, or
* have a directory name starting with p5-*, or
* depend on a package starting with p5-
like last time, for 5.18, where this didn't lead to complaints.
Let me know if you have any this time.
a) refer 'perl' in their Makefile, or
b) have a directory name of p5-*, or
c) have any dependency on any p5-* package
Like last time, where this caused no complaints.
This release adds several performance features for the Hydra process manager, support for communicator-specific tuning of eager/rendezvous thresholds, and fixes several bugs present in 3.0.3. Please use this release instead of 3.0.3.
A new stable release of MPICH, 3.0.3, is now available for download. This release adds several performance features for MPI-RMA and fixes several bugs present in 3.0.2. Please use this release instead of 3.0.2.
===============================================================================
Changes in 3.0
===============================================================================
# MPI-3: All MPI-3 features are now implemented and the MPI_VERSION
bumped up to 3.0.
# OVERALL: Added support for ARM-v7 native atomics
# MPE: MPE is now separated out of MPICH and can be downloaded/used
as a separate package.
# PM/PMI: Upgraded to hwloc-1.6
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r10344:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich-3.0
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich-3.0?action=follow_copy&rev=HEAD&stop_rev=10344&mode=follow_copy
===============================================================================
Changes in 1.5
===============================================================================
# OVERALL: Nemesis now supports an "--enable-yield=..." configure
option for better performance/behavior when oversubscribing
processes to cores. Some form of this option is enabled by default
on Linux, Darwin, and systems that support sched_yield().
# OVERALL: Added support for Intel Many Integrated Core (MIC)
architecture: shared memory, TCP/IP, and SCIF based communication.
# OVERALL: Added support for IBM BG/Q architecture. Thanks to IBM
for the contribution.
# MPI-3: const support has been added to mpi.h, although it is
disabled by default. It can be enabled on a per-translation unit
basis with "#define MPICH2_CONST const".
# MPI-3: Added support for MPIX_Type_create_hindexed_block.
# MPI-3: The new MPI-3 nonblocking collective functions are now
available as "MPIX_" functions (e.g., "MPIX_Ibcast").
# MPI-3: The new MPI-3 neighborhood collective routines are now available as
"MPIX_" functions (e.g., "MPIX_Neighbor_allgather").
# MPI-3: The new MPI-3 MPI_Comm_split_type function is now available
as an "MPIX_" function.
# MPI-3: The new MPI-3 tools interface is now available as "MPIX_T_"
functions. This is a beta implementation right now with several
limitations, including no support for multithreading. Several
performance variables related to CH3's message matching are exposed
through this interface.
# MPI-3: The new MPI-3 matched probe functionality is supported via
the new routines MPIX_Mprobe, MPIX_Improbe, MPIX_Mrecv, and
MPIX_Imrecv.
# MPI-3: The new MPI-3 nonblocking communicator duplication routine,
MPIX_Comm_idup, is now supported. It will only work for
single-threaded programs at this time.
# MPI-3: MPIX_Comm_reenable_anysource support
# MPI-3: Native MPIX_Comm_create_group support (updated version of
the prior MPIX_Group_comm_create routine).
# MPI-3: MPI_Intercomm_create's internal communication no longer interferes
with point-to-point communication, even if point-to-point operations on the
parent communicator use the same tag or MPI_ANY_TAG.
# MPI-3: Eliminated the possibility of interference between
MPI_Intercomm_create and point-to-point messaging operations.
# Build system: Completely revamped build system to rely fully on
autotools. Parallel builds ("make -j8" and similar) are now supported.
# Build system: rename "./maint/updatefiles" --> "./autogen.sh" and
"configure.in" --> "configure.ac"
# JUMPSHOT: Improvements to Jumpshot to handle thousands of
timelines, including performance improvements to slog2 in such
cases.
# JUMPSHOT: Added navigation support to locate chosen drawable's ends
when viewport has been scrolled far from the drawable.
# PM/PMI: Added support for memory binding policies.
# PM/PMI: Various improvements to the process binding support in
Hydra. Several new pre-defined binding options are provided.
# PM/PMI: Upgraded to hwloc-1.5
# PM/PMI: Several improvements to PBS support to natively use the PBS
launcher.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r8478:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.5
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.5?action=follow_copy&rev=HEAD&stop_rev=8478&mode=follow_copy
===============================================================================
Changes in 1.4.1
===============================================================================
# OVERALL: Several improvements to the ARMCI API implementation
within MPICH2.
# Build system: Added beta support for DESTDIR while installing
MPICH2.
# PM/PMI: Upgrade hwloc to 1.2.1rc2.
# PM/PMI: Initial support for the PBS launcher.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r8675:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.4.1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.4.1?action=follow_copy&rev=HEAD&stop_rev=8675&mode=follow_copy
===============================================================================
Changes in 1.4
===============================================================================
# OVERALL: Improvements to fault tolerance for collective
operations. Thanks to Rui Wang @ ICT for reporting several of these
issues.
# OVERALL: Improvements to the universe size detection. Thanks to
Yauheni Zelenko for reporting this issue.
# OVERALL: Bug fixes for Fortran attributes on some systems. Thanks
to Nicolai Stange for reporting this issue.
# OVERALL: Added new ARMCI API implementation (experimental).
# OVERALL: Added new MPIX_Group_comm_create function to allow
non-collective creation of sub-communicators.
# FORTRAN: Bug fixes in the MPI_DIST_GRAPH_ Fortran bindings.
# PM/PMI: Support for a manual "none" launcher in Hydra to allow for
higher-level tools to be built on top of Hydra. Thanks to Justin
Wozniak for reporting this issue, for providing several patches for
the fix, and testing it.
# PM/PMI: Bug fixes in Hydra to handle non-uniform layouts of hosts
better. Thanks to the MVAPICH group at OSU for reporting this issue
and testing it.
# PM/PMI: Bug fixes in Hydra to handle cases where only a subset of
the available launchers or resource managers are compiled
in. Thanks to Satish Balay @ Argonne for reporting this issue.
# PM/PMI: Support for a different username to be provided for each
host; this only works for launchers that support this (such as
SSH).
# PM/PMI: Bug fixes for using Hydra on AIX machines. Thanks to
Kitrick Sheets @ NCSA for reporting this issue and providing the
first draft of the patch.
# PM/PMI: Bug fixes in memory allocation/management for environment
variables that was showing up on older platforms. Thanks to Steven
Sutphen for reporting the issue and providing detailed analysis to
track down the bug.
# PM/PMI: Added support for providing a configuration file to pick
the default options for Hydra. Thanks to Saurabh T. for reporting
the issues with the current implementation and working with us to
improve this option.
# PM/PMI: Improvements to the error code returned by Hydra.
# PM/PMI: Bug fixes for handling "=" in environment variable values in
hydra.
# PM/PMI: Upgrade the hwloc version to 1.2.
# COLLECTIVES: Performance and memory usage improvements for MPI_Bcast
in certain cases.
# VALGRIND: Fix incorrect Valgrind client request usage when MPICH2 is
built for memory debugging.
# BUILD SYSTEM: "--enable-fast" and "--disable-error-checking" are once
again valid simultaneous options to configure.
# TEST SUITE: Several new tests for MPI RMA operations.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r7838:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.4
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.4?action=follow_copy&rev=HEAD&stop_rev=7838&mode=follow_copy
===============================================================================
Changes in 1.3.2
===============================================================================
# OVERALL: MPICH2 now recognizes the OSX mach_absolute_time as a
native timer type.
# OVERALL: Performance improvements to MPI_Comm_split on large
systems.
# OVERALL: Several improvements to error returns capabilities in the
presence of faults.
# PM/PMI: Several fixes and improvements to Hydra's process binding
capability.
# PM/PMI: Upgrade the hwloc version to 1.1.1.
# PM/PMI: Allow users to sort node lists allocated by resource
managers in Hydra.
# PM/PMI: Improvements to signal handling. Now Hydra respects Ctrl-Z
signals and passes on the signal to the application.
# PM/PMI: Improvements to STDOUT/STDERR handling including improved
support for rank prepending on output. Improvements to STDIN
handling for applications being run in the background.
# PM/PMI: Split the bootstrap servers into "launchers" and "resource
managers", allowing the user to pick a different resource manager
from the launcher. For example, the user can now pick the "SLURM"
resource manager and "SSH" as the launcher.
# PM/PMI: The MPD process manager is deprecated.
# PM/PMI: The PLPA process binding library support is deprecated.
# WINDOWS: Adding support for gfortran and 64-bit gcc libs.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r7457:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.3.2
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.3.2?action=follow_copy&rev=HEAD&stop_rev=7457&mode=follow_copy
Changes in 1.3.1
# OVERALL: MPICH2 is now fully compliant with the CIFTS FTB standard
MPI events (based on the draft standard).
# OVERALL: Major improvements to RMA performance for long lists of
RMA operations.
# OVERALL: Performance improvements for Group_translate_ranks.
# COLLECTIVES: Collective algorithm selection thresholds can now be controlled
at runtime via environment variables.
# ROMIO: PVFS error codes are now mapped to MPI error codes.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r7350:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.3.1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.3.1?action=follow_copy&rev=HEAD&stop_rev=7350&mode=follow_copy
Changes in 1.3
# OVERALL: Initial support for fine-grained threading in
ch3:nemesis:tcp.
# OVERALL: Support for Asynchronous Communication Progress.
# OVERALL: The ssm and shm channels have been removed.
# OVERALL: Checkpoint/restart support using BLCR.
# OVERALL: Improved tolerance to process and communication failures
when error handler is set to MPI_ERRORS_RETURN. If a communication
operation fails (e.g., due to a process failure) MPICH2 will return
an error, and further communication to that process is not
possible. However, communication with other processes will still
proceed normally. Note, however, that the behavior collective
operations on communicators containing the failed process is
undefined, and may give incorrect results or hang some processes.
# OVERALL: Experimental support for inter-library dependencies.
# PM/PMI: Hydra is now the default process management framework
replacing MPD.
# PM/PMI: Added dynamic process support for Hydra.
# PM/PMI: Added support for LSF, SGE and POE in Hydra.
# PM/PMI: Added support for CPU and memory/cache topology aware
process-core binding.
# DEBUGGER: Improved support and bug fixes in the Totalview support.
# Build system: Replaced F90/F90FLAGS by FC/FCFLAGS. F90/F90FLAGS are
not longer supported in the configure.
# Multi-compiler support: On systems where C compiler that is used to
build mpich2 libraries supports multiple weak symbols and multiple aliases,
the Fortran binding built in the mpich2 libraries can handle different
Fortran compilers (than the one used to build mpich2). Details in README.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r5762:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.3
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.3?action=follow_copy&rev=HEAD&stop_rev=5762&mode=follow_copy
===============================================================================
Changes in 1.2.1
===============================================================================
# OVERALL: Improved support for fine-grained multithreading.
# OVERALL: Improved integration with Valgrind for debugging builds of MPICH2.
# PM/PMI: Initial support for hwloc process-core binding library in
Hydra.
# PM/PMI: Updates to the PMI-2 code to match the PMI-2 API and
wire-protocol draft.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r5425:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.2.1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.2.1?action=follow_copy&rev=HEAD&stop_rev=5425&mode=follow_copy
===============================================================================
Changes in 1.2
===============================================================================
# OVERALL: Support for MPI-2.2
# OVERALL: Several fixes to Nemesis/MX.
# WINDOWS: Performance improvements to Nemesis/windows.
# PM/PMI: Scalability and performance improvements to Hydra using
PMI-1.1 process-mapping features.
# PM/PMI: Support for process-binding for hyperthreading enabled
systems in Hydra.
# PM/PMI: Initial support for PBS as a resource management kernel in
Hydra.
# PM/PMI: PMI2 client code is now officially included in the release.
# TEST SUITE: Support to run the MPICH2 test suite through valgrind.
# Several other minor bug fixes, memory leak fixes, and code cleanup.
A full list of changes is available using:
svn log -r5025:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.2
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.2?action=follow_copy&rev=HEAD&stop_rev=5025&mode=follow_copy
===============================================================================
Changes in 1.1.1p1
===============================================================================
- OVERALL: Fixed an invalid read in the dataloop code for zero count types.
- OVERALL: Fixed several bugs in ch3:nemesis:mx (tickets #744,#760;
also change r5126).
- BUILD SYSTEM: Several fixes for functionality broken in 1.1.1 release,
including MPICH2LIB_xFLAGS and extra libraries living in $LIBS instead of
$LDFLAGS. Also, '-lpthread' should no longer be duplicated in link lines.
- BUILD SYSTEM: MPICH2 shared libraries are now compatible with glibc versioned
symbols on Linux, such as those present in the MX shared libraries.
- BUILD SYSTEM: Minor tweaks to improve compilation under the nvcc CUDA
compiler.
- PM/PMI: Fix mpd incompatibility with python2.3 introduced in mpich2-1.1.1.
- PM/PMI: Several fixes to hydra, including memory leak fixes and process
binding issues.
- TEST SUITE: Correct invalid arguments in the coll2 and coll3 tests.
- Several other minor bug fixes, memory leak fixes, and code cleanup. A full
list of changes is available using:
svn log -r5032:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.1.1p1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.1.1p1?action=follow_copy&rev=HEAD&stop_rev=5032&mode=follow_copy
===============================================================================
Changes in 1.1.1
===============================================================================
# OVERALL: Improved support for Boost MPI.
# PM/PMI: Significantly improved time taken by MPI_Init with Nemesis and MPD on
large numbers of processes.
# PM/PMI: Improved support for hybrid MPI-UPC program launching with
Hydra.
# PM/PMI: Improved support for process-core binding with Hydra.
# PM/PMI: Preliminary support for PMI-2. Currently supported only
with Hydra.
# Many other bug fixes, memory leak fixes and code cleanup. A full
list of changes is available using:
svn log -r4655:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.1.1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.1.1?action=follow_copy&rev=HEAD&stop_rev=4655&mode=follow_copy
===============================================================================
Changes in 1.1
===============================================================================
- OVERALL: Added MPI 2.1 support.
- OVERALL: Nemesis is now the default configuration channel with a
completely new TCP communication module.
- OVERALL: Windows support for nemesis.
- OVERALL: Added a new Myrinet MX network module for nemesis.
- OVERALL: Initial support for shared-memory aware collective
communication operations. Currently MPI_Bcast, MPI_Reduce, MPI_Allreduce,
and MPI_Scan.
- OVERALL: Improved handling of MPI Attributes.
- OVERALL: Support for BlueGene/P through the DCMF library (thanks to
IBM for the patch).
- OVERALL: Experimental support for fine-grained multithreading
- OVERALL: Added dynamic processes support for Nemesis.
- OVERALL: Added automatic as well as statically runtime configurable
receive timeout variation for MPD (thanks to OSU for the patch).
- OVERALL: Improved performance for MPI_Allgatherv, MPI_Gatherv, and MPI_Alltoall.
- PM/PMI: Initial support for the new Hydra process management
framework (current support is for ssh, rsh, fork and a preliminary
version of slurm).
- ROMIO: Added support for MPI_Type_create_resized and
MPI_Type_create_indexed_block datatypes in ROMIO.
- ROMIO: Optimized Lustre ADIO driver (thanks to Weikuan Yu for
initial work and Sun for further improvements).
- Many other bug fixes, memory leak fixes and code cleanup. A full
list of changes is available using:
svn log -r813:HEAD https://svn.mcs.anl.gov/repos/mpi/mpich2/tags/release/mpich2-1.1
... or at the following link:
https://trac.mcs.anl.gov/projects/mpich2/log/mpich2/tags/release/mpich2-1.1?action=follow_copy&rev=HEAD&stop_rev=813&mode=follow_copy
Changes since 1.0.7:
- Added support for MPI 2.1
- Added support for MPI_Type_create_resized and
MPI_Type_create_indexed_block datatypes in ROMIO.
- Bug fixes, memory leak fixes and code cleanup.
MPICH2 is an all-new implementation of MPI from the group at Argonne
National Laboratory. It shares many goals with the original MPICH but
no actual code. It is a portable, high-performance implementation of
the entire MPI-2 standard. This release has all MPI-2 functions and
features required by the standard with the exception of support for the
"external32" portable I/O format.