| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697 | // Copyright (C) 2005-2006 Douglas Gregor <doug.gregor -at- gmail.com>.// Use, modification and distribution is subject to the Boost Software// License, Version 1.0. (See accompanying file LICENSE_1_0.txt or copy at// http://www.boost.org/LICENSE_1_0.txt)// Message Passing Interface 1.1 -- Section 4. MPI Collectives/** @file collectives.hpp * *  This header contains MPI collective operations, which implement *  various parallel algorithms that require the coordination of all *  processes within a communicator. The header @c collectives_fwd.hpp *  provides forward declarations for each of these operations. To *  include only specific collective algorithms, use the headers @c *  boost/mpi/collectives/algorithm_name.hpp. */#ifndef BOOST_MPI_COLLECTIVES_HPP#define BOOST_MPI_COLLECTIVES_HPP#include <boost/mpi/communicator.hpp>#include <boost/mpi/inplace.hpp>#include <vector>namespace boost { namespace mpi {/** *  @brief Gather the values stored at every process into vectors of *  values from each process. * *  @c all_gather is a collective algorithm that collects the values *  stored at each process into a vector of values indexed by the *  process number they came from. The type @c T of the values may be *  any type that is serializable or has an associated MPI data type. * *  When the type @c T has an associated MPI data type, this routine *  invokes @c MPI_Allgather to gather the values. * *    @param comm The communicator over which the all-gather will *    occur. * *    @param in_value The value to be transmitted by each process. To *    gather an array of values, @c in_values points to the @c n local *    values to be transmitted. * *    @param out_values A vector or pointer to storage that will be *    populated with the values from each process, indexed by the *    process ID number. If it is a vector, the vector will be resized *    accordingly. */template<typename T>voidall_gather(const communicator& comm, const T& in_value,            std::vector<T>& out_values);/** * \overload */template<typename T>voidall_gather(const communicator& comm, const T& in_value, T* out_values);/** * \overload */template<typename T>voidall_gather(const communicator& comm, const T* in_values, int n,           std::vector<T>& out_values);/** * \overload */template<typename T>voidall_gather(const communicator& comm, const T* in_values, int n, T* out_values);/** *  @brief Combine the values stored by each process into a single *  value available to all processes. * *  @c all_reduce is a collective algorithm that combines the values *  stored by each process into a single value available to all *  processes. The values are combined in a user-defined way, *  specified via a function object. The type @c T of the values may *  be any type that is serializable or has an associated MPI data *  type. One can think of this operation as a @c all_gather, followed *  by an @c std::accumulate() over the gather values and using the *  operation @c op. * *  When the type @c T has an associated MPI data type, this routine *  invokes @c MPI_Allreduce to perform the reduction. If possible, *  built-in MPI operations will be used; otherwise, @c all_reduce() *  will create a custom MPI_Op for the call to MPI_Allreduce. * *    @param comm The communicator over which the reduction will *    occur. *    @param value The local value to be combined with the local *    values of every other process. For reducing arrays, @c in_values *    is a pointer to the local values to be reduced and @c n is the *    number of values to reduce. See @c reduce for more information. * *    If wrapped in a @c inplace_t object, combine the usage of both *    input and $c out_value and the local value will be overwritten *    (a convenience function @c inplace is provided for the wrapping). * *    @param out_value Will receive the result of the reduction *    operation. If this parameter is omitted, the outgoing value will *    instead be returned. * *    @param op The binary operation that combines two values of type *    @c T and returns a third value of type @c T. For types @c T that has *    ssociated MPI data types, @c op will either be translated into *    an @c MPI_Op (via @c MPI_Op_create) or, if possible, mapped *    directly to a built-in MPI operation. See @c is_mpi_op in the @c *    operations.hpp header for more details on this mapping. For any *    non-built-in operation, commutativity will be determined by the *    @c is_commmutative trait (also in @c operations.hpp): users are *    encouraged to mark commutative operations as such, because it *    gives the implementation additional lattitude to optimize the *    reduction operation. * *    @param n Indicated the size of the buffers of array type. *    @returns If no @p out_value parameter is supplied, returns the *    result of the reduction operation. */template<typename T, typename Op>voidall_reduce(const communicator& comm, const T* value, int n, T* out_value,            Op op);/** * \overload */template<typename T, typename Op>voidall_reduce(const communicator& comm, const T& value, T& out_value, Op op);/** * \overload */template<typename T, typename Op>T all_reduce(const communicator& comm, const T& value, Op op);/** * \overload */template<typename T, typename Op>voidall_reduce(const communicator& comm, inplace_t<T*> value, int n,           Op op);/** * \overload */template<typename T, typename Op>voidall_reduce(const communicator& comm, inplace_t<T> value, Op op);/** *  @brief Send data from every process to every other process. * *  @c all_to_all is a collective algorithm that transmits @c p values *  from every process to every other process. On process i, jth value *  of the @p in_values vector is sent to process j and placed in the *  ith position of the @p out_values vector in process @p j. The type *  @c T of the values may be any type that is serializable or has an *  associated MPI data type. If @c n is provided, then arrays of @p n *  values will be transferred from one process to another. * *  When the type @c T has an associated MPI data type, this routine *  invokes @c MPI_Alltoall to scatter the values. * *    @param comm The communicator over which the all-to-all *    communication will occur. * *    @param in_values A vector or pointer to storage that contains *    the values to send to each process, indexed by the process ID *    number. * *    @param out_values A vector or pointer to storage that will be *    updated to contain the values received from other processes. The *    jth value in @p out_values will come from the procss with rank j. */template<typename T>voidall_to_all(const communicator& comm, const std::vector<T>& in_values,           std::vector<T>& out_values);/** * \overload */template<typename T>void all_to_all(const communicator& comm, const T* in_values, T* out_values);/** * \overload */template<typename T>voidall_to_all(const communicator& comm, const std::vector<T>& in_values, int n,           std::vector<T>& out_values);/** * \overload */template<typename T>void all_to_all(const communicator& comm, const T* in_values, int n, T* out_values);/** * @brief Broadcast a value from a root process to all other * processes. * * @c broadcast is a collective algorithm that transfers a value from * an arbitrary @p root process to every other process that is part of * the given communicator. The @c broadcast algorithm can transmit any * Serializable value, values that have associated MPI data types, * packed archives, skeletons, and the content of skeletons; see the * @c send primitive for communicators for a complete list. The type * @c T shall be the same for all processes that are a part of the * communicator @p comm, unless packed archives are being transferred: * with packed archives, the root sends a @c packed_oarchive or @c * packed_skeleton_oarchive whereas the other processes receive a * @c packed_iarchive or @c packed_skeleton_iarchve, respectively. * * When the type @c T has an associated MPI data type, this routine * invokes @c MPI_Bcast to perform the broadcast. * *   @param comm The communicator over which the broadcast will *   occur. * *   @param value The value (or values, if @p n is provided) to be *   transmitted (if the rank of @p comm is equal to @p root) or *   received (if the rank of @p comm is not equal to @p root). When *   the @p value is a @c skeleton_proxy, only the skeleton of the *   object will be broadcast. In this case, the @p root will build a *   skeleton from the object help in the proxy and all of the *   non-roots will reshape the objects held in their proxies based on *   the skeleton sent from the root. * *   @param n When supplied, the number of values that the pointer @p *   values points to, for broadcasting an array of values. The value *   of @p n must be the same for all processes in @p comm. * *   @param root The rank/process ID of the process that will be *   transmitting the value. */template<typename T>void broadcast(const communicator& comm, T& value, int root);/** * \overload */template<typename T>void broadcast(const communicator& comm, T* values, int n, int root);/** * \overload */template<typename T>void broadcast(const communicator& comm, skeleton_proxy<T>& value, int root);/** * \overload */template<typename T>voidbroadcast(const communicator& comm, const skeleton_proxy<T>& value, int root);/** *  @brief Gather the values stored at every process into a vector at *  the root process. * *  @c gather is a collective algorithm that collects the values *  stored at each process into a vector of values at the @p root *  process. This vector is indexed by the process number that the *  value came from. The type @c T of the values may be any type that *  is serializable or has an associated MPI data type. * *  When the type @c T has an associated MPI data type, this routine *  invokes @c MPI_Gather to gather the values. * *    @param comm The communicator over which the gather will occur. * *    @param in_value The value to be transmitted by each process. For *    gathering arrays of values, @c in_values points to storage for *    @c n*comm.size() values. * *    @param out_values A vector or pointer to storage that will be *    populated with the values from each process, indexed by the *    process ID number. If it is a vector, it will be resized *    accordingly. For non-root processes, this parameter may be *    omitted. If it is still provided, however, it will be unchanged. * *    @param root The process ID number that will collect the *    values. This value must be the same on all processes. */template<typename T>voidgather(const communicator& comm, const T& in_value, std::vector<T>& out_values,       int root);/** * \overload */template<typename T>voidgather(const communicator& comm, const T& in_value, T* out_values, int root);/** * \overload */template<typename T>void gather(const communicator& comm, const T& in_value, int root);/** * \overload */template<typename T>voidgather(const communicator& comm, const T* in_values, int n,        std::vector<T>& out_values, int root);/** * \overload */template<typename T>voidgather(const communicator& comm, const T* in_values, int n, T* out_values,        int root);/** * \overload */template<typename T>void gather(const communicator& comm, const T* in_values, int n, int root);/** *  @brief Similar to boost::mpi::gather with the difference that the number *  of values to be send by non-root processes can vary. * *    @param comm The communicator over which the gather will occur. * *    @param in_values The array of values to be transmitted by each process. * *    @param in_size For each non-root process this specifies the size *    of @p in_values. * *    @param out_values A pointer to storage that will be populated with *    the values from each process. For non-root processes, this parameter *    may be omitted. If it is still provided, however, it will be unchanged. * *    @param sizes A vector containing the number of elements each non-root *    process will send. * *    @param displs A vector such that the i-th entry specifies the *    displacement (relative to @p out_values) from which to take the ingoing *    data at the @p root process. Overloaded versions for which @p displs is *    omitted assume that the data is to be placed contiguously at the root process. * *    @param root The process ID number that will collect the *    values. This value must be the same on all processes. */template<typename T>voidgatherv(const communicator& comm, const std::vector<T>& in_values,        T* out_values, const std::vector<int>& sizes, const std::vector<int>& displs,        int root);/** * \overload */template<typename T>voidgatherv(const communicator& comm, const T* in_values, int in_size,        T* out_values, const std::vector<int>& sizes, const std::vector<int>& displs,        int root);/** * \overload */template<typename T>void gatherv(const communicator& comm, const std::vector<T>& in_values, int root);/** * \overload */template<typename T>void gatherv(const communicator& comm, const T* in_values, int in_size, int root);/** * \overload */template<typename T>voidgatherv(const communicator& comm, const T* in_values, int in_size,        T* out_values, const std::vector<int>& sizes, int root);/** * \overload */template<typename T>voidgatherv(const communicator& comm, const std::vector<T>& in_values,        T* out_values, const std::vector<int>& sizes, int root);/** *  @brief Scatter the values stored at the root to all processes *  within the communicator. * *  @c scatter is a collective algorithm that scatters the values *  stored in the @p root process (inside a vector) to all of the *  processes in the communicator. The vector @p out_values (only *  significant at the @p root) is indexed by the process number to *  which the corresponding value will be sent. The type @c T of the *  values may be any type that is serializable or has an associated *  MPI data type. * *  When the type @c T has an associated MPI data type, this routine *  invokes @c MPI_Scatter to scatter the values. * *    @param comm The communicator over which the scatter will occur. * *    @param in_values A vector or pointer to storage that will contain *    the values to send to each process, indexed by the process rank. *    For non-root processes, this parameter may be omitted. If it is *    still provided, however, it will be unchanged. * *    @param out_value The value received by each process. When *    scattering an array of values, @p out_values points to the @p n *    values that will be received by each process. * *    @param root The process ID number that will scatter the *    values. This value must be the same on all processes. */template<typename T>voidscatter(const communicator& comm, const std::vector<T>& in_values, T& out_value,        int root);/** * \overload */template<typename T>voidscatter(const communicator& comm, const T* in_values, T& out_value, int root);/** * \overload */template<typename T>void scatter(const communicator& comm, T& out_value, int root);/** * \overload */template<typename T>voidscatter(const communicator& comm, const std::vector<T>& in_values,         T* out_values, int n, int root);/** * \overload */template<typename T>voidscatter(const communicator& comm, const T* in_values, T* out_values, int n,        int root);/** * \overload */template<typename T>void scatter(const communicator& comm, T* out_values, int n, int root);/** *  @brief Similar to boost::mpi::scatter with the difference that the number *  of values stored at the root process does not need to be a multiple of *  the communicator's size. * *    @param comm The communicator over which the scatter will occur. * *    @param in_values A vector or pointer to storage that will contain *    the values to send to each process, indexed by the process rank. *    For non-root processes, this parameter may be omitted. If it is *    still provided, however, it will be unchanged. * *    @param sizes A vector containing the number of elements each non-root *    process will receive. * *    @param displs A vector such that the i-th entry specifies the *    displacement (relative to @p in_values) from which to take the outgoing *    data to process i. Overloaded versions for which @p displs is omitted *    assume that the data is contiguous at the @p root process. * *    @param out_values The array of values received by each process. * *    @param out_size For each non-root process this will contain the size *    of @p out_values. * *    @param root The process ID number that will scatter the *    values. This value must be the same on all processes. */template<typename T>voidscatterv(const communicator& comm, const std::vector<T>& in_values,         const std::vector<int>& sizes, const std::vector<int>& displs,         T* out_values, int out_size, int root);/** * \overload */template<typename T>voidscatterv(const communicator& comm, const T* in_values,         const std::vector<int>& sizes, const std::vector<int>& displs,         T* out_values, int out_size, int root);/** * \overload */template<typename T>void scatterv(const communicator& comm, T* out_values, int out_size, int root);/** * \overload */template<typename T>voidscatterv(const communicator& comm, const T* in_values,         const std::vector<int>& sizes, T* out_values, int root);/** * \overload */template<typename T>voidscatterv(const communicator& comm, const std::vector<T>& in_values,         const std::vector<int>& sizes, T* out_values, int root);/** *  @brief Combine the values stored by each process into a single *  value at the root. * *  @c reduce is a collective algorithm that combines the values *  stored by each process into a single value at the @c root. The *  values can be combined arbitrarily, specified via a function *  object. The type @c T of the values may be any type that is *  serializable or has an associated MPI data type. One can think of *  this operation as a @c gather to the @p root, followed by an @c *  std::accumulate() over the gathered values and using the operation *  @c op.  * *  When the type @c T has an associated MPI data type, this routine *  invokes @c MPI_Reduce to perform the reduction. If possible, *  built-in MPI operations will be used; otherwise, @c reduce() will *  create a custom MPI_Op for the call to MPI_Reduce. * *    @param comm The communicator over which the reduction will *    occur. * *    @param in_value The local value to be combined with the local *    values of every other process. For reducing arrays, @c in_values *    contains a pointer to the local values. In this case, @c n is *    the number of values that will be reduced. Reduction occurs *    independently for each of the @p n values referenced by @p *    in_values, e.g., calling reduce on an array of @p n values is *    like calling @c reduce @p n separate times, one for each *    location in @p in_values and @p out_values. * *    @param out_value Will receive the result of the reduction *    operation, but only for the @p root process. Non-root processes *    may omit if parameter; if they choose to supply the parameter, *    it will be unchanged. For reducing arrays, @c out_values *    contains a pointer to the storage for the output values. * *    @param op The binary operation that combines two values of type *    @c T into a third value of type @c T. For types @c T that has *    ssociated MPI data types, @c op will either be translated into *    an @c MPI_Op (via @c MPI_Op_create) or, if possible, mapped *    directly to a built-in MPI operation. See @c is_mpi_op in the @c *    operations.hpp header for more details on this mapping. For any *    non-built-in operation, commutativity will be determined by the *    @c is_commmutative trait (also in @c operations.hpp): users are *    encouraged to mark commutative operations as such, because it *    gives the implementation additional lattitude to optimize the *    reduction operation. * *    @param root The process ID number that will receive the final, *    combined value. This value must be the same on all processes. */template<typename T, typename Op>voidreduce(const communicator& comm, const T& in_value, T& out_value, Op op,       int root);/** * \overload */template<typename T, typename Op>void reduce(const communicator& comm, const T& in_value, Op op, int root);/** * \overload */template<typename T, typename Op>voidreduce(const communicator& comm, const T* in_values, int n, T* out_values,        Op op, int root);/** * \overload */template<typename T, typename Op>void reduce(const communicator& comm, const T* in_values, int n, Op op, int root);/** *  @brief Compute a prefix reduction of values from all processes in *  the communicator. * *  @c scan is a collective algorithm that combines the values stored *  by each process with the values of all processes with a smaller *  rank. The values can be arbitrarily combined, specified via a *  function object @p op. The type @c T of the values may be any type *  that is serializable or has an associated MPI data type. One can *  think of this operation as a @c gather to some process, followed *  by an @c std::prefix_sum() over the gathered values using the *  operation @c op. The ith process returns the ith value emitted by *  @c std::prefix_sum(). * *  When the type @c T has an associated MPI data type, this routine *  invokes @c MPI_Scan to perform the reduction. If possible, *  built-in MPI operations will be used; otherwise, @c scan() will *  create a custom @c MPI_Op for the call to MPI_Scan. * *    @param comm The communicator over which the prefix reduction *    will occur. * *    @param in_value The local value to be combined with the local *    values of other processes. For the array variant, the @c *    in_values parameter points to the @c n local values that will be *    combined. * *    @param out_value If provided, the ith process will receive the *    value @c op(in_value[0], op(in_value[1], op(..., in_value[i]) *    ... )). For the array variant, @c out_values contains a pointer *    to storage for the @c n output values. The prefix reduction *    occurs independently for each of the @p n values referenced by *    @p in_values, e.g., calling scan on an array of @p n values is *    like calling @c scan @p n separate times, one for each location *    in @p in_values and @p out_values. * *    @param op The binary operation that combines two values of type *    @c T into a third value of type @c T. For types @c T that has *    ssociated MPI data types, @c op will either be translated into *    an @c MPI_Op (via @c MPI_Op_create) or, if possible, mapped *    directly to a built-in MPI operation. See @c is_mpi_op in the @c *    operations.hpp header for more details on this mapping. For any *    non-built-in operation, commutativity will be determined by the *    @c is_commmutative trait (also in @c operations.hpp). * *    @returns If no @p out_value parameter is provided, returns the *    result of prefix reduction. */template<typename T, typename Op>voidscan(const communicator& comm, const T& in_value, T& out_value, Op op);/** * \overload */template<typename T, typename Op>Tscan(const communicator& comm, const T& in_value, Op op);/** * \overload */template<typename T, typename Op>voidscan(const communicator& comm, const T* in_values, int n, T* out_values, Op op);} } // end namespace boost::mpi#endif // BOOST_MPI_COLLECTIVES_HPP#ifndef BOOST_MPI_COLLECTIVES_FORWARD_ONLY// Include implementations of each of the collectives#  include <boost/mpi/collectives/all_gather.hpp>#  include <boost/mpi/collectives/all_reduce.hpp>#  include <boost/mpi/collectives/all_to_all.hpp>#  include <boost/mpi/collectives/broadcast.hpp>#  include <boost/mpi/collectives/gather.hpp>#  include <boost/mpi/collectives/gatherv.hpp>#  include <boost/mpi/collectives/scatter.hpp>#  include <boost/mpi/collectives/scatterv.hpp>#  include <boost/mpi/collectives/reduce.hpp>#  include <boost/mpi/collectives/scan.hpp>#endif
 |