| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225 | // Copyright (C) 2004-2006 The Trustees of Indiana University.// Copyright (C) 2002 Brad King and Douglas Gregor// Use, modification and distribution is subject to the Boost Software// License, Version 1.0. (See accompanying file LICENSE_1_0.txt or copy at// http://www.boost.org/LICENSE_1_0.txt)//  Authors: Douglas Gregor//           Andrew Lumsdaine//           Brian Barrett#ifndef BOOST_PARALLEL_GRAPH_PAGE_RANK_HPP#define BOOST_PARALLEL_GRAPH_PAGE_RANK_HPP#ifndef BOOST_GRAPH_USE_MPI#error "Parallel BGL files should not be included unless <boost/graph/use_mpi.hpp> has been included"#endif#include <boost/assert.hpp>#include <boost/graph/overloading.hpp>#include <boost/graph/page_rank.hpp>#include <boost/graph/distributed/concepts.hpp>#include <boost/property_map/parallel/distributed_property_map.hpp>#include <boost/property_map/parallel/caching_property_map.hpp>#include <boost/graph/parallel/algorithm.hpp>#include <boost/graph/parallel/container_traits.hpp>// #define WANT_MPI_ONESIDED 1namespace boost { namespace graph { namespace distributed {namespace detail {#ifdef WANT_MPI_ONESIDED  template<typename Graph, typename RankMap, typename owner_map_t>  void page_rank_step(const Graph& g, RankMap from_rank, MPI_Win to_win,                       typename property_traits<RankMap>::value_type damping,                      owner_map_t owner)  {    typedef typename property_traits<RankMap>::value_type rank_type;    int me, ret;    MPI_Comm_rank(MPI_COMM_WORLD, &me);    // MPI_Accumulate is not required to store the value of the data    // being sent, only the address.  The value of the memory location    // must not change until the end of the access epoch, meaning the    // call to MPI_Fence.  We therefore store the updated value back    // into the from_rank map before the accumulate rather than using    // a temporary.  We're going to reset the values in the from_rank    // before the end of page_rank_step() anyway, so this isn't a huge    // deal.  But MPI-2 One-sided is an abomination.    BGL_FORALL_VERTICES_T(u, g, Graph) {      put(from_rank, u, (damping * get(from_rank, u) / out_degree(u, g)));      BGL_FORALL_ADJ_T(u, v, g, Graph) {        ret = MPI_Accumulate(&(from_rank[u]),                             1, MPI_DOUBLE,                             get(owner, v), local(v),                              1, MPI_DOUBLE, MPI_SUM, to_win);        BOOST_ASSERT(MPI_SUCCESS == ret);      }    }    MPI_Win_fence(0, to_win);    // Set new rank maps for the other map.  Do this now to get around    // the stupid synchronization rules of MPI-2 One-sided    BGL_FORALL_VERTICES_T(v, g, Graph) put(from_rank, v, rank_type(1 - damping));  }#endif  template<typename T>  struct rank_accumulate_reducer {    BOOST_STATIC_CONSTANT(bool, non_default_resolver = true);    template<typename K>    T operator()(const K&) const { return T(0); }    template<typename K>    T operator()(const K&, const T& x, const T& y) const { return x + y; }  };} // end namespace detailtemplate<typename Graph, typename RankMap, typename Done, typename RankMap2>voidpage_rank_impl(const Graph& g, RankMap rank_map, Done done,                typename property_traits<RankMap>::value_type damping,               typename graph_traits<Graph>::vertices_size_type n,               RankMap2 rank_map2){  typedef typename property_traits<RankMap>::value_type rank_type;  int me;  MPI_Comm_rank(MPI_COMM_WORLD, &me);  typename property_map<Graph, vertex_owner_t>::const_type    owner = get(vertex_owner, g);  (void)owner;  typedef typename boost::graph::parallel::process_group_type<Graph>    ::type process_group_type;  typedef typename process_group_type::process_id_type process_id_type;  process_group_type pg = process_group(g);  process_id_type id = process_id(pg);  BOOST_ASSERT(me == id);  rank_type initial_rank = rank_type(rank_type(1) / n);  BGL_FORALL_VERTICES_T(v, g, Graph) put(rank_map, v, initial_rank);#ifdef WANT_MPI_ONESIDED  BOOST_ASSERT(sizeof(rank_type) == sizeof(double));  bool to_map_2 = true;  MPI_Win win, win2;  MPI_Win_create(&(rank_map[*(vertices(g).first)]),                  sizeof(double) * num_vertices(g),                 sizeof(double),                  MPI_INFO_NULL, MPI_COMM_WORLD, &win);  MPI_Win_set_name(win, "rank_map_win");  MPI_Win_create(&(rank_map2[*(vertices(g).first)]),                  sizeof(double) * num_vertices(g),                 sizeof(double),                  MPI_INFO_NULL, MPI_COMM_WORLD, &win2);  MPI_Win_set_name(win, "rank_map2_win");  // set initial rank maps for the first iteration...  BGL_FORALL_VERTICES_T(v, g, Graph) put(rank_map2, v, rank_type(1 - damping));  MPI_Win_fence(0, win);  MPI_Win_fence(0, win2);  while ((to_map_2 && !done(rank_map, g)) ||         (!to_map_2 && !done(rank_map2, g))) {    if (to_map_2) {      graph::distributed::detail::page_rank_step(g, rank_map, win2, damping, owner);      to_map_2 = false;    } else {      graph::distributed::detail::page_rank_step(g, rank_map2, win, damping, owner);      to_map_2 = true;    }  }  synchronize(boost::graph::parallel::process_group(g));  MPI_Win_free(&win);  MPI_Win_free(&win2);#else    // The ranks accumulate after each step.  rank_map.set_reduce(detail::rank_accumulate_reducer<rank_type>());  rank_map2.set_reduce(detail::rank_accumulate_reducer<rank_type>());  rank_map.set_consistency_model(boost::parallel::cm_flush | boost::parallel::cm_reset);  rank_map2.set_consistency_model(boost::parallel::cm_flush | boost::parallel::cm_reset);  bool to_map_2 = true;  while ((to_map_2 && !done(rank_map, g)) ||         (!to_map_2 && !done(rank_map2, g))) {    /**     * PageRank can implemented slightly more efficiently on a     * bidirectional graph than on an incidence graph. However,     * distributed PageRank requires that we have the rank of the     * source vertex available locally, so we force the incidence     * graph implementation, which pushes rank from source to     * target.     */    typedef incidence_graph_tag category;    if (to_map_2) {      graph::detail::page_rank_step(g, rank_map, rank_map2, damping,                                    category());      to_map_2 = false;    } else {      graph::detail::page_rank_step(g, rank_map2, rank_map, damping,                                    category());      to_map_2 = true;    }    using boost::graph::parallel::process_group;    synchronize(process_group(g));  }  rank_map.reset();#endif        if (!to_map_2)    BGL_FORALL_VERTICES_T(v, g, Graph) put(rank_map, v, get(rank_map2, v));}template<typename Graph, typename RankMap, typename Done, typename RankMap2>voidpage_rank(const Graph& g, RankMap rank_map, Done done,           typename property_traits<RankMap>::value_type damping,          typename graph_traits<Graph>::vertices_size_type n,          RankMap2 rank_map2          BOOST_GRAPH_ENABLE_IF_MODELS_PARM(Graph, distributed_graph_tag)) {  (page_rank_impl)(g, rank_map, done, damping, n, rank_map2);}template<typename MutableGraph>voidremove_dangling_links(MutableGraph& g                      BOOST_GRAPH_ENABLE_IF_MODELS_PARM(MutableGraph,                                                         distributed_graph_tag)){  typename graph_traits<MutableGraph>::vertices_size_type old_n;  do {    old_n = num_vertices(g);    typename graph_traits<MutableGraph>::vertex_iterator vi, vi_end;    for (boost::tie(vi, vi_end) = vertices(g); vi != vi_end; /* in loop */) {      typename graph_traits<MutableGraph>::vertex_descriptor v = *vi++;      if (out_degree(v, g) == 0) {        clear_vertex(v, g);        remove_vertex(v, g);      }    }  } while (num_vertices(g) < old_n);}} // end namespace distributedusing distributed::page_rank;using distributed::remove_dangling_links;} } // end namespace boost::graph#endif // BOOST_PARALLEL_GRAPH_PAGE_RANK_HPP
 |