36 MPI_Allgather( sendTo.data(), comm_size, MPI_INT,
92 const std::map<int,MessageType>& messages,
97 MPI_Comm_rank( comm, &rank);
98 MPI_Comm_size( comm, &comm_size);
99 thrust::host_vector<int> sendTo( comm_size, 0 ), recvFrom( comm_size, 0);
100 thrust::host_vector<int> global( comm_size*comm_size);
101 auto flat_vals = detail::flatten_values( messages);
102 for(
auto& send : flat_vals)
104 sendTo[send.first] = send.second.size();
107 MPI_Allgather( sendTo.data(), comm_size, MPI_INT,
108 global.data(), comm_size, MPI_INT,
110 for(
int i=0; i<comm_size; i++)
111 recvFrom[i] = global[i*comm_size+rank];
113 thrust::host_vector<int> accS( comm_size), accR(comm_size);
114 thrust::exclusive_scan( sendTo.begin(), sendTo.end(), accS.begin());
115 thrust::exclusive_scan( recvFrom.begin(), recvFrom.end(), accR.begin());
117 auto send = detail::flatten_map( flat_vals);
119 thrust::host_vector<value_type> recv(
120 thrust::reduce( recvFrom.begin(), recvFrom.end()));
121 void * send_ptr = thrust::raw_pointer_cast( send.data());
122 const int * sendTo_ptr = thrust::raw_pointer_cast( sendTo.data());
123 const int * accS_ptr = thrust::raw_pointer_cast( accS.data());
124 void * recv_ptr = thrust::raw_pointer_cast( recv.data());
125 const int * recvFrom_ptr = thrust::raw_pointer_cast( recvFrom.data());
126 const int * accR_ptr = thrust::raw_pointer_cast( accR.data());
127 MPI_Datatype type = dg::getMPIDataType<value_type>();
128 MPI_Alltoallv( send_ptr, sendTo_ptr, accS_ptr, type,
129 recv_ptr, recvFrom_ptr, accR_ptr, type,
131 return detail::make_map_t<MessageType>( recv, detail::make_size_map( recvFrom) );
149void mpi_gather(
const thrust::host_vector<std::array<int,2>>& gather_map,
150 const ContainerType& gatherFrom, ContainerType& result, MPI_Comm comm)
152 thrust::host_vector<int> bufferIdx;
153 auto gather_m = gIdx2unique_idx( gather_map, bufferIdx);
155 auto gather = detail::flatten_map( send_map);
156 auto size_map = detail::get_size_map( send_map);
157 ContainerType res( gather.size());
158 thrust::gather( gather.begin(), gather.end(), gatherFrom.begin(), res.begin());
159 std::map<int,ContainerType> result_map = detail::make_map_t<ContainerType>( res, size_map);
160 std::map<int,ContainerType> recv =
mpi_permute( result_map, comm);
161 ContainerType flat = detail::flatten_map( recv);
162 result.resize( bufferIdx.size());
163 thrust::gather( bufferIdx.begin(), bufferIdx.end(), flat.begin(), result.begin());
185void mpi_scatter(
const thrust::host_vector<std::array<int,2>>& scatter_map,
186 const ContainerType& toScatter, ContainerType& result,
188 bool resize_result =
false
191 thrust::host_vector<int> bufferIdx;
193 auto scatter_m = gIdx2unique_idx( scatter_map, bufferIdx);
195 auto scatter = detail::flatten_map( recv_map);
197 auto size_map = detail::get_size_map( scatter_m);
199 ContainerType to_send( toScatter);
200 thrust::scatter( toScatter.begin(), toScatter.end(), bufferIdx.begin(), to_send.begin());
202 auto send = detail::make_map_t<ContainerType>( to_send, size_map);
204 auto res = detail::flatten_map( result_map);
206 result.resize( res.size());
207 thrust::scatter( res.begin(), res.end(), scatter.begin(), result.begin());
227 thrust::host_vector<Integer> seq( p.size());
228 thrust::host_vector<std::array<Integer,2>> seq_arr( p.size());
229 thrust::sequence( seq.begin(), seq.end());
231 MPI_Comm_rank( comm, &rank);
233 for(
unsigned u=0; u<seq.size(); u++)
234 seq_arr[u] = {rank, seq[u]};
235 thrust::host_vector<std::array<Integer,2>> sort_map;
void mpi_gather(const thrust::host_vector< std::array< int, 2 > > &gather_map, const ContainerType &gatherFrom, ContainerType &result, MPI_Comm comm)
Un-optimized distributed gather operation.
Definition mpi_permutation.h:149
thrust::host_vector< std::array< Integer, 2 > > mpi_invert_permutation(const thrust::host_vector< std::array< Integer, 2 > > &p, MPI_Comm comm)
Invert a globally bijective index map.
Definition mpi_permutation.h:224
void mpi_scatter(const thrust::host_vector< std::array< int, 2 > > &scatter_map, const ContainerType &toScatter, ContainerType &result, MPI_Comm comm, bool resize_result=false)
Un-optimized distributed scatter operation.
Definition mpi_permutation.h:185
std::map< int, MessageType > mpi_permute(const std::map< int, MessageType > &messages, MPI_Comm comm)
Exchange messages between processes in a communicator.
Definition mpi_permutation.h:91