|
| MPIKroneckerGather (MPI_Comm comm=MPI_COMM_NULL) |
| no communication
|
|
| MPIKroneckerGather (unsigned left_size, const std::map< int, thrust::host_vector< int > > &recvIdx, unsigned n, unsigned num_cols, unsigned right_size, MPI_Comm comm_1d) |
| Construct from communication pattern.
|
|
template<template< typename > typename OtherVector> |
| MPIKroneckerGather (const MPIKroneckerGather< OtherVector > &src) |
| Construct from other execution policy.
|
|
MPI_Comm | communicator () const |
| The internal MPI communicator used.
|
|
unsigned | buffer_size () const |
| Number of pointers in receive buffer equals number of indices in recvIdx .
|
|
unsigned | chunk_size () const |
| n*left_size*right_size
|
|
bool | isCommunicating () const |
| True if the gather/scatter operation involves actual MPI communication.
|
|
template<class ContainerType > |
void | global_gather_init (const ContainerType &gatherFrom) const |
| \( w' = P_{G,MPI} G_2 v\). Globally (across processes) asynchronously gather data into a buffer
|
|
template<class ContainerType > |
void | global_gather_wait (const ContainerType &gatherFrom, Vector< const dg::get_value_type< ContainerType > * > &buffer_ptrs) const |
| Wait for asynchronous communication to finish and gather received data into buffer and return pointers to it.
|
|
template<template< typename > typename Vector>
struct dg::MPIKroneckerGather< Vector >
Communicator for asynchronous communication of MPISparseBlockMat
.
This is a version of MPIGather
that is optimised for the Kronecker type communication pattern ("Hyperblocks") present in our EllSparseBlockMat
It avoids data movement to the greatest possible extent and exposes send and receive buffers via pointers.
Imagine a communicator with Cartesian topology and further imagine that the grid topology is also Cartesian (vectors form a box) in Nd dimensions. A Sparseblockmat typically requires to gather slices of given index from other processes in a 1d communicator. This class provides pointers to these other indices. The pointers either reference data in an internal communication buffer (since it involves communciation to get the layers from neighboring processes) another buffer (if mpi communication requires to reorder input data) or the input vector itself (if the communication goes along the last dimension there is no need to reorder, in fact, here is the main gain we get from the pointer approach, we save on unnecessary data copies, which might be significant in cases where the communication to computation ratio is high). The size of the data each pointer references is the halo size, buffer_size()
The communication is done asynchronously i.e. the user can initiate the communication and signal when the results are needed at a later stage.
- Note
- If the number of neighboring processes in the given direction is 1, the buffer size is 0 and all members return immediately.
-
the pointers may alias each other (if the input contains less than 4 layers)
-
the corresponding gather map is of general type and the communication can also be modeled in
MPIGather
- Template Parameters
-
Vector | a thrust Vector e.g. thrust::host_vector or thrust::device_vector Determines the internal buffer type of the \( G_2\) gather operation |
- See also
- dg::RowColDistMat
template<template< typename > typename Vector>
template<class ContainerType >
Wait for asynchronous communication to finish and gather received data into buffer and return pointers to it.
Call MPI_Waitall
on internal MPI_Request
variables and manage host memory in case of cuda-unaware MPI. After this call returns it is safe to use the buffer and the gatherFrom
variable from the corresponding global_gather_init
call
- Parameters
-
gatherFrom | source vector v; data is collected from this vector |
buffer_ptrs | (write only) pointers coresponding to recvIdx from the constructor |