5#include <thrust/host_vector.h>
6#include <thrust/device_vector.h>
51 error = MPI_Init_thread(&argc, &argv, MPI_THREAD_FUNNELED, &provided);
52 assert( error == MPI_SUCCESS &&
"Threaded MPI lib required!\n");
54 MPI_Init(&argc, &argv);
56#if THRUST_DEVICE_SYSTEM==THRUST_DEVICE_SYSTEM_CUDA
58 MPI_Comm_rank( MPI_COMM_WORLD, &rank);
60 cudaGetDeviceCount(&num_devices);
63 std::cerr <<
"# No CUDA capable devices found on rank "<<rank<<std::endl;
64 MPI_Abort(MPI_COMM_WORLD, -1);
67 int device = rank % num_devices;
68 cudaSetDevice( device);
81std::vector<T>
mpi_read_as(
unsigned num, MPI_Comm comm, std::istream& is = std::cin)
84 MPI_Comm_rank( comm, &rank);
85 std::vector<T> nums(num);
88 for(
unsigned u=0; u<num; u++)
91 MPI_Bcast( &nums[0], num, getMPIDataType<T>(), 0, comm);
112inline void mpi_read_grid(
unsigned& n, std::vector<unsigned>& N, MPI_Comm comm,
113 std::istream& is = std::cin,
bool verbose =
true, std::ostream& os = std::cout)
116 MPI_Comm_rank( comm, &rank);
117 const std::string Ns[6] = {
"Nx",
"Ny",
"Nz",
"Nu",
"Nv",
"Nw"};
118 unsigned ndims = N.size();
120 if(rank == 0 and verbose)
122 os <<
"# Type n, "<<Ns[0];
123 for(
unsigned u=1; u<ndims; u++)
124 os <<
" and "<< Ns[u];
129 for(
unsigned u=0; u<ndims; u++)
131 if(rank == 0 and verbose)
133 os <<
"# On the grid "<<n;
134 for(
unsigned u=0; u<ndims; u++)
145inline void mpi_read_grid(
unsigned& n, std::vector<unsigned*> N, MPI_Comm comm,
146 std::istream& is = std::cin,
bool verbose =
true, std::ostream& os = std::cout)
148 std::vector<unsigned> Ns( N.size());
150 for(
unsigned u=0; u<Ns.size(); u++)
178 std::vector<int> periods,
bool reorder =
true)
181 MPI_Comm_size( comm_old, &size);
182 assert( dims.size() == periods.size());
183 int ndims = dims.size();
184 int re = (int)reorder;
185 int err = MPI_Dims_create( size, ndims, &dims[0]);
186 if( err != MPI_SUCCESS)
188 "Cannot create Cartesian dimensions from given dimensions and size "<<size);
190 for(
int u=0; u<(int)ndims; u++)
195 "ERROR: Process partition needs to match total number of processes! "
196 <<size<<
" vs "<<reduce);
199 err = MPI_Cart_create( comm_old, ndims, &dims[0], &periods[0], re, &comm_cart);
200 if( err != MPI_SUCCESS)
202 "Cannot create Cartesian comm from given communicator");
231 std::vector<dg::bc> bcs,
232 std::istream& is = std::cin,
233 MPI_Comm comm_old = MPI_COMM_WORLD,
236 std::ostream& os = std::cout)
239 MPI_Comm_rank( comm_old, &rank);
240 MPI_Comm_size( comm_old, &size);
241 if(rank==0 && verbose)os <<
"# MPI v"<<MPI_VERSION<<
"."<<MPI_SUBVERSION<<std::endl;
242 unsigned ndims = bcs.size();
244 std::vector<int> periods( ndims);
245 for(
unsigned u=0; u<ndims; u++)
256 const std::string nps[6] = {
"npx",
"npy",
"npz",
"npu",
"npv",
"npw"};
257 os <<
"# Type "<<nps[0];
258 for(
unsigned u=1; u<ndims; u++)
259 os <<
" and "<< nps[u];
270 num_threads = omp_get_max_threads( );
272 os <<
"# Computing with "<<np[0];
273 for(
unsigned u=1; u<ndims; u++)
275 os <<
" processes x " << num_threads<<
" threads = "
276 << size*num_threads<<
" total"<<std::endl;
279#if THRUST_DEVICE_SYSTEM==THRUST_DEVICE_SYSTEM_CUDA
281 cudaGetDevice( &device);
282 if( rank==0 and verbose)
284 std::cout <<
"# MPI is "
285 <<(cuda_aware_mpi ?
"cuda-aware" :
"NOT cuda-aware")
288 if(verbose)std::cout <<
"# Rank "<<rank<<
" computes with device "<<device<<
" !"<<std::endl;
303std::cin,
bool verbose =
true)
317is = std::cin,
bool verbose =
true)
332std::istream& is = std::cin,
bool verbose =
true )
334 comm =
mpi_cart_create( {bcx, bcy, bcz}, is, MPI_COMM_WORLD,
true, verbose,
350std::istream& is = std::cin,
bool verbose =
true )
369unsigned& Ny, MPI_Comm& comm, std::istream& is = std::cin,
bool verbose =
true
390unsigned& Nx,
unsigned& Ny,
unsigned& Nz, MPI_Comm& comm, std::istream& is =
391std::cin,
bool verbose =
true )
393 comm =
mpi_cart_create( {bcx, bcy, bcz}, is, MPI_COMM_WORLD,
true, verbose,
395 mpi_read_grid( n, {&Nx, &Ny, &Nz}, comm, is, verbose, std::cout);
class intended for the use in throw statements
Definition exceptions.h:83
small class holding a stringstream
Definition exceptions.h:29
#define _ping_
Definition exceptions.h:12
bc
Switch between boundary conditions.
Definition enums.h:15
@ PER
periodic boundaries
Definition enums.h:16
void mpi_init3d(dg::bc bcx, dg::bc bcy, dg::bc bcz, MPI_Comm &comm, std::istream &is=std::cin, bool verbose=true)
DEPRECATED: Short for.
Definition mpi_init.h:331
void mpi_init1d(dg::bc bcx, MPI_Comm &comm, std::istream &is=std::cin, bool verbose=true)
DEPRECATED: Short for.
Definition mpi_init.h:302
void mpi_init2d(dg::bc bcx, dg::bc bcy, MPI_Comm &comm, std::istream &is=std::cin, bool verbose=true)
DEPRECATED: Short for.
Definition mpi_init.h:316
void mpi_init(int argc, char *argv[])
Convencience shortcut: Calls MPI_Init or MPI_Init_thread and inits CUDA devices.
Definition mpi_init.h:47
MPI_Comm mpi_cart_create(MPI_Comm comm_old, std::vector< int > dims, std::vector< int > periods, bool reorder=true)
Convenience call to MPI_Cart_create preceded by MPI_Dims_create.
Definition mpi_init.h:177
void mpi_read_grid(unsigned &n, std::vector< unsigned > &N, MPI_Comm comm, std::istream &is=std::cin, bool verbose=true, std::ostream &os=std::cout)
Read in grid sizes from is.
Definition mpi_init.h:112
std::vector< T > mpi_read_as(unsigned num, MPI_Comm comm, std::istream &is=std::cin)
Read num values from is and broadcast to all processes as type T.
Definition mpi_init.h:81
This is the namespace for all functions and classes defined and used by the discontinuous Galerkin li...