Discontinuous Galerkin Library
#include "dg/algorithm.h"
Loading...
Searching...
No Matches
mpi_init.h
Go to the documentation of this file.
1#pragma once
2
3#include <iostream>
4#include <cassert>
5#include <thrust/host_vector.h>
6#include <thrust/device_vector.h> //declare THRUST_DEVICE_SYSTEM
7#include "config.h"
8#include "mpi_datatype.h"
9#include "../enums.h"
10
17namespace dg
18{
19
22
47inline void mpi_init( int argc, char* argv[])
48{
49#ifdef _OPENMP
50 int provided, error;
51 error = MPI_Init_thread(&argc, &argv, MPI_THREAD_FUNNELED, &provided);
52 assert( error == MPI_SUCCESS && "Threaded MPI lib required!\n");
53#else
54 MPI_Init(&argc, &argv);
55#endif
56#if THRUST_DEVICE_SYSTEM==THRUST_DEVICE_SYSTEM_CUDA
57 int rank;
58 MPI_Comm_rank( MPI_COMM_WORLD, &rank);
59 int num_devices=0;
60 cudaGetDeviceCount(&num_devices);
61 if(num_devices == 0)
62 {
63 std::cerr << "# No CUDA capable devices found on rank "<<rank<<std::endl;
64 MPI_Abort(MPI_COMM_WORLD, -1);
65 exit(-1);
66 }
67 int device = rank % num_devices; //assume # of gpus/node is fixed
68 cudaSetDevice( device);
69#endif//THRUST_DEVICE_SYSTEM==THRUST_DEVICE_SYSTEM_CUDA
70}
71
80template<class T>
81std::vector<T> mpi_read_as( unsigned num, MPI_Comm comm, std::istream& is = std::cin)
82{
83 int rank;
84 MPI_Comm_rank( comm, &rank);
85 std::vector<T> nums(num);
86 if( rank == 0)
87 {
88 for( unsigned u=0; u<num; u++)
89 is >> nums[u];
90 }
91 MPI_Bcast( &nums[0], num, getMPIDataType<T>(), 0, comm);
92 return nums;
93}
94
112inline void mpi_read_grid( unsigned& n, std::vector<unsigned>& N, MPI_Comm comm,
113 std::istream& is = std::cin, bool verbose = true, std::ostream& os = std::cout)
114{
115 int rank;
116 MPI_Comm_rank( comm, &rank);
117 const std::string Ns[6] = {"Nx", "Ny", "Nz", "Nu", "Nv", "Nw"};
118 unsigned ndims = N.size();
119 assert( ndims > 0);
120 if(rank == 0 and verbose)
121 {
122 os << "# Type n, "<<Ns[0];
123 for( unsigned u=1; u<ndims; u++)
124 os << " and "<< Ns[u];
125 os << std::endl;
126 }
127 auto vals = mpi_read_as<unsigned>( 1 + ndims, comm, is);
128 n = vals[0];
129 for( unsigned u=0; u<ndims; u++)
130 N[u] = vals[1+u];
131 if(rank == 0 and verbose)
132 {
133 os << "# On the grid "<<n;
134 for( unsigned u=0; u<ndims; u++)
135 os << " x "<< N[u];
136 os << std::endl;
137 }
138}
139
145inline void mpi_read_grid( unsigned& n, std::vector<unsigned*> N, MPI_Comm comm,
146 std::istream& is = std::cin, bool verbose = true, std::ostream& os = std::cout)
147{
148 std::vector<unsigned> Ns( N.size());
149 mpi_read_grid( n, Ns, comm, is, verbose, os);
150 for( unsigned u=0; u<Ns.size(); u++)
151 *N[u] = Ns[u];
152}
153
177inline MPI_Comm mpi_cart_create( MPI_Comm comm_old, std::vector<int> dims,
178 std::vector<int> periods, bool reorder = true)
179{
180 int size;
181 MPI_Comm_size( comm_old, &size);
182 assert( dims.size() == periods.size());
183 int ndims = dims.size();
184 int re = (int)reorder;
185 int err = MPI_Dims_create( size, ndims, &dims[0]);
186 if( err != MPI_SUCCESS)
187 throw Error(Message(_ping_)<<
188 "Cannot create Cartesian dimensions from given dimensions and size "<<size);
189 int reduce = 1;
190 for( int u=0; u<(int)ndims; u++)
191 reduce *= dims[u];
192 if( size != reduce)
193 {
194 throw Error(Message(_ping_)<<
195 "ERROR: Process partition needs to match total number of processes! "
196 <<size<< " vs "<<reduce);
197 }
198 MPI_Comm comm_cart;
199 err = MPI_Cart_create( comm_old, ndims, &dims[0], &periods[0], re, &comm_cart);
200 if( err != MPI_SUCCESS)
201 throw Error(Message(_ping_)<<
202 "Cannot create Cartesian comm from given communicator");
203 return comm_cart;
204}
205
230inline MPI_Comm mpi_cart_create(
231 std::vector<dg::bc> bcs,
232 std::istream& is = std::cin,
233 MPI_Comm comm_old = MPI_COMM_WORLD,
234 bool reorder = true,
235 bool verbose = true,
236 std::ostream& os = std::cout)
237{
238 int rank, size;
239 MPI_Comm_rank( comm_old, &rank);
240 MPI_Comm_size( comm_old, &size);
241 if(rank==0 && verbose)os << "# MPI v"<<MPI_VERSION<<"."<<MPI_SUBVERSION<<std::endl;
242 unsigned ndims = bcs.size();
243 assert( ndims != 0);
244 std::vector<int> periods( ndims);
245 for( unsigned u=0; u<ndims; u++)
246 {
247 if(bcs[u] == dg::PER)
248 periods[u] = true;
249 else
250 periods[u] = false;
251 }
252 if( rank == 0)
253 {
254 if(verbose)
255 {
256 const std::string nps[6] = {"npx", "npy", "npz", "npu", "npv", "npw"};
257 os << "# Type "<<nps[0];
258 for( unsigned u=1; u<ndims; u++)
259 os << " and "<< nps[u];
260 os << std::endl;
261 }
262 }
263 auto np = mpi_read_as<int>( ndims, comm_old, is);
264 if( rank == 0)
265 {
266 if(verbose)
267 {
268 int num_threads = 1;
269#ifdef _OPENMP
270 num_threads = omp_get_max_threads( );
271#endif //omp
272 os << "# Computing with "<<np[0];
273 for( unsigned u=1; u<ndims; u++)
274 os << " x " <<np[u];
275 os << " processes x " << num_threads<<" threads = "
276 << size*num_threads<<" total"<<std::endl;
277 }
278 }
279#if THRUST_DEVICE_SYSTEM==THRUST_DEVICE_SYSTEM_CUDA
280 int device=0;
281 cudaGetDevice( &device);
282 if( rank==0 and verbose)
283 {
284 std::cout << "# MPI is "
285 <<(cuda_aware_mpi ? "cuda-aware" : "NOT cuda-aware")
286 <<"!\n";
287 }
288 if(verbose)std::cout << "# Rank "<<rank<<" computes with device "<<device<<" !"<<std::endl;
289#endif//THRUST_DEVICE_SYSTEM==THRUST_DEVICE_SYSTEM_CUDA
290 return dg::mpi_cart_create( comm_old, np, periods, reorder);
291}
292
293
302inline void mpi_init1d( dg::bc bcx, MPI_Comm& comm, std::istream& is =
303std::cin, bool verbose = true)
304{
305 comm = mpi_cart_create( {bcx}, is, MPI_COMM_WORLD, true, verbose,
306 std::cout);
307}
316inline void mpi_init2d( dg::bc bcx, dg::bc bcy, MPI_Comm& comm, std::istream&
317is = std::cin, bool verbose = true)
318{
319 comm = mpi_cart_create( {bcx, bcy}, is, MPI_COMM_WORLD, true, verbose,
320 std::cout);
321}
322
331inline void mpi_init3d( dg::bc bcx, dg::bc bcy, dg::bc bcz, MPI_Comm& comm,
332std::istream& is = std::cin, bool verbose = true )
333{
334 comm = mpi_cart_create( {bcx, bcy, bcz}, is, MPI_COMM_WORLD, true, verbose,
335 std::cout);
336}
337
349inline void mpi_init1d( dg::bc bcx, unsigned& n, unsigned& N, MPI_Comm& comm,
350std::istream& is = std::cin, bool verbose = true )
351{
352 comm = mpi_cart_create( {bcx}, is, MPI_COMM_WORLD, true, verbose,
353 std::cout);
354 mpi_read_grid( n, {&N}, comm, is, verbose, std::cout);
355}
356
368inline void mpi_init2d( dg::bc bcx, dg::bc bcy, unsigned& n, unsigned& Nx,
369unsigned& Ny, MPI_Comm& comm, std::istream& is = std::cin, bool verbose = true
370)
371{
372 comm = mpi_cart_create( {bcx, bcy}, is, MPI_COMM_WORLD, true, verbose,
373 std::cout);
374 mpi_read_grid( n, {&Nx, &Ny}, comm, is, verbose, std::cout);
375}
376
377
389inline void mpi_init3d( dg::bc bcx, dg::bc bcy, dg::bc bcz, unsigned& n,
390unsigned& Nx, unsigned& Ny, unsigned& Nz, MPI_Comm& comm, std::istream& is =
391std::cin, bool verbose = true )
392{
393 comm = mpi_cart_create( {bcx, bcy, bcz}, is, MPI_COMM_WORLD, true, verbose,
394 std::cout);
395 mpi_read_grid( n, {&Nx, &Ny, &Nz}, comm, is, verbose, std::cout);
396}
397
399
400} //namespace dg
class intended for the use in throw statements
Definition exceptions.h:83
small class holding a stringstream
Definition exceptions.h:29
enums
#define _ping_
Definition exceptions.h:12
bc
Switch between boundary conditions.
Definition enums.h:15
@ PER
periodic boundaries
Definition enums.h:16
void mpi_init3d(dg::bc bcx, dg::bc bcy, dg::bc bcz, MPI_Comm &comm, std::istream &is=std::cin, bool verbose=true)
DEPRECATED: Short for.
Definition mpi_init.h:331
void mpi_init1d(dg::bc bcx, MPI_Comm &comm, std::istream &is=std::cin, bool verbose=true)
DEPRECATED: Short for.
Definition mpi_init.h:302
void mpi_init2d(dg::bc bcx, dg::bc bcy, MPI_Comm &comm, std::istream &is=std::cin, bool verbose=true)
DEPRECATED: Short for.
Definition mpi_init.h:316
void mpi_init(int argc, char *argv[])
Convencience shortcut: Calls MPI_Init or MPI_Init_thread and inits CUDA devices.
Definition mpi_init.h:47
MPI_Comm mpi_cart_create(MPI_Comm comm_old, std::vector< int > dims, std::vector< int > periods, bool reorder=true)
Convenience call to MPI_Cart_create preceded by MPI_Dims_create.
Definition mpi_init.h:177
void mpi_read_grid(unsigned &n, std::vector< unsigned > &N, MPI_Comm comm, std::istream &is=std::cin, bool verbose=true, std::ostream &os=std::cout)
Read in grid sizes from is.
Definition mpi_init.h:112
std::vector< T > mpi_read_as(unsigned num, MPI_Comm comm, std::istream &is=std::cin)
Read num values from is and broadcast to all processes as type T.
Definition mpi_init.h:81
This is the namespace for all functions and classes defined and used by the discontinuous Galerkin li...