Extension: Geometries
#include "dg/geometries/geometries.h"
Loading...
Searching...
No Matches
mpi_fieldaligned.h
Go to the documentation of this file.
1#pragma once
2
3#include "dg/algorithm.h"
4#include "fieldaligned.h"
5#include "dg/backend/timer.h"
6
7namespace dg{
8namespace geo{
9
11
12template <class ProductMPIGeometry, class MIMatrix, class LocalContainer>
13struct Fieldaligned< ProductMPIGeometry, MIMatrix, MPI_Vector<LocalContainer> >
14{
15 Fieldaligned(){}
16 template <class Limiter>
18 const ProductMPIGeometry& grid,
21 Limiter limit = FullLimiter(),
22 double eps = 1e-5,
23 unsigned mx=12, unsigned my=12,
24 double deltaPhi = -1, std::string interpolation_method = "linear-nearest",
25 bool benchmark = true):
26 Fieldaligned( dg::geo::createBHat(vec), grid, bcx, bcy, limit, eps,
27 mx, my, deltaPhi, interpolation_method, benchmark)
28 {
29 }
30 template <class Limiter>
32 const ProductMPIGeometry& grid,
35 Limiter limit = FullLimiter(),
36 double eps = 1e-5,
37 unsigned mx=12, unsigned my=12,
38 double deltaPhi = -1, std::string interpolation_method = "linear-nearest",
39 bool benchmark = true);
40 template<class ...Params>
41 void construct( Params&& ...ps)
42 {
43 //construct and swap
44 *this = Fieldaligned( std::forward<Params>( ps)...);
45 }
46
47 dg::bc bcx()const{
48 return m_bcx;
49 }
50 dg::bc bcy()const{
51 return m_bcy;
52 }
53
54 void set_boundaries( dg::bc bcz, double left, double right)
55 {
56 m_bcz = bcz;
57 dg::blas1::copy( left, m_left);
58 dg::blas1::copy( right, m_right);
59 }
60
61 void set_boundaries( dg::bc bcz, const MPI_Vector<LocalContainer>& left, const MPI_Vector<LocalContainer>& right)
62 {
63 m_bcz = bcz;
64 m_left = left;
65 m_right = right;
66 }
67
68 void set_boundaries( dg::bc bcz, const MPI_Vector<LocalContainer>& global, double scal_left, double scal_right)
69 {
70 dg::split( global, m_f, *m_g);
71 dg::blas1::axpby( scal_left, m_f[0], 0., m_left);
72 dg::blas1::axpby( scal_right, m_f[m_g->local().Nz()-1], 0., m_right);
73 m_bcz = bcz;
74 }
75
76 void operator()(enum whichMatrix which, const MPI_Vector<LocalContainer>& in, MPI_Vector<LocalContainer>& out);
77
78 double deltaPhi() const{return m_deltaPhi;}
79 const MPI_Vector<LocalContainer>& hbm()const {
80 return m_hbm;
81 }
82 const MPI_Vector<LocalContainer>& hbp()const {
83 return m_hbp;
84 }
85 const MPI_Vector<LocalContainer>& sqrtG()const {
86 return m_G;
87 }
88 const MPI_Vector<LocalContainer>& sqrtGm()const {
89 return m_Gm;
90 }
91 const MPI_Vector<LocalContainer>& sqrtGp()const {
92 return m_Gp;
93 }
94 const MPI_Vector<LocalContainer>& bphi()const {
95 return m_bphi;
96 }
97 const MPI_Vector<LocalContainer>& bphiM()const {
98 return m_bphiM;
99 }
100 const MPI_Vector<LocalContainer>& bphiP()const {
101 return m_bphiP;
102 }
103 const MPI_Vector<LocalContainer>& bbm()const {
104 return m_bbm;
105 }
106 const MPI_Vector<LocalContainer>& bbo()const {
107 return m_bbo;
108 }
109 const MPI_Vector<LocalContainer>& bbp()const {
110 return m_bbp;
111 }
112 const ProductMPIGeometry& grid() const{return *m_g;}
113
114 template< class BinaryOp, class UnaryOp>
115 MPI_Vector<LocalContainer> evaluate( BinaryOp f, UnaryOp g, unsigned p0,
116 unsigned rounds) const;
117 std::string method() const{return m_interpolation_method;}
118 private:
119 void ePlus( enum whichMatrix which, const MPI_Vector<LocalContainer>& in, MPI_Vector<LocalContainer>& out);
120 void eMinus(enum whichMatrix which, const MPI_Vector<LocalContainer>& in, MPI_Vector<LocalContainer>& out);
121 void zero(enum whichMatrix which, const MPI_Vector<LocalContainer>& in, MPI_Vector<LocalContainer>& out);
122 MIMatrix m_plus, m_zero, m_minus, m_plusT, m_minusT; //2d interpolation matrices
123 MPI_Vector<LocalContainer> m_hbm, m_hbp; //3d size
124 MPI_Vector<LocalContainer> m_G, m_Gm, m_Gp; //3d size
125 MPI_Vector<LocalContainer> m_bphi, m_bphiM, m_bphiP; //3d size
126 MPI_Vector<LocalContainer> m_bbm, m_bbp, m_bbo; //3d size
127
128 MPI_Vector<LocalContainer> m_left, m_right; //2d size
129 MPI_Vector<LocalContainer> m_limiter; //2d size
130 mutable MPI_Vector<LocalContainer> m_ghostM, m_ghostP; //2d size
131 mutable std::vector<MPI_Vector<dg::View<const LocalContainer>> > m_f;
132 mutable std::vector<MPI_Vector<dg::View<LocalContainer>> > m_temp;
134 dg::bc m_bcx, m_bcy, m_bcz;
135 unsigned m_Nz, m_perp_size, m_mx, m_my;
136 double m_deltaPhi, m_eps;
137 std::string m_interpolation_method;
138 dg::geo::CylindricalVectorLvl1 m_vec; // to reconstruct adjoint
139 unsigned m_coords2, m_sizeZ; //number of processes in z
140 //we need to manually send data through the host for cuda-unaware-mpi
141 mutable thrust::host_vector<double> m_buffer; //2d size
142 dg::detail::MPIContiguousGather m_from_minus, m_from_plus;
143 bool m_have_adjoint = false;
144 void updateAdjoint( )
145 {
146 auto vol = dg::tensor::volume(m_g->metric()), vol2d0(vol);
147 auto vol2d = dg::split( vol, *m_g);
148 dg::assign( vol2d[0], vol2d0);
149 dg::ClonePtr<dg::aMPIGeometry2d> grid_transform( m_g->perp_grid()) ;
150 dg::ClonePtr<dg::aGeometry2d> global_grid_magnetic;
151 std::array<thrust::host_vector<double>,3> yp_trafo, ym_trafo;
152 thrust::host_vector<double> hbp, hbm;
153 thrust::host_vector<bool> in_boxp, in_boxm;
154
155 make_matrices( m_vec, grid_transform, global_grid_magnetic,
156 m_bcx, m_bcy, m_eps, m_mx, m_my, m_deltaPhi,
157 m_interpolation_method,
158 false, true, vol2d0, hbp, hbm,
159 in_boxp, in_boxm,
160 yp_trafo, ym_trafo);
161 }
162
163 void make_matrices(
165 const dg::ClonePtr<dg::aMPIGeometry2d>& grid_transform,
166 dg::ClonePtr<dg::aGeometry2d>& global_grid_magnetic,
167 dg::bc bcx, dg::bc bcy, double eps,
168 unsigned mx, unsigned my,
169 double deltaPhi, std::string interpolation_method,
170 bool benchmark, bool make_adjoint,
171 const MPI_Vector<thrust::host_vector<double>>& vol2d0,
172 thrust::host_vector<double>& hbp,
173 thrust::host_vector<double>& hbm,
174 thrust::host_vector<bool>& in_boxp,
175 thrust::host_vector<bool>& in_boxm,
176 std::array<thrust::host_vector<double>,3>& yp_trafo,
177 std::array<thrust::host_vector<double>,3>& ym_trafo
178 )
179 {
180 int rank;
181 MPI_Comm_rank( m_g->communicator(), &rank);
182 std::string inter_m, project_m, fine_m;
183 detail::parse_method( interpolation_method, inter_m, project_m, fine_m);
184 if( benchmark && rank==0)
185 std::cout << "# Interpolation method: \""<<inter_m
186 << "\" projection method: \""<<project_m
187 <<"\" fine grid \""<<fine_m<<"\"\n";
189 // grid_trafo -> grid_equi -> grid_fine -> grid_equi -> grid_trafo
190 dg::Timer t;
191 if( benchmark) t.tic();
192 // We do not need metric of grid_equidist or or grid_fine
193 // We only need grid_fine_local and grid_equidist_global: multiplying cell numbers on an MPI grid would redistribute points for non-equipartition
194 // So we make them RealGrid
195 dg::RealGrid2d<double> grid_equidist_global( grid_transform->global()) ;
196 dg::RealGrid2d<double> grid_fine_local( grid_transform->local());
197 grid_equidist_global.set( 1, grid_equidist_global.shape(0), grid_equidist_global.shape(1));
198 dg::ClonePtr<dg::aMPIGeometry2d> grid_magnetic = grid_transform;//INTEGRATE HIGH ORDER GRID
199 grid_magnetic->set( grid_transform->n() < 3 ? 4 : 7, grid_magnetic->Nx(), grid_magnetic->Ny());
200 global_grid_magnetic = grid_magnetic->global_geometry();
201 // For project method "const" we round up to the nearest multiple of n
202 if( project_m != "dg" && fine_m == "dg")
203 {
204 unsigned rx = mx % grid_transform->nx(), ry = my % grid_transform->ny();
205 if( 0 != rx || 0 != ry)
206 {
207 if(rank==0)std::cerr << "#Warning: for projection method \"const\" mx and my must be multiples of nx and ny! Rounding up for you ...\n";
208 mx = mx + grid_transform->nx() - rx;
209 my = my + grid_transform->ny() - ry;
210 }
211 }
212 if( fine_m == "equi")
213 grid_fine_local.set( 1, grid_fine_local.shape(0), grid_fine_local.shape(1));
214 grid_fine_local.multiplyCellNumbers((double)mx, (double)my);
215 if( benchmark)
216 {
217 t.toc();
218 if(rank==0) std::cout << "# DS: High order grid gen took: "<<t.diff()<<"\n";
219 t.tic();
220 }
222 std::array<thrust::host_vector<double>,3> yp, ym;
223 detail::integrate_all_fieldlines2d( vec, *global_grid_magnetic,
224 grid_transform->local(), yp_trafo, vol2d0.data(), hbp, in_boxp,
225 deltaPhi, eps);
226 detail::integrate_all_fieldlines2d( vec, *global_grid_magnetic,
227 grid_transform->local(), ym_trafo, vol2d0.data(), hbm, in_boxm,
228 -deltaPhi, eps);
229 dg::HVec Xf = dg::evaluate( dg::cooX2d, grid_fine_local);
230 dg::HVec Yf = dg::evaluate( dg::cooY2d, grid_fine_local);
231 {
233 grid_transform->local(), dg::NEU, dg::NEU, grid_transform->n() < 3 ? "cubic" : "dg");
234 yp.fill(dg::evaluate( dg::zero, grid_fine_local));
235 ym = yp;
236 for( int i=0; i<2; i++)
237 {
238 dg::blas2::symv( interpolate, yp_trafo[i], yp[i]);
239 dg::blas2::symv( interpolate, ym_trafo[i], ym[i]);
240 }
241 } // release memory for interpolate matrix
242 if(benchmark)
243 {
244 t.toc();
245 if(rank==0) std::cout << "# DS: Fieldline integration took: "<<t.diff()<<"\n";
246 t.tic();
247 }
249 { // free memory after use
250 dg::IHMatrix fine, projection, multi, temp;
251 if( project_m == "dg")
252 projection = dg::create::projection( grid_transform->global(), grid_fine_local);
253 else
254 projection = dg::create::projection( grid_equidist_global, grid_fine_local, project_m);
255
256 std::array<dg::HVec*,3> xcomp{ &yp[0], &Xf, &ym[0]};
257 std::array<dg::HVec*,3> ycomp{ &yp[1], &Yf, &ym[1]};
258 std::array<MIMatrix*,3> result{ &m_plus, &m_zero, &m_minus};
259 std::array<MIMatrix*,3> resultT{ &m_plusT, &m_zero, &m_minusT};
260 for( unsigned u=0; u<3; u++)
261 {
262 if( inter_m == "dg")
263 {
264 multi = projection * dg::create::interpolation( *xcomp[u], *ycomp[u],
265 grid_transform->global(), bcx, bcy, "dg");
266 multi = dg::convertGlobal2LocalRows( multi, *grid_transform);
267 }
268 else
269 {
270 multi = projection * dg::create::interpolation( *xcomp[u], *ycomp[u],
271 grid_equidist_global, bcx, bcy, inter_m) *
272 dg::create::backproject( grid_transform->global()); // from dg to equidist
273 multi = dg::convertGlobal2LocalRows( multi, *grid_transform);
274 }
275
276 if( project_m != "dg")
277 {
278 multi = dg::create::inv_backproject( grid_transform->local()) * multi;
279 }
280 dg::MIHMatrix mpi = dg::make_mpi_matrix( multi, *grid_transform); //, tempT;
281 dg::blas2::transfer( mpi, *result[u]);
282 if( make_adjoint and u != 1)
283 {
284 dg::IHMatrix multiT = multi.transpose();
285 // multiT is column distributed
286 // multiT has global rows and local column indices
287 dg::convertLocal2GlobalCols( multiT, *grid_transform);
288 // now multiT has global rows and global column indices
289 auto mat = dg::convertGlobal2LocalRows( multiT, *grid_transform);
290 // now mat is row distributed with global column indices
291 auto mpi_mat = dg::make_mpi_matrix( mat, *grid_transform);
292 dg::blas2::transfer( mpi_mat, *resultT[u]);
293 m_have_adjoint = true;
294 }
295 }
296 }
297 if( benchmark)
298 {
299 t.toc();
300 if(rank==0) std::cout << "# DS: Multiplication PI took: "<<t.diff()<<"\n";
301 }
302 }
303};
305template<class MPIGeometry, class MIMatrix, class LocalContainer>
306template <class Limiter>
307Fieldaligned<MPIGeometry, MIMatrix, MPI_Vector<LocalContainer> >::Fieldaligned(
309 const MPIGeometry& grid,
310 dg::bc bcx, dg::bc bcy, Limiter limit, double eps,
311 unsigned mx, unsigned my,
312 double deltaPhi, std::string interpolation_method, bool benchmark
313 ):
314 m_g(grid), m_bcx(bcx), m_bcy(bcy), m_bcz(grid.bcz()),
315 m_Nz( grid.local().Nz()), m_mx(mx), m_my(my), m_eps(eps),
316 m_interpolation_method(interpolation_method),
317 m_vec(vec)
318{
319 int rank;
320 MPI_Comm_rank( grid.communicator(), &rank);
321 int dims[3], periods[3], coords[3];
322 MPI_Cart_get( m_g->communicator(), 3, dims, periods, coords);
323 m_coords2 = coords[2], m_sizeZ = dims[2];
324
326 if( (grid.bcx() == PER && bcx != PER) || (grid.bcx() != PER && bcx == PER) )
327 throw( dg::Error(dg::Message(_ping_)<<"Fieldaligned: Got conflicting periodicity in x. The grid says "<<bc2str(grid.bcx())<<" while the parameter says "<<bc2str(bcx)));
328 if( (grid.bcy() == PER && bcy != PER) || (grid.bcy() != PER && bcy == PER) )
329 throw( dg::Error(dg::Message(_ping_)<<"Fieldaligned: Got conflicting boundary conditions in y. The grid says "<<bc2str(grid.bcy())<<" while the parameter says "<<bc2str(bcy)));
330 if( deltaPhi <=0) deltaPhi = grid.hz();
331 m_deltaPhi = deltaPhi; // store for evaluate
332
333 auto vol = dg::tensor::volume(grid.metric()), vol2d0(vol);
334 auto vol2d = dg::split( vol, grid);
335 dg::assign( vol2d[0], vol2d0);
336 dg::ClonePtr<dg::aMPIGeometry2d> grid_transform( grid.perp_grid()) ;
337 dg::ClonePtr<dg::aGeometry2d> global_grid_magnetic;
338 std::array<thrust::host_vector<double>,3> yp_trafo, ym_trafo;
339 thrust::host_vector<double> hbp, hbm;
340 thrust::host_vector<bool> in_boxp, in_boxm;
341
342 make_matrices( vec, grid_transform, global_grid_magnetic,
343 bcx, bcy, eps, mx, my, m_deltaPhi, interpolation_method,
344 benchmark, false, vol2d0, hbp, hbm,
345 in_boxp, in_boxm,
346 yp_trafo, ym_trafo);
348 dg::HVec hbphi( yp_trafo[2]), hbphiP(hbphi), hbphiM(hbphi);
349 auto tmp = dg::pullback( vec.z(), *grid_transform);
350 hbphi = tmp.data();
351 //this is a pullback bphi( R(zeta, eta), Z(zeta, eta)):
352 if( dynamic_cast<const dg::CartesianMPIGrid2d*>( grid_transform.get()))
353 {
354 for( unsigned i=0; i<hbphiP.size(); i++)
355 {
356 hbphiP[i] = vec.z()(yp_trafo[0][i], yp_trafo[1][i]);
357 hbphiM[i] = vec.z()(ym_trafo[0][i], ym_trafo[1][i]);
358 }
359 }
360 else
361 {
362 dg::HVec Ihbphi = dg::pullback( vec.z(), *global_grid_magnetic);
363 dg::HVec Lhbphi = dg::forward_transform( Ihbphi, *global_grid_magnetic);
364 for( unsigned i=0; i<yp_trafo[0].size(); i++)
365 {
366 hbphiP[i] = dg::interpolate( dg::lspace, Lhbphi, yp_trafo[0][i],
367 yp_trafo[1][i], *global_grid_magnetic);
368 hbphiM[i] = dg::interpolate( dg::lspace, Lhbphi, ym_trafo[0][i],
369 ym_trafo[1][i], *global_grid_magnetic);
370 }
371 }
372 dg::assign3dfrom2d( dg::MHVec(hbphi, grid_transform->communicator()), m_bphi, grid);
373 dg::assign3dfrom2d( dg::MHVec(hbphiM, grid_transform->communicator()), m_bphiM, grid);
374 dg::assign3dfrom2d( dg::MHVec(hbphiP, grid_transform->communicator()), m_bphiP, grid);
375
376 dg::assign3dfrom2d( dg::MHVec(yp_trafo[2], grid_transform->communicator()), m_Gp, grid);
377 dg::assign3dfrom2d( dg::MHVec(ym_trafo[2], grid_transform->communicator()), m_Gm, grid);
378 MPI_Vector<LocalContainer> weights = dg::create::weights( grid);
379 m_G = vol;
380 dg::blas1::pointwiseDot( m_G, weights, m_G);
381 dg::blas1::pointwiseDot( m_Gp, weights, m_Gp);
382 dg::blas1::pointwiseDot( m_Gm, weights, m_Gm);
383
384 dg::assign( dg::evaluate( dg::zero, grid), m_hbm);
385 m_temp = dg::split( m_hbm, grid); //3d vector
386 m_f = dg::split( (const MPI_Vector<LocalContainer>&)m_hbm, grid);
387 dg::assign3dfrom2d( dg::MHVec(hbp, grid_transform->communicator()), m_hbp, grid);
388 dg::assign3dfrom2d( dg::MHVec(hbm, grid_transform->communicator()), m_hbm, grid);
389 dg::blas1::scal( m_hbm, -1.);
391 thrust::host_vector<double> bbm( in_boxp.size(),0.), bbo(bbm), bbp(bbm);
392 for( unsigned i=0; i<in_boxp.size(); i++)
393 {
394 if( !in_boxp[i] && !in_boxm[i])
395 bbo[i] = 1.;
396 else if( !in_boxp[i] && in_boxm[i])
397 bbp[i] = 1.;
398 else if( in_boxp[i] && !in_boxm[i])
399 bbm[i] = 1.;
400 // else all are 0
401 }
402 dg::assign3dfrom2d( dg::MHVec(bbm, grid_transform->communicator()), m_bbm, grid);
403 dg::assign3dfrom2d( dg::MHVec(bbo, grid_transform->communicator()), m_bbo, grid);
404 dg::assign3dfrom2d( dg::MHVec(bbp, grid_transform->communicator()), m_bbp, grid);
405
406
408 m_perp_size = grid_transform->local().size();
409 dg::assign( dg::pullback(limit, *grid_transform), m_limiter);
410 dg::assign( dg::evaluate(dg::zero, *grid_transform), m_left);
411 m_ghostM = m_ghostP = m_right = m_left;
413 int source, dest;
414 dg::detail::MsgChunk chunk { 0, (int)m_perp_size};
415
416 MPI_Cart_shift( grid.comm(2), 0, +1, &source, &dest);
417 std::map<int, thrust::host_vector<dg::detail::MsgChunk>> recvMsgP =
418 {{ dest, thrust::host_vector<dg::detail::MsgChunk>( 1, chunk)}};
419 m_from_plus = dg::detail::MPIContiguousGather( recvMsgP, grid.comm(2));
420
421 MPI_Cart_shift( grid.comm(2), 0, -1, &source, &dest);
422 std::map<int, thrust::host_vector<dg::detail::MsgChunk>> recvMsgM =
423 {{ dest, thrust::host_vector<dg::detail::MsgChunk>( 1, chunk)}};
424 m_from_minus = dg::detail::MPIContiguousGather( recvMsgM, grid.comm(2));
425}
426
427
428template<class G, class M, class container>
429void Fieldaligned<G, M, MPI_Vector<container> >::operator()(enum
430 whichMatrix which, const MPI_Vector<container>& f,
431 MPI_Vector<container>& fe)
432{
433 if(which == einsPlus || which == einsMinusT) ePlus( which, f, fe);
434 if(which == einsMinus || which == einsPlusT) eMinus( which, f, fe);
435 if( which == einsPlus || which == einsMinusT ) ePlus( which, f, fe);
436 else if(which == einsMinus || which == einsPlusT ) eMinus( which, f, fe);
437 else if(which == zeroMinus || which == zeroPlus ||
438 which == zeroMinusT|| which == zeroPlusT ||
439 which == zeroForw ) zero( which, f, fe);
440}
441template< class G, class M, class container>
442void Fieldaligned<G, M, MPI_Vector<container> >::zero( enum whichMatrix which, const MPI_Vector<container>& f, MPI_Vector<container>& f0)
443{
444 dg::split( f, m_f, *m_g);
445 dg::split( f0, m_temp, *m_g);
446 //1. compute 2d interpolation in every plane and store in m_temp
447 for( unsigned i0=0; i0<m_Nz; i0++)
448 {
449 if(which == zeroPlus)
450 dg::blas2::symv( m_plus, m_f[i0], m_temp[i0]);
451 else if(which == zeroMinus)
452 dg::blas2::symv( m_minus, m_f[i0], m_temp[i0]);
453 else if(which == zeroPlusT)
454 {
455 if( ! m_have_adjoint) updateAdjoint( );
456 dg::blas2::symv( m_plusT, m_f[i0], m_temp[i0]);
457 }
458 else if(which == zeroMinusT)
459 {
460 if( ! m_have_adjoint) updateAdjoint( );
461 dg::blas2::symv( m_minusT, m_f[i0], m_temp[i0]);
462 }
463 else if( which == zeroForw)
464 {
465 if ( m_interpolation_method != "dg" )
466 {
467 dg::blas2::symv( m_zero, m_f[i0], m_temp[i0]);
468 }
469 else
470 dg::blas1::copy( m_f[i0], m_temp[i0]);
471 }
472 }
473}
474
475template<class G, class M, class container>
476void Fieldaligned<G,M, MPI_Vector<container> >::ePlus( enum whichMatrix which, const MPI_Vector<container>& f, MPI_Vector<container>& fpe )
477{
478 dg::split( f, m_f, *m_g);
479 dg::split( fpe, m_temp, *m_g);
480 MPI_Vector<dg::View<container>> send_buf(
481 {m_ghostP.data().data(), m_ghostP.size()}, m_g->get_perp_comm());
482 //1. compute 2d interpolation in every plane and store in m_temp
483 for( unsigned i0=0; i0<m_Nz; i0++)
484 {
485 // If communication necessary we write the symv result into send buffer
486 bool comm_plane = (m_sizeZ != 1 and i0 == m_Nz -1);
487 unsigned ip = (i0==m_Nz-1) ? 0:i0+1;
488 if(which == einsPlus)
489 dg::blas2::symv( m_plus, m_f[ip], comm_plane ? send_buf : m_temp[i0]);
490 else if(which == einsMinusT)
491 {
492 if( ! m_have_adjoint) updateAdjoint( );
493 dg::blas2::symv( m_minusT, m_f[ip], comm_plane ? send_buf : m_temp[i0]);
494 }
495 }
496
497 //2. communicate halo in z
498 if( m_sizeZ != 1)
499 {
500 unsigned i0 = m_Nz-1;
501 m_from_plus.global_gather_init( send_buf.data(), m_temp[i0].data());
502 m_from_plus.global_gather_wait( m_temp[i0].data());
503 }
504
505 //3. apply right boundary conditions in last plane
506 unsigned i0=m_Nz-1;
507 if( m_bcz != dg::PER && m_g->local().z1() == m_g->global().z1())
508 {
509 if( m_bcz == dg::DIR || m_bcz == dg::NEU_DIR)
510 dg::blas1::axpby( 2, m_right, -1., m_f[i0], m_ghostP);
511 if( m_bcz == dg::NEU || m_bcz == dg::DIR_NEU)
512 dg::blas1::axpby( m_deltaPhi, m_right, 1., m_f[i0], m_ghostP);
513 //interlay ghostcells with periodic cells: L*g + (1-L)*fpe
514 dg::blas1::axpby( 1., m_ghostP, -1., m_temp[i0], m_ghostP);
515 dg::blas1::pointwiseDot( 1., m_limiter, m_ghostP, 1., m_temp[i0]);
516 }
517}
518
519template<class G, class M, class container>
520void Fieldaligned<G, M, MPI_Vector<container> >::eMinus( enum
521 whichMatrix which, const MPI_Vector<container>& f, MPI_Vector<container>& fme )
522{
523 int rank;
524 MPI_Comm_rank(m_g->communicator(), &rank);
525 dg::split( f, m_f, *m_g);
526 dg::split( fme, m_temp, *m_g);
527 MPI_Vector<dg::View<container>> send_buf(
528 {m_ghostM.data().data(), m_ghostM.size()}, m_g->get_perp_comm());
529 //1. compute 2d interpolation in every plane and store in m_temp
530 for( unsigned i0=0; i0<m_Nz; i0++)
531 {
532 // If communication necessary we write the symv result into send buffer
533 bool comm_plane = (m_sizeZ != 1 and i0 == 0);
534 unsigned im = (i0==0) ? m_Nz-1:i0-1;
535 if(which == einsPlusT)
536 {
537 if( ! m_have_adjoint) updateAdjoint( );
538 dg::blas2::symv( m_plusT, m_f[im], comm_plane ? send_buf : m_temp[i0]);
539 }
540 else if(which == einsMinus)
541 dg::blas2::symv( m_minus, m_f[im], comm_plane ? send_buf : m_temp[i0]);
542 }
543
544 //2. communicate halo in z
545 if( m_sizeZ != 1)
546 {
547 unsigned i0 = 0;
548 m_from_minus.global_gather_init( send_buf.data(), m_temp[i0].data());
549 m_from_minus.global_gather_wait( m_temp[i0].data());
550 }
551
552 //3. apply left boundary conditions in first plane
553 unsigned i0=0;
554 if( m_bcz != dg::PER && m_g->local().z0() == m_g->global().z0())
555 {
556 if( m_bcz == dg::DIR || m_bcz == dg::DIR_NEU)
557 dg::blas1::axpby( 2., m_left, -1., m_f[i0], m_ghostM);
558 if( m_bcz == dg::NEU || m_bcz == dg::NEU_DIR)
559 dg::blas1::axpby( -m_deltaPhi, m_left, 1., m_f[i0], m_ghostM);
560 //interlay ghostcells with periodic cells: L*g + (1-L)*fme
561 dg::blas1::axpby( 1., m_ghostM, -1., m_temp[i0], m_ghostM);
562 dg::blas1::pointwiseDot( 1., m_limiter, m_ghostM, 1., m_temp[i0]);
563 }
564}
565
566template<class G, class M, class container>
567template< class BinaryOp, class UnaryOp>
568MPI_Vector<container> Fieldaligned<G,M, MPI_Vector<container> >::evaluate( BinaryOp binary, UnaryOp unary, unsigned p0, unsigned rounds) const
569{
570 //idea: simply apply I+/I- enough times on the init2d vector to get the result in each plane
571 //unary function is always such that the p0 plane is at x=0
572 assert( p0 < m_g->global().Nz());
573 const dg::ClonePtr<aMPIGeometry2d> g2d = m_g->perp_grid();
574 MPI_Vector<container> init2d = dg::pullback( binary, *g2d);
575 MPI_Vector<container> zero2d = dg::evaluate( dg::zero, *g2d);
576 unsigned globalNz = m_g->global().Nz();
577
578 MPI_Vector<container> temp(init2d), tempP(init2d), tempM(init2d);
579 MPI_Vector<container> vec3d = dg::evaluate( dg::zero, *m_g);
580 std::vector<MPI_Vector<container> > plus2d(globalNz, zero2d), minus2d(plus2d), result(plus2d);
581 unsigned turns = rounds;
582 if( turns ==0) turns++;
583 //first apply Interpolation many times, scale and store results
584 for( unsigned r=0; r<turns; r++)
585 for( unsigned i0=0; i0<globalNz; i0++)
586 {
587 dg::blas1::copy( init2d, tempP);
588 dg::blas1::copy( init2d, tempM);
589 unsigned rep = r*globalNz + i0;
590 for(unsigned k=0; k<rep; k++)
591 {
593 dg::blas2::symv( m_minus, tempP, temp);
594 temp.swap( tempP);
596 dg::blas2::symv( m_plus, tempM, temp);
597 temp.swap( tempM);
598 }
599 dg::blas1::scal( tempP, unary( (double)rep*m_deltaPhi ) );
600 dg::blas1::scal( tempM, unary( -(double)rep*m_deltaPhi ) );
601 dg::blas1::axpby( 1., tempP, 1., plus2d[i0]);
602 dg::blas1::axpby( 1., tempM, 1., minus2d[i0]);
603 }
604 //now we have the plus and the minus filaments
605 if( rounds == 0) //there is a limiter
606 {
607 for( unsigned i0=0; i0<m_Nz; i0++)
608 {
609 int idx = (int)(i0+m_coords2*m_Nz) - (int)p0;
610 if(idx>=0)
611 result[i0] = plus2d[idx];
612 else
613 result[i0] = minus2d[abs(idx)];
614 thrust::copy( result[i0].data().begin(), result[i0].data().end(),
615 vec3d.data().begin() + i0*m_perp_size);
616 }
617 }
618 else //sum up plus2d and minus2d
619 {
620 for( unsigned i0=0; i0<globalNz; i0++)
621 {
622 unsigned revi0 = (globalNz - i0)%globalNz; //reverted index
623 dg::blas1::axpby( 1., plus2d[i0], 0., result[i0]);
624 dg::blas1::axpby( 1., minus2d[revi0], 1., result[i0]);
625 }
626 dg::blas1::axpby( -1., init2d, 1., result[0]);
627 for(unsigned i0=0; i0<m_Nz; i0++)
628 {
629 int idx = ((int)i0 + m_coords2*m_Nz -(int)p0 + globalNz)%globalNz; //shift index
630 thrust::copy( result[idx].data().begin(), result[idx].data().end(),
631 vec3d.data().begin() + i0*m_perp_size);
632 }
633 }
634 return vec3d;
635}
636
637
639
640
644template<class BinaryOp, class UnaryOp>
646 const aProductMPIGeometry3d& grid,
647 const CylindricalVectorLvl0& vec,
648 const BinaryOp& binary,
649 const UnaryOp& unary,
650 unsigned p0,
651 unsigned rounds,
652 double eps = 1e-5)
653{
654 unsigned Nz = grid.Nz();
655 const dg::ClonePtr<aMPIGeometry2d> g2d = grid.perp_grid();
656 // Construct for field-aligned output
657 dg::MHVec vec3d = dg::evaluate( dg::zero, grid);
658 dg::MHVec tempP = dg::evaluate( dg::zero, *g2d), tempM( tempP);
659 std::vector<dg::MHVec> plus2d(Nz, tempP), minus2d(plus2d), result(plus2d);
660 dg::MHVec init2d = dg::pullback( binary, *g2d);
661 std::array<dg::HVec,3> yy0{
662 dg::evaluate( dg::cooX2d, g2d->local()),
663 dg::evaluate( dg::cooY2d, g2d->local()),
664 dg::evaluate( dg::zero, g2d->local())}, yy1(yy0), xx0( yy0), xx1(yy0); //s
665 dg::geo::detail::DSFieldCylindrical3 cyl_field(vec);
666 double deltaPhi = grid.hz();
667 double phiM0 = 0., phiP0 = 0.;
668 unsigned turns = rounds;
669 if( turns == 0) turns++;
670 for( unsigned r=0; r<turns; r++)
671 for( unsigned i0=0; i0<Nz; i0++)
672 {
673 unsigned rep = r*Nz + i0;
674 if( rep == 0)
675 tempM = tempP = init2d;
676 else
677 {
679 "Dormand-Prince-7-4-5", std::array<double,3>{0,0,0});
681 cyl_field, dg::pid_control, dg::fast_l2norm, eps, 1e-10);
682 for( unsigned i=0; i<g2d->local().size(); i++)
683 {
684 // minus direction needs positive integration!
685 double phiM1 = phiM0 + deltaPhi;
686 std::array<double,3>
687 coords0{yy0[0][i],yy0[1][i],yy0[2][i]}, coords1;
688 odeint.integrate_in_domain( phiM0, coords0, phiM1,
689 coords1, deltaPhi, g2d->global(), eps);
690 yy1[0][i] = coords1[0], yy1[1][i] = coords1[1], yy1[2][i] =
691 coords1[2];
692 tempM.data()[i] = binary( yy1[0][i], yy1[1][i]);
693
694 // plus direction needs negative integration!
695 double phiP1 = phiP0 - deltaPhi;
696 coords0 = std::array<double,3>{xx0[0][i],xx0[1][i],xx0[2][i]};
697 odeint.integrate_in_domain( phiP0, coords0, phiP1,
698 coords1, -deltaPhi, g2d->global(), eps);
699 xx1[0][i] = coords1[0], xx1[1][i] = coords1[1], xx1[2][i] =
700 coords1[2];
701 tempP.data()[i] = binary( xx1[0][i], xx1[1][i]);
702 }
703 std::swap( yy0, yy1);
704 std::swap( xx0, xx1);
705 phiM0 += deltaPhi;
706 phiP0 -= deltaPhi;
707 }
708 dg::blas1::scal( tempM, unary( -(double)rep*deltaPhi ) );
709 dg::blas1::scal( tempP, unary( (double)rep*deltaPhi ) );
710 dg::blas1::axpby( 1., tempM, 1., minus2d[i0]);
711 dg::blas1::axpby( 1., tempP, 1., plus2d[i0]);
712 }
713 //now we have the plus and the minus filaments
714 int dims[3], periods[3], coords[3];
715 MPI_Cart_get( grid.communicator(), 3, dims, periods, coords);
716 unsigned coords2 = coords[2];
717 if( rounds == 0) //there is a limiter
718 {
719 for( unsigned i0=0; i0<grid.local().Nz(); i0++)
720 {
721 int idx = (int)(i0+coords2*grid.local().Nz()) - (int)p0;
722 if(idx>=0)
723 result[i0] = plus2d[idx];
724 else
725 result[i0] = minus2d[abs(idx)];
726 thrust::copy( result[i0].data().begin(), result[i0].data().end(),
727 vec3d.data().begin() + i0*g2d->local().size());
728 }
729 }
730 else //sum up plus2d and minus2d
731 {
732 for( unsigned i0=0; i0<Nz; i0++)
733 {
734 unsigned revi0 = (Nz - i0)%Nz; //reverted index
735 dg::blas1::axpby( 1., plus2d[i0], 0., result[i0]);
736 dg::blas1::axpby( 1., minus2d[revi0], 1., result[i0]);
737 }
738 dg::blas1::axpby( -1., init2d, 1., result[0]);
739 for( unsigned i0=0; i0<grid.local().Nz(); i0++)
740 {
741 int idx = ((int)i0 +coords2*grid.local().Nz()-(int)p0 + Nz)%Nz;
742 //shift index
743 thrust::copy( result[idx].data().begin(), result[idx].data().end(),
744 vec3d.data().begin() + i0*g2d->local().size());
745 }
746 }
747 return vec3d;
748}
749
750}//namespace geo
751}//namespace dg
DG_DEVICE T zero(T x, Ts ...xs)
DG_DEVICE double cooY2d(double x, double y)
DG_DEVICE double cooX2d(double x, double y)
void copy(const ContainerTypeIn &source, ContainerTypeOut &target)
void axpby(value_type alpha, const ContainerType1 &x, value_type1 beta, ContainerType &y)
void pointwiseDot(value_type alpha, const ContainerType1 &x1, const ContainerType2 &x2, value_type1 beta, ContainerType &y)
void assign(const from_ContainerType &from, ContainerType &to, Params &&... ps)
void scal(ContainerType &x, value_type alpha)
void transfer(const MatrixType &x, AnotherMatrixType &y)
void symv(MatrixType &&M, const ContainerType1 &x, ContainerType2 &y)
std::string bc2str(bc bcx)
auto weights(const Topology &g)
auto evaluate(Functor &&f, const Topology &g)
whichMatrix
Enum for the use in Fieldaligned.
Definition fieldaligned.h:16
thrust::host_vector< double > fieldaligned_evaluate(const aProductGeometry3d &grid, const CylindricalVectorLvl0 &vec, const BinaryOp &binary, const UnaryOp &unary, unsigned p0, unsigned rounds, double eps=1e-5)
Evaluate a 2d functor and transform to all planes along the fieldlines
Definition fieldaligned.h:1014
ONE FullLimiter
Full Limiter means there is a limiter everywhere.
Definition fieldaligned.h:30
Topology::host_vector forward_transform(const typename Topology::host_vector &in, const Topology &g)
real_type interpolate(dg::space sp, const host_vector &v, real_type x, const RealGrid1d< real_type > &g, dg::bc bcx=dg::NEU)
dg::MIHMatrix_t< typename MPITopology::value_type > projection(const MPITopology &g_new, const MPITopology &g_old, std::string method="dg")
dg::SparseMatrix< int, real_type, thrust::host_vector > interpolation(const RecursiveHostVector &x, const aRealTopology< real_type, Nd > &g, std::array< dg::bc, Nd > bcx, std::string method="dg")
CylindricalVectorLvl1 createBHat(const TokamakMagneticField &mag)
Contravariant components of the magnetic unit vector field and its Divergence and derivative in cylin...
Definition magnetic_field.h:934
void convertLocal2GlobalCols(dg::IHMatrix_t< real_type > &local, const ConversionPolicy &policy)
dg::IHMatrix_t< real_type > convertGlobal2LocalRows(const dg::IHMatrix_t< real_type > &global, const ConversionPolicy &row_policy)
dg::MIHMatrix_t< real_type > make_mpi_matrix(const dg::IHMatrix_t< real_type > &global_cols, const ConversionPolicy &col_policy)
Geometry::host_vector pullback(const Functor &f, const Geometry &g)
void assign3dfrom2d(const host_vector &in2d, Container &out, const Topology &grid)
dg::IHMatrix_t< real_type > inv_backproject(const aRealTopology< real_type, Nd > &g)
dg::IHMatrix_t< real_type > backproject(const aRealTopology< real_type, Nd > &g)
void split(SharedContainer &in, std::vector< View< SharedContainer > > &out, const aRealTopology3d< real_type > &grid)
ContainerType volume(const SparseTensor< ContainerType > &t)
static auto pid_control
static auto fast_l2norm
thrust::host_vector< double > HVec
Cloneable * get()
const container & data() const
SparseMatrix transpose() const
double diff() const
void toc()
void tic()
real_type hz() const
MPI_Comm communicator() const
unsigned Nz() const
const RealGrid< real_type, Nd > & local() const
aRealMPIGeometry2d< real_type > * perp_grid() const
Definition fluxfunctions.h:416
This struct bundles a vector field and its divergence.
Definition fluxfunctions.h:444
const CylindricalFunctor & z() const
z-component of the vector
Definition fluxfunctions.h:474
dg::bc bcx() const
Definition fieldaligned.h:322
const container & hbp() const
Distance between the planes .
Definition fieldaligned.h:399
const container & bphi() const
bphi
Definition fieldaligned.h:419
const container & bbp() const
Mask plus, 1 if fieldline intersects wall in plus direction but not in minus direction,...
Definition fieldaligned.h:444
container evaluate(BinaryOp binary, UnaryOp unary, unsigned p0, unsigned rounds) const
Evaluate a 2d functor and transform to all planes along the fieldline
const container & sqrtG() const
Volume form (including weights) .
Definition fieldaligned.h:404
void operator()(enum whichMatrix which, const container &in, container &out)
Apply the interpolation to three-dimensional vectors.
const container & sqrtGm() const
Volume form on minus plane (including weights) .
Definition fieldaligned.h:409
void set_boundaries(dg::bc bcz, double left, double right)
Set boundary conditions in the limiter region.
Definition fieldaligned.h:339
std::string method() const
Return the interpolation_method string given in the constructor.
Definition fieldaligned.h:509
const container & hbm() const
Distance between the planes and the boundary .
Definition fieldaligned.h:394
const container & bphiM() const
bphi on minus plane
Definition fieldaligned.h:424
dg::bc bcy() const
Definition fieldaligned.h:325
double deltaPhi() const
Definition fieldaligned.h:391
Fieldaligned()
do not allocate memory; no member call except construct is valid
Definition fieldaligned.h:274
void construct(Params &&...ps)
Perfect forward parameters to one of the constructors.
Definition fieldaligned.h:316
const container & bbo() const
Mask both, 1 if fieldline intersects wall in plus direction and in minus direction,...
Definition fieldaligned.h:439
const container & bbm() const
Mask minus, 1 if fieldline intersects wall in minus direction but not in plus direction,...
Definition fieldaligned.h:434
const container & sqrtGp() const
Volume form on plus plane (including weights) .
Definition fieldaligned.h:414
const container & bphiP() const
bphi on plus plane
Definition fieldaligned.h:429
const ProductGeometry & grid() const
Grid used for construction.
Definition fieldaligned.h:448
A tokamak field as given by R0, Psi and Ipol plus Meta-data like shape and equilibrium.
Definition magnetic_field.h:165