Discontinuous Galerkin Library
#include "dg/algorithm.h"
|
Namespaces | |
namespace | dg::x |
Macros | |
#define | DG_DEVICE |
Expands to __host__ __device__ if compiled with nvcc else is empty. | |
Typedefs | |
template<class T > | |
using | dg::HVec_t = thrust::host_vector<T> |
Host Vector. | |
using | dg::HVec = thrust::host_vector<double> |
Host Vector. | |
using | dg::cHVec = thrust::host_vector<thrust::complex<double>> |
complex Host Vector | |
using | dg::iHVec = thrust::host_vector<int> |
integer Host Vector | |
using | dg::fHVec = thrust::host_vector<float> |
Host Vector. | |
using | dg::DVec = thrust::device_vector<double> |
Device Vector. The device can be an OpenMP parallelized cpu or a gpu. This depends on the value of the macro THRUST_DEVICE_SYSTEM, which can be either THRUST_DEVICE_SYSTEM_OMP for openMP or THRUST_DEVICE_SYSTEM_CUDA for a gpu or THRUST_DEVICE_SYSTEM_CPP for a cpu. | |
using | dg::iDVec = thrust::device_vector<int> |
integer Device Vector | |
using | dg::cDVec = thrust::device_vector<thrust::complex<double>> |
complex Device Vector | |
using | dg::fDVec = thrust::device_vector<float> |
Device Vector. The device can be an OpenMP parallelized cpu or a gpu. This depends on the value of the macro THRUST_DEVICE_SYSTEM, which can be either THRUST_DEVICE_SYSTEM_OMP for openMP or THRUST_DEVICE_SYSTEM_CUDA for a gpu or THRUST_DEVICE_SYSTEM_CPP for a cpu. | |
template<class T > | |
using | dg::HMatrix_t = EllSparseBlockMat<T, thrust::host_vector> |
using | dg::HMatrix = EllSparseBlockMat<double, thrust::host_vector> |
Host Matrix for derivatives. | |
using | dg::fHMatrix = EllSparseBlockMat<float, thrust::host_vector> |
Host Matrix for derivatives. | |
using | dg::DMatrix = EllSparseBlockMat<double, thrust::device_vector> |
Device Matrix for derivatives. | |
using | dg::fDMatrix = EllSparseBlockMat<float, thrust::device_vector> |
Device Matrix for derivatives. | |
template<class real_type > | |
using | dg::IHMatrix_t = dg::SparseMatrix<int, real_type, thrust::host_vector> |
template<class real_type > | |
using | dg::IDMatrix_t = dg::SparseMatrix<int, real_type, thrust::device_vector> |
using | dg::IHMatrix = IHMatrix_t<double> |
using | dg::IDMatrix = IDMatrix_t<double> |
template<class T > | |
using | dg::MHVec_t = dg::MPI_Vector<dg::HVec_t<T> > |
MPI Host Vector s.a. dg::HVec_t. | |
using | dg::MHVec = dg::MPI_Vector<dg::HVec > |
MPI Host Vector s.a. dg::HVec. | |
using | dg::cMHVec = dg::MPI_Vector<dg::cHVec > |
MPI Host Vector s.a. dg::cHVec. | |
using | dg::fMHVec = dg::MPI_Vector<dg::fHVec > |
MPI Host Vector s.a. dg::fHVec. | |
using | dg::MDVec = dg::MPI_Vector<dg::DVec > |
MPI Device Vector s.a. dg::DVec. | |
using | dg::cMDVec = dg::MPI_Vector<dg::cDVec > |
MPI Device Vector s.a. dg::cDVec. | |
using | dg::fMDVec = dg::MPI_Vector<dg::fDVec > |
MPI Device Vector s.a. dg::fDVec. | |
template<class T > | |
using | dg::CooMat_t = dg::CooSparseBlockMat<T, thrust::host_vector> |
using | dg::CooMat = dg::CooSparseBlockMat<double, thrust::host_vector> |
using | dg::fCooMat = dg::CooSparseBlockMat<float, thrust::host_vector> |
using | dg::DCooMat = dg::CooSparseBlockMat<double, thrust::device_vector> |
using | dg::fDCooMat = dg::CooSparseBlockMat<float, thrust::device_vector> |
template<class T > | |
using | dg::MHMatrix_t = dg::MPISparseBlockMat<thrust::host_vector, dg::HMatrix_t<T>, dg::CooMat_t<T>> |
MPI Host Matrix for derivatives. | |
using | dg::MHMatrix = dg::MPISparseBlockMat<thrust::host_vector, dg::HMatrix, dg::CooMat> |
MPI Host Matrix for derivatives. | |
using | dg::fMHMatrix = dg::MPISparseBlockMat<thrust::host_vector, dg::fHMatrix, dg::fCooMat> |
MPI Host Matrix for derivatives. | |
using | dg::MDMatrix = dg::MPISparseBlockMat<thrust::device_vector, dg::DMatrix, dg::DCooMat> |
MPI Device Matrix for derivatives. | |
using | dg::fMDMatrix = dg::MPISparseBlockMat<thrust::device_vector, dg::fDMatrix, dg::fDCooMat> |
MPI Device Matrix for derivatives. | |
template<class real_type > | |
using | dg::MIHMatrix_t = MPIDistMat< thrust::host_vector, IHMatrix_t<real_type> > |
template<class real_type > | |
using | dg::MIDMatrix_t = MPIDistMat< thrust::device_vector, IDMatrix_t<real_type> > |
using | dg::MIHMatrix = MIHMatrix_t<double> |
using | dg::MIDMatrix = MIDMatrix_t<double> |
#define DG_DEVICE |
Expands to __host__ __device__ if compiled with nvcc else is empty.
using dg::cDVec = thrust::device_vector<thrust::complex<double>> |
complex Device Vector
using dg::cHVec = thrust::host_vector<thrust::complex<double>> |
complex Host Vector
using dg::cMDVec = dg::MPI_Vector<dg::cDVec > |
MPI Device Vector s.a. dg::cDVec.
using dg::cMHVec = dg::MPI_Vector<dg::cHVec > |
MPI Host Vector s.a. dg::cHVec.
using dg::CooMat = dg::CooSparseBlockMat<double, thrust::host_vector> |
using dg::CooMat_t = dg::CooSparseBlockMat<T, thrust::host_vector> |
using dg::DCooMat = dg::CooSparseBlockMat<double, thrust::device_vector> |
using dg::DMatrix = EllSparseBlockMat<double, thrust::device_vector> |
Device Matrix for derivatives.
using dg::DVec = thrust::device_vector<double> |
Device Vector. The device can be an OpenMP parallelized cpu or a gpu. This depends on the value of the macro THRUST_DEVICE_SYSTEM, which can be either THRUST_DEVICE_SYSTEM_OMP for openMP or THRUST_DEVICE_SYSTEM_CUDA for a gpu or THRUST_DEVICE_SYSTEM_CPP for a cpu.
using dg::fCooMat = dg::CooSparseBlockMat<float, thrust::host_vector> |
using dg::fDCooMat = dg::CooSparseBlockMat<float, thrust::device_vector> |
using dg::fDMatrix = EllSparseBlockMat<float, thrust::device_vector> |
Device Matrix for derivatives.
using dg::fDVec = thrust::device_vector<float> |
Device Vector. The device can be an OpenMP parallelized cpu or a gpu. This depends on the value of the macro THRUST_DEVICE_SYSTEM, which can be either THRUST_DEVICE_SYSTEM_OMP for openMP or THRUST_DEVICE_SYSTEM_CUDA for a gpu or THRUST_DEVICE_SYSTEM_CPP for a cpu.
using dg::fHMatrix = EllSparseBlockMat<float, thrust::host_vector> |
Host Matrix for derivatives.
using dg::fHVec = thrust::host_vector<float> |
Host Vector.
using dg::fMDMatrix = dg::MPISparseBlockMat<thrust::device_vector, dg::fDMatrix, dg::fDCooMat> |
MPI Device Matrix for derivatives.
using dg::fMDVec = dg::MPI_Vector<dg::fDVec > |
MPI Device Vector s.a. dg::fDVec.
using dg::fMHMatrix = dg::MPISparseBlockMat<thrust::host_vector, dg::fHMatrix, dg::fCooMat> |
MPI Host Matrix for derivatives.
using dg::fMHVec = dg::MPI_Vector<dg::fHVec > |
MPI Host Vector s.a. dg::fHVec.
using dg::HMatrix = EllSparseBlockMat<double, thrust::host_vector> |
Host Matrix for derivatives.
using dg::HMatrix_t = EllSparseBlockMat<T, thrust::host_vector> |
using dg::HVec = thrust::host_vector<double> |
Host Vector.
using dg::HVec_t = thrust::host_vector<T> |
Host Vector.
using dg::IDMatrix = IDMatrix_t<double> |
using dg::IDMatrix_t = dg::SparseMatrix<int, real_type, thrust::device_vector> |
using dg::iDVec = thrust::device_vector<int> |
integer Device Vector
using dg::IHMatrix = IHMatrix_t<double> |
using dg::IHMatrix_t = dg::SparseMatrix<int, real_type, thrust::host_vector> |
using dg::iHVec = thrust::host_vector<int> |
integer Host Vector
using dg::MDMatrix = dg::MPISparseBlockMat<thrust::device_vector, dg::DMatrix, dg::DCooMat> |
MPI Device Matrix for derivatives.
using dg::MDVec = dg::MPI_Vector<dg::DVec > |
MPI Device Vector s.a. dg::DVec.
using dg::MHMatrix = dg::MPISparseBlockMat<thrust::host_vector, dg::HMatrix, dg::CooMat> |
MPI Host Matrix for derivatives.
using dg::MHMatrix_t = dg::MPISparseBlockMat<thrust::host_vector, dg::HMatrix_t<T>, dg::CooMat_t<T>> |
MPI Host Matrix for derivatives.
using dg::MHVec = dg::MPI_Vector<dg::HVec > |
MPI Host Vector s.a. dg::HVec.
using dg::MHVec_t = dg::MPI_Vector<dg::HVec_t<T> > |
MPI Host Vector s.a. dg::HVec_t.
using dg::MIDMatrix = MIDMatrix_t<double> |
using dg::MIDMatrix_t = MPIDistMat< thrust::device_vector, IDMatrix_t<real_type> > |
using dg::MIHMatrix = MIHMatrix_t<double> |
using dg::MIHMatrix_t = MPIDistMat< thrust::host_vector, IHMatrix_t<real_type> > |