diff options
author | 2019-04-29 10:06:29 -0700 | |
---|---|---|
committer | 2019-04-29 10:06:29 -0700 | |
commit | eb62aba2cf0f94489f0a91ffb1dec776c7cb9b37 (patch) | |
tree | a2f84d363d26b6debe63ac1a40eb79625254b559 /Source/Particles/WarpXParticleContainer.cpp | |
parent | d528e233a7710dc497174c825620f1de7323faa5 (diff) | |
parent | db115f923353e4fe483f13c5de50e9fe9dc701a9 (diff) | |
download | WarpX-eb62aba2cf0f94489f0a91ffb1dec776c7cb9b37.tar.gz WarpX-eb62aba2cf0f94489f0a91ffb1dec776c7cb9b37.tar.zst WarpX-eb62aba2cf0f94489f0a91ffb1dec776c7cb9b37.zip |
Merge branch 'dev' into RZgeometry
Diffstat (limited to 'Source/Particles/WarpXParticleContainer.cpp')
-rw-r--r-- | Source/Particles/WarpXParticleContainer.cpp | 208 |
1 files changed, 77 insertions, 131 deletions
diff --git a/Source/Particles/WarpXParticleContainer.cpp b/Source/Particles/WarpXParticleContainer.cpp index 46f6fc1fc..ecfbd8793 100644 --- a/Source/Particles/WarpXParticleContainer.cpp +++ b/Source/Particles/WarpXParticleContainer.cpp @@ -235,12 +235,12 @@ WarpXParticleContainer::AddNParticles (int lev, void WarpXParticleContainer::DepositCurrent(WarpXParIter& pti, - RealVector& wp, RealVector& uxp, - RealVector& uyp, RealVector& uzp, - MultiFab& jx, MultiFab& jy, MultiFab& jz, - MultiFab* cjx, MultiFab* cjy, MultiFab* cjz, - const long np_current, const long np, - int thread_num, int lev, Real dt ) + RealVector& wp, RealVector& uxp, + RealVector& uyp, RealVector& uzp, + MultiFab& jx, MultiFab& jy, MultiFab& jz, + MultiFab* cjx, MultiFab* cjy, MultiFab* cjz, + const long np_current, const long np, + int thread_num, int lev, Real dt ) { Real *jx_ptr, *jy_ptr, *jz_ptr; const std::array<Real,3>& xyzmin_tile = WarpX::LowerCorner(pti.tilebox(), lev); @@ -263,7 +263,16 @@ WarpXParticleContainer::DepositCurrent(WarpXParIter& pti, // Deposit charge for particles that are not in the current buffers if (np_current > 0) - { + { +#ifdef AMREX_USE_GPU + jx_ptr = jx[pti].dataPtr(); + jy_ptr = jy[pti].dataPtr(); + jz_ptr = jz[pti].dataPtr(); + + auto jxntot = jx[pti].length(); + auto jyntot = jy[pti].length(); + auto jzntot = jz[pti].length(); +#else tbx.grow(ngJ); tby.grow(ngJ); tbz.grow(ngJ); @@ -275,31 +284,15 @@ WarpXParticleContainer::DepositCurrent(WarpXParIter& pti, jx_ptr = local_jx[thread_num].dataPtr(); jy_ptr = local_jy[thread_num].dataPtr(); jz_ptr = local_jz[thread_num].dataPtr(); - - auto jxarr = local_jx[thread_num].array(); - amrex::ParallelFor(tbx, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - jxarr(i,j,k) = 0.0; - }); - - auto jyarr = local_jy[thread_num].array(); - amrex::ParallelFor(tby, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - jyarr(i,j,k) = 0.0; - }); - - auto jzarr = local_jz[thread_num].array(); - amrex::ParallelFor(tbz, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - jzarr(i,j,k) = 0.0; - }); + + local_jx[thread_num].setVal(0.0); + local_jy[thread_num].setVal(0.0); + local_jz[thread_num].setVal(0.0); auto jxntot = local_jx[thread_num].length(); auto jyntot = local_jy[thread_num].length(); auto jzntot = local_jz[thread_num].length(); +#endif BL_PROFILE_VAR_START(blp_pxr_cd); if (j_is_nodal) { @@ -389,43 +382,35 @@ WarpXParticleContainer::DepositCurrent(WarpXParIter& pti, } BL_PROFILE_VAR_STOP(blp_pxr_cd); - + +#ifndef AMREX_USE_GPU BL_PROFILE_VAR_START(blp_accumulate); - - const auto local_jx_arr = local_jx[thread_num].array(); - auto global_jx_arr = jx.array(pti); - amrex::ParallelFor(tbx, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - Gpu::Atomic::Add(&global_jx_arr(i, j, k), local_jx_arr(i, j, k)); - }); - - const auto local_jy_arr = local_jy[thread_num].array(); - auto global_jy_arr = jy.array(pti); - amrex::ParallelFor(tby, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - Gpu::Atomic::Add(&global_jy_arr(i, j, k), local_jy_arr(i, j, k)); - }); - - const auto local_jz_arr = local_jz[thread_num].array(); - auto global_jz_arr = jz.array(pti); - amrex::ParallelFor(tbz, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - Gpu::Atomic::Add(&global_jz_arr(i, j, k), local_jz_arr(i, j, k)); - }); + + jx[pti].atomicAdd(local_jx[thread_num], tbx, tbx, 0, 0, 1); + jy[pti].atomicAdd(local_jy[thread_num], tby, tby, 0, 0, 1); + jz[pti].atomicAdd(local_jz[thread_num], tbz, tbz, 0, 0, 1); BL_PROFILE_VAR_STOP(blp_accumulate); - } +#endif + } // Deposit charge for particles that are in the current buffers if (np_current < np) - { + { const IntVect& ref_ratio = WarpX::RefRatio(lev-1); const Box& ctilebox = amrex::coarsen(pti.tilebox(),ref_ratio); const std::array<Real,3>& cxyzmin_tile = WarpX::LowerCorner(ctilebox, lev-1); +#ifdef AMREX_USE_GPU + jx_ptr = (*cjx)[pti].dataPtr(); + jy_ptr = (*cjy)[pti].dataPtr(); + jz_ptr = (*cjz)[pti].dataPtr(); + + auto jxntot = jx[pti].length(); + auto jyntot = jy[pti].length(); + auto jzntot = jz[pti].length(); +#else + tbx = amrex::convert(ctilebox, WarpX::jx_nodal_flag); tby = amrex::convert(ctilebox, WarpX::jy_nodal_flag); tbz = amrex::convert(ctilebox, WarpX::jz_nodal_flag); @@ -441,31 +426,15 @@ WarpXParticleContainer::DepositCurrent(WarpXParIter& pti, jy_ptr = local_jy[thread_num].dataPtr(); jz_ptr = local_jz[thread_num].dataPtr(); - auto jxarr = local_jx[thread_num].array(); - amrex::ParallelFor(tbx, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - jxarr(i,j,k) = 0.0; - }); - - auto jyarr = local_jy[thread_num].array(); - amrex::ParallelFor(tby, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - jyarr(i,j,k) = 0.0; - }); - - auto jzarr = local_jz[thread_num].array(); - amrex::ParallelFor(tbz, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - jzarr(i,j,k) = 0.0; - }); + local_jx[thread_num].setVal(0.0); + local_jy[thread_num].setVal(0.0); + local_jz[thread_num].setVal(0.0); auto jxntot = local_jx[thread_num].length(); auto jyntot = local_jy[thread_num].length(); auto jzntot = local_jz[thread_num].length(); - +#endif + long ncrse = np - np_current; BL_PROFILE_VAR_START(blp_pxr_cd); if (j_is_nodal) { @@ -557,33 +526,15 @@ WarpXParticleContainer::DepositCurrent(WarpXParIter& pti, BL_PROFILE_VAR_STOP(blp_pxr_cd); +#ifndef AMREX_USE_GPU BL_PROFILE_VAR_START(blp_accumulate); - const auto local_jx_arr = local_jx[thread_num].array(); - auto global_jx_arr = cjx->array(pti); - amrex::ParallelFor(tbx, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - Gpu::Atomic::Add(&global_jx_arr(i, j, k), local_jx_arr(i, j, k)); - }); - - const auto local_jy_arr = local_jy[thread_num].array(); - auto global_jy_arr = cjy->array(pti); - amrex::ParallelFor(tby, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - Gpu::Atomic::Add(&global_jy_arr(i, j, k), local_jy_arr(i, j, k)); - }); - - const auto local_jz_arr = local_jz[thread_num].array(); - auto global_jz_arr = cjz->array(pti); - amrex::ParallelFor(tbz, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - Gpu::Atomic::Add(&global_jz_arr(i, j, k), local_jz_arr(i, j, k)); - }); + (*cjx)[pti].atomicAdd(local_jx[thread_num], tbx, tbx, 0, 0, 1); + (*cjy)[pti].atomicAdd(local_jy[thread_num], tby, tby, 0, 0, 1); + (*cjz)[pti].atomicAdd(local_jz[thread_num], tbz, tbz, 0, 0, 1); BL_PROFILE_VAR_STOP(blp_accumulate); +#endif } }; @@ -610,20 +561,22 @@ WarpXParticleContainer::DepositCharge ( WarpXParIter& pti, RealVector& wp, // Deposit charge for particles that are not in the current buffers if (np_current > 0) - { + { const std::array<Real, 3>& xyzmin = xyzmin_tile; + +#ifdef AMREX_USE_GPU + data_ptr = (*rhomf)[pti].dataPtr(); + auto rholen = (*rhomf)[pti].length(); +#else tile_box.grow(ngRho); local_rho[thread_num].resize(tile_box); - auto rhoarr = local_rho[thread_num].array(); - amrex::ParallelFor(tile_box, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - rhoarr(i,j,k) = 0.0; - }); - data_ptr = local_rho[thread_num].dataPtr(); auto rholen = local_rho[thread_num].length(); + + local_rho[thread_num].setVal(0.0); +#endif + #if (AMREX_SPACEDIM == 3) const long nx = rholen[0]-1-2*ngRho; const long ny = rholen[1]-1-2*ngRho; @@ -652,39 +605,36 @@ WarpXParticleContainer::DepositCharge ( WarpXParIter& pti, RealVector& wp, #endif BL_PROFILE_VAR_STOP(blp_pxr_chd); +#ifndef AMREX_USE_GPU BL_PROFILE_VAR_START(blp_accumulate); - const auto local_rho_arr = local_rho[thread_num].array(); - auto global_rho_arr = rhomf->array(pti); - amrex::ParallelFor(tile_box, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - Gpu::Atomic::Add(&global_rho_arr(i, j, k, icomp), local_rho_arr(i, j, k)); - }); + (*rhomf)[pti].atomicAdd(local_rho[thread_num], tile_box, tile_box, 0, icomp, 1); BL_PROFILE_VAR_STOP(blp_accumulate); - } +#endif + } // Deposit charge for particles that are in the current buffers if (np_current < np) - { + { const IntVect& ref_ratio = WarpX::RefRatio(lev-1); const Box& ctilebox = amrex::coarsen(pti.tilebox(), ref_ratio); const std::array<Real,3>& cxyzmin_tile = WarpX::LowerCorner(ctilebox, lev-1); +#ifdef AMREX_USE_GPU + data_ptr = (*crhomf)[pti].dataPtr(); + auto rholen = (*crhomf)[pti].length(); +#else tile_box = amrex::convert(ctilebox, IntVect::TheUnitVector()); tile_box.grow(ngRho); local_rho[thread_num].resize(tile_box); - auto rhoarr = local_rho[thread_num].array(); - amrex::ParallelFor(tile_box, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - rhoarr(i,j,k) = 0.0; - }); - data_ptr = local_rho[thread_num].dataPtr(); auto rholen = local_rho[thread_num].length(); + + local_rho[thread_num].setVal(0.0); +#endif + #if (AMREX_SPACEDIM == 3) const long nx = rholen[0]-1-2*ngRho; const long ny = rholen[1]-1-2*ngRho; @@ -715,17 +665,13 @@ WarpXParticleContainer::DepositCharge ( WarpXParIter& pti, RealVector& wp, #endif BL_PROFILE_VAR_STOP(blp_pxr_chd); +#ifndef AMREX_USE_GPU BL_PROFILE_VAR_START(blp_accumulate); - const auto local_rho_arr = local_rho[thread_num].array(); - auto global_rho_arr = crhomf->array(pti); - amrex::ParallelFor(tile_box, - [=] AMREX_GPU_DEVICE (int i, int j, int k) noexcept - { - Gpu::Atomic::Add(&global_rho_arr(i, j, k, icomp), local_rho_arr(i, j, k)); - }); - + (*crhomf)[pti].atomicAdd(local_rho[thread_num], tile_box, tile_box, 0, icomp, 1); + BL_PROFILE_VAR_STOP(blp_accumulate); +#endif } }; |