Page Menu
Home
c4science
Search
Configure Global Search
Log In
Files
F77716462
lal_lj_coul_msm.cpp
No One
Temporary
Actions
Download File
Edit File
Delete File
View Transforms
Subscribe
Mute Notifications
Award Token
Subscribers
None
File Metadata
Details
File Info
Storage
Attached
Created
Fri, Aug 16, 01:04
Size
6 KB
Mime Type
text/x-c++
Expires
Sun, Aug 18, 01:04 (1 d, 23 h)
Engine
blob
Format
Raw Data
Handle
19912258
Attached To
rLAMMPS lammps
lal_lj_coul_msm.cpp
View Options
/***************************************************************************
lj_coul_msm.cpp
-------------------
Trung Dac Nguyen (ORNL)
Class for acceleration of the lj/cut/coul/msm pair style.
__________________________________________________________________________
This file is part of the LAMMPS Accelerator Library (LAMMPS_AL)
__________________________________________________________________________
begin :
email : brownw@ornl.gov
***************************************************************************/
#if defined(USE_OPENCL)
#include "lj_coul_msm_cl.h"
#elif defined(USE_CUDART)
const char *lj_coul_msm=0;
#else
#include "lj_coul_msm_cubin.h"
#endif
#include "lal_lj_coul_msm.h"
#include <cassert>
using namespace LAMMPS_AL;
#define LJCoulMSMT LJCoulMSM<numtyp, acctyp>
extern Device<PRECISION,ACC_PRECISION> device;
template <class numtyp, class acctyp>
LJCoulMSMT::LJCoulMSM() : BaseCharge<numtyp,acctyp>(),
_allocated(false) {
}
template <class numtyp, class acctyp>
LJCoulMSMT::~LJCoulMSM() {
clear();
}
template <class numtyp, class acctyp>
int LJCoulMSMT::bytes_per_atom(const int max_nbors) const {
return this->bytes_per_atom_atomic(max_nbors);
}
template <class numtyp, class acctyp>
int LJCoulMSMT::init(const int ntypes,
double **host_cutsq, double **host_lj1,
double **host_lj2, double **host_lj3,
double **host_lj4, double **host_gcons,
double **host_dgcons, double **host_offset,
double *host_special_lj, const int nlocal,
const int nall, const int max_nbors,
const int maxspecial, const double cell_size,
const double gpu_split, FILE *_screen,
double **host_cut_ljsq, const double host_cut_coulsq,
double *host_special_coul, const int order,
const double qqrd2e) {
int success;
success=this->init_atomic(nlocal,nall,max_nbors,maxspecial,cell_size,gpu_split,
_screen,lj_coul_msm,"k_lj_coul_msm");
if (success!=0)
return success;
// If atom type constants fit in shared memory use fast kernel
int lj_types=ntypes;
shared_types=false;
int max_shared_types=this->device->max_shared_types();
if (lj_types<=max_shared_types && this->_block_size>=max_shared_types) {
lj_types=max_shared_types;
shared_types=true;
}
_lj_types=lj_types;
// Allocate a host write buffer for data initialization
UCL_H_Vec<numtyp> host_write(lj_types*lj_types*32,*(this->ucl_device),
UCL_WRITE_ONLY);
for (int i=0; i<lj_types*lj_types; i++)
host_write[i]=0.0;
lj1.alloc(lj_types*lj_types,*(this->ucl_device),UCL_READ_ONLY);
this->atom->type_pack4(ntypes,lj_types,lj1,host_write,host_lj1,host_lj2,
host_cutsq, host_cut_ljsq);
lj3.alloc(lj_types*lj_types,*(this->ucl_device),UCL_READ_ONLY);
this->atom->type_pack4(ntypes,lj_types,lj3,host_write,host_lj3,host_lj4,
host_offset);
// pack gcons and dgcons
int nrows, ncols;
nrows = 7;
ncols = 7;
UCL_H_Vec<numtyp> dview_gcons(nrows*ncols,*(this->ucl_device),
UCL_WRITE_ONLY);
for (int ix=0; ix<nrows; ix++)
for (int iy=0; iy<ncols; iy++)
dview_gcons[ix*ncols+iy]=host_gcons[ix][iy];
gcons.alloc(nrows*ncols,*(this->ucl_device),UCL_READ_ONLY);
ucl_copy(gcons,dview_gcons,false);
gcons_tex.get_texture(*(this->pair_program),"gcons_tex");
gcons_tex.bind_float(gcons,1);
nrows = 7;
ncols = 6;
UCL_H_Vec<numtyp> dview_dgcons(nrows*ncols,*(this->ucl_device),
UCL_WRITE_ONLY);
for (int ix=0; ix<nrows; ix++)
for (int iy=0; iy<ncols; iy++)
dview_dgcons[ix*ncols+iy]=host_dgcons[ix][iy];
dgcons.alloc(nrows*ncols,*(this->ucl_device),UCL_READ_ONLY);
ucl_copy(dgcons,dview_dgcons,false);
dgcons_tex.get_texture(*(this->pair_program),"dgcons_tex");
dgcons_tex.bind_float(dgcons,1);
sp_lj.alloc(8,*(this->ucl_device),UCL_READ_ONLY);
for (int i=0; i<4; i++) {
host_write[i]=host_special_lj[i];
host_write[i+4]=host_special_coul[i];
}
ucl_copy(sp_lj,host_write,8,false);
_cut_coulsq=host_cut_coulsq;
_qqrd2e=qqrd2e;
_order=order;
_allocated=true;
this->_max_bytes=lj1.row_bytes()+lj3.row_bytes()+
gcons.row_bytes()+dgcons.row_bytes()+sp_lj.row_bytes();
return 0;
}
template <class numtyp, class acctyp>
void LJCoulMSMT::clear() {
if (!_allocated)
return;
_allocated=false;
lj1.clear();
lj3.clear();
gcons.clear();
dgcons.clear();
sp_lj.clear();
this->clear_atomic();
}
template <class numtyp, class acctyp>
double LJCoulMSMT::host_memory_usage() const {
return this->host_memory_usage_atomic()+sizeof(LJCoulMSM<numtyp,acctyp>);
}
// ---------------------------------------------------------------------------
// Calculate energies, forces, and torques
// ---------------------------------------------------------------------------
template <class numtyp, class acctyp>
void LJCoulMSMT::loop(const bool _eflag, const bool _vflag) {
// Compute the block size and grid size to keep all cores busy
const int BX=this->block_size();
int eflag, vflag;
if (_eflag)
eflag=1;
else
eflag=0;
if (_vflag)
vflag=1;
else
vflag=0;
int GX=static_cast<int>(ceil(static_cast<double>(this->ans->inum())/
(BX/this->_threads_per_atom)));
int ainum=this->ans->inum();
int nbor_pitch=this->nbor->nbor_pitch();
this->time_pair.start();
if (shared_types) {
this->k_pair_fast.set_size(GX,BX);
this->k_pair_fast.run(&this->atom->x, &lj1, &lj3, &gcons, &dgcons, &sp_lj,
&this->nbor->dev_nbor, &this->_nbor_data->begin(),
&this->ans->force, &this->ans->engv, &eflag,
&vflag, &ainum, &nbor_pitch, &this->atom->q,
&_cut_coulsq, &_qqrd2e, &_order,
&this->_threads_per_atom);
} else {
this->k_pair.set_size(GX,BX);
this->k_pair.run(&this->atom->x, &lj1, &lj3, &gcons, &dgcons,
&_lj_types, &sp_lj, &this->nbor->dev_nbor,
&this->_nbor_data->begin(), &this->ans->force,
&this->ans->engv, &eflag, &vflag, &ainum,
&nbor_pitch, &this->atom->q, &_cut_coulsq,
&_qqrd2e, &_order, &this->_threads_per_atom);
}
this->time_pair.stop();
}
template class LJCoulMSM<PRECISION,ACC_PRECISION>;
Event Timeline
Log In to Comment