[gpaw-users] vdW calculations
Jens Jørgen Mortensen
jensj at fysik.dtu.dk
Fri Mar 26 13:21:39 CET 2010
On Fri, 2010-03-26 at 13:41 +0200, Janne Blomqvist wrote:
> Jess Wellendorff Pedersen wrote:
> > Dear Jussi,
> >
> > You are right that the vdW-DF functional uses revPBE setups and these
> > must therefore be generated manually.
> >
> > A general status of the FFT implementation of the vdW-DF in GPAW:
> > 1) Forces work, and we expect them to work well.
> > 2) The spatial integral over all electron densities that vdW-interact
> > with each other is handled by an FFT, which is why the vdW-DF takes more
> > memory than usual GGA calculations. This is highly dependent on system
> > size: The extra memory used grows rapidly as one approaches large unit
> > cells, say 20x20x20 Å. Also, periodic boundary conditions takes less
> > memory than non-periodic ones.
> > 3) Apart from the memory issue, the functional is reasonably ready-to-go
> > as is, and as a starting point one can use it without giving any
> > arguments for the functional. I.e. just do calc=GPAW(...,
> > xc='vdW-DF',.....). It works in parallel with up to 20 processors. When
> > in parallel, each processor handles one of the 20 cubic interpolation
> > splines, so if only 16 processors are used, 12 processors must wait for
> > 4 processors to handle the remaining 4 splines.
>
> In my experience, the memory consumption is a big problem in practice.
> On a big machine with only 1 GB per core (louhi, should be familiar to
> Jussi!), I have used N*160 cores. As there are 8 cores per node, with
> 160 cores there is then only 1 vdw job per node, and the vdw job can use
> all the remaining memory on a node. Sometimes, even that is not enough.
>
> On way to slightly reduce memory consumption is to use real->complex
> FFT's instead of complex-complex. Jens-Jörgen posted a patch to trac,
> and I have a half-done effort at this lying around in my home dir, but
> AFAIK nothing has been committed to trunk yet.
My patch seemed to work, but I didn't have time to polish and test it
carefully. If someone could finish that work, that would be very
welcome!
JJ
> I suppose another step would be to domain decompose the arrays, and
> implement a MPI distributed FFT, but that's a lot of work.
>
> Performance-wise, I'm happy with the current FFT implementation.
> Compared to the GGA calculation, the calculation time for the vdw
> correction seems to be quite insignificant, at least for decent sized
> systems (I don't actually have numbers, just my seat-of-the-pants feeling).
>
> > 4) A word of caution, however: The vdW-DF implementation is still not as
> > robust and ready-to-go as GGAs. One recently realized issue concerns the
> > value of cutoff-radius used for density-density interactions (rcut in
> > the code). Default is 125 Bohr, but for some reason this large cut-off
> > sometimes leads to diverging interaction energies when increasing the
> > unit cell size. Most often so for small systems such as molecules, rare
> > gas dimers etc. The reason for this is still unresolved. For metallic
> > adsorption systems, however, rcut=125 Bohr is well suited.
>
> One thing I have noticed is that the effective potential seems a bit
> weird. With a GGA calculation, the xy-averaged potential (surface calc,
> see
> https://svn.fysik.dtu.dk/projects/gpaw/trunk/doc/exercises/surface/work_function.py)
> looks like one would expect, but the same system calculated with vdw the
> potential seems 'spiky'.
>
> >
> > Sincerely,
> > Jess Wellendorff
> > CAMD ph.d student
> >
> >
> > Jussi Enkovaara wrote:
> >> Jussi Enkovaara wrote:
> >>> Hi,
> >>> it seems that vdW functional uses revPBE functional for some parts.
> >>> As revPBE setups are not
> >>> distributed in gpaw-setups-0.5.3574.tar.gz, these have to be
> >>> generated manually, am I right?
> >>>
> >>> Are there some other things one has to consider when peforming
> >>> self-consistent vdW calculations?
> >>
> >> One question already came to my mind, do forces work correctly with vdW?
> >>
> >> Best regards,
> >> Jussi
> >> _______________________________________________
> >> gpaw-users mailing list
> >> gpaw-users at listserv.fysik.dtu.dk
> >> https://listserv.fysik.dtu.dk/mailman/listinfo/gpaw-users
> > _______________________________________________
> > gpaw-users mailing list
> > gpaw-users at listserv.fysik.dtu.dk
> > https://listserv.fysik.dtu.dk/mailman/listinfo/gpaw-users
>
>
> --
> Janne Blomqvist
> _______________________________________________
> gpaw-users mailing list
> gpaw-users at listserv.fysik.dtu.dk
> https://listserv.fysik.dtu.dk/mailman/listinfo/gpaw-users
More information about the gpaw-users
mailing list