[gpaw-users] general comment on memory leaks.

abhishek khetan askhetan at gmail.com
Mon Jan 18 13:26:16 CET 2016


I tried using the cluster interactively, and it gives me the output as
below. I couldn't make the r_memusage function work but its easily visible
that the memory requirements are quite modest. I do not know why there is
seg fault when I allocate it in the regular cluster for production jobs.

  ___ ___ ___ _ _ _
 |   |   |_  | | | |
 | | | | | . | | | |
 |__ |  _|___|_____|  0.12.0.13279
 |___|_|

User:   ak498084 at linuxbmc0002.rz.RWTH-Aachen.DE
Date:   Mon Jan 18 13:22:24 2016
Arch:   x86_64
Pid:    20443
Python: 2.7.9
gpaw:   /home/ak498084/Utility/GPAW/gpaw_devel/gpaw-0.12/gpaw
_gpaw:
/home/ak498084/Utility/GPAW/gpaw_devel/gpaw-0.12/build/bin.linux-x86_64-2.7/gpaw-python
ase:    /home/ak498084/Utility/GPAW/gpaw_devel/ase/ase (version 3.10.0)
numpy:
/usr/local_rwth/sw/python/2.7.9/x86_64/lib/python2.7/site-packages/numpy
(version 1.9.1)
scipy:
/usr/local_rwth/sw/python/2.7.9/x86_64/lib/python2.7/site-packages/scipy
(version 0.15.1)
units:  Angstrom and eV
cores:  32

Memory estimate
---------------
Process memory now: 75.02 MiB
Calculator  1145.24 MiB
    Density  56.04 MiB
        Arrays  15.91 MiB
        Localized functions  35.58 MiB
        Mixer  4.55 MiB
    Hamiltonian  23.19 MiB
        Arrays  11.82 MiB
        XC  0.00 MiB
        Poisson  8.81 MiB
        vbar  2.56 MiB
    Wavefunctions  1066.01 MiB
        Arrays psit_nG  523.69 MiB
        Eigensolver  2.29 MiB
        Projections  2.06 MiB
        Projectors  4.17 MiB
        Overlap op  533.81 MiB


On Mon, Jan 18, 2016 at 1:01 PM, abhishek khetan <askhetan at gmail.com> wrote:

> Dear Marcin, and Ask,
>
> I am indeed on this cluster. And I have already used both these tools.
> When I use the r_memusage (to check the peak physical memory), the peak
> physical memory is in the order of a few MBs and the process gets killed
> right as the beginning with the output only as:
>
>  |   |   |_  | | | |
>  | | | | | . | | | |
>  |__ |  _|___|_____|  0.12.0.13279
>  |___|_|
>
>
> The same is not the case when I take a pre-converged systems and run the
> r_memusage script. It shows me a good 2.5 GBs (and rising) before I kill
> the process, as I can see its running fine. This is what I mean by saying
> that the allocation doesn't even start for these unconverged cases. Using
> eigensolver=RMM_DIIS(keep_htpsit=False), has the exact same problems. Is
> there a way, I can trick gpaw into giving the cluster much less of a
> requirement. I want to try this because, as I have mentioned, at the peak
> condition my jobs don't need more than 2 GB per core and I'm providing it 8
> GB usually (albiet, to no use).
>
> Best,
>
>
> On Sat, Jan 16, 2016 at 1:10 PM, Marcin Dulak <mdul at dtu.dk> wrote:
>
>> Hi,
>>
>> are you one this cluster?
>> https://doc.itc.rwth-aachen.de/display/CC/r_memusage
>>
>> https://doc.itc.rwth-aachen.de/display/CC/Resource+limitations+on+dialog+systems
>> It may be that the batch system (LSF) kills your jobs that exceed given
>> resident memory.
>> The two links above may help you to diagnose that.
>> I recall GPAW's memory estimate is not very accurate for standard
>> ground-state, PW or GRID mode jobs
>> (~20%) and may be very inaccurate (order of magnitude) for VDW or LCAO
>> jobs (Ask correct me if this is not the case anymore).
>>
>> Best regards,
>>
>> Marcin
>> _______________________________________________
>> gpaw-users mailing list
>> gpaw-users at listserv.fysik.dtu.dk
>> https://listserv.fysik.dtu.dk/mailman/listinfo/gpaw-users
>>
>
>
>
> --
> || radhe radhe ||
>
> abhishek
>



-- 
|| radhe radhe ||

abhishek
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listserv.fysik.dtu.dk/pipermail/gpaw-users/attachments/20160118/c9d13b82/attachment.html>


More information about the gpaw-users mailing list