[gpaw-users] general comment on memory leaks.
abhishek khetan
askhetan at gmail.com
Mon Jan 18 13:01:03 CET 2016
Dear Marcin, and Ask,
I am indeed on this cluster. And I have already used both these tools. When
I use the r_memusage (to check the peak physical memory), the peak physical
memory is in the order of a few MBs and the process gets killed right as
the beginning with the output only as:
| | |_ | | | |
| | | | | . | | | |
|__ | _|___|_____| 0.12.0.13279
|___|_|
The same is not the case when I take a pre-converged systems and run the
r_memusage script. It shows me a good 2.5 GBs (and rising) before I kill
the process, as I can see its running fine. This is what I mean by saying
that the allocation doesn't even start for these unconverged cases. Using
eigensolver=RMM_DIIS(keep_htpsit=False), has the exact same problems. Is
there a way, I can trick gpaw into giving the cluster much less of a
requirement. I want to try this because, as I have mentioned, at the peak
condition my jobs don't need more than 2 GB per core and I'm providing it 8
GB usually (albiet, to no use).
Best,
On Sat, Jan 16, 2016 at 1:10 PM, Marcin Dulak <mdul at dtu.dk> wrote:
> Hi,
>
> are you one this cluster?
> https://doc.itc.rwth-aachen.de/display/CC/r_memusage
>
> https://doc.itc.rwth-aachen.de/display/CC/Resource+limitations+on+dialog+systems
> It may be that the batch system (LSF) kills your jobs that exceed given
> resident memory.
> The two links above may help you to diagnose that.
> I recall GPAW's memory estimate is not very accurate for standard
> ground-state, PW or GRID mode jobs
> (~20%) and may be very inaccurate (order of magnitude) for VDW or LCAO
> jobs (Ask correct me if this is not the case anymore).
>
> Best regards,
>
> Marcin
> _______________________________________________
> gpaw-users mailing list
> gpaw-users at listserv.fysik.dtu.dk
> https://listserv.fysik.dtu.dk/mailman/listinfo/gpaw-users
>
--
|| radhe radhe ||
abhishek
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listserv.fysik.dtu.dk/pipermail/gpaw-users/attachments/20160118/ef5ca9ad/attachment.html>
More information about the gpaw-users
mailing list