[gpaw-users] Problem running parallel GPAW

Jesper Rude Selknæs jesperrude at gmail.com
Mon Apr 30 16:53:32 CEST 2012


Hi list

I did a bit more investigation. From
https://wiki.fysik.dtu.dk/gpaw/install/installationguide.html#running-tests
 i ran the rank example:

[gpaw]$ mpirun -np 2 gpaw-python -c "import gpaw.mpi as mpi; print mpi.rank"

which ran flawlessly for up to 10 nodes (120  cores total).

I then ran [gpaw]$ mpirun -np 2 gpaw-python gpaw/test/spinpol.py

which fails as soon as I go beyounf the one node. It runs fine for 12
cores within the same node. I know that tehre is no speedup (actaully
tehre is heavy speeddown) using this many cores, but it executes
correctly, which is all i wants for the moment. Part of output ftom
spinpol.py on 2 nodes is given below, as it is quit large.

Any ideas?

R.

Jesper



File "/usr/lib/python2.6/site-packages/ase/atoms.py", line 548, in
get_potential_energy
    self.initialize(atoms)
  File "/usr/lib64/python2.6/site-packages/gpaw/paw.py", line 591, in initialize
    return self._calc.get_potential_energy(self)
  File "/usr/lib64/python2.6/site-packages/gpaw/aseinterface.py", line
37, in get_potential_energy
    self.calculate(atoms, converge=True)
  File "/usr/lib64/python2.6/site-packages/gpaw/paw.py", line 221, in calculate
    self.initialize(atoms)
  File "/usr/lib64/python2.6/site-packages/gpaw/paw.py", line 591, in initialize
    magmom_a, par.hund)
    return self._calc.get_potential_energy(self)
  File "/usr/lib64/python2.6/site-packages/gpaw/aseinterface.py", line
37, in get_potential_energy
    self.calculate(atoms, converge=True)
  File "/usr/lib64/python2.6/site-packages/gpaw/paw.py", line 221, in calculate
  File "/usr/lib64/python2.6/site-packages/gpaw/density.py", line 74,
in initialize
    allocate=False)
  File "/usr/lib64/python2.6/site-packages/gpaw/transformers.py", line
144, in Transformer
    self.initialize(atoms)
  File "/usr/lib64/python2.6/site-packages/gpaw/paw.py", line 591, in initialize
    t = _Transformer(gdin, gdout, nn, dtype, allocate)
  File "/usr/lib64/python2.6/site-packages/gpaw/transformers.py", line
49, in __init__
    gdout.n_c[0] * gdout.n_c[1])
AssertionError
    magmom_a, par.hund)
  File "/usr/lib64/python2.6/site-packages/gpaw/density.py", line 74,
in initialize
    allocate=False)
  File "/usr/lib64/python2.6/site-packages/gpaw/transformers.py", line
144, in Transformer
    t = _Transformer(gdin, gdout, nn, dtype, allocate)
  File "/usr/lib64/python2.6/site-packages/gpaw/transformers.py", line
49, in __init__
    gdout.n_c[0] * gdout.n_c[1])
    magmom_a, par.hund)
  File "/usr/lib64/python2.6/site-packages/gpaw/density.py", line 74,
in initialize
GPAW CLEANUP (node 1): <type 'exceptions.AssertionError'> occurred.
Calling MPI_Abort!
    allocate=False)
  File "/usr/lib64/python2.6/site-packages/gpaw/transformers.py", line
144, in Transformer
AssertionError
    t = _Transformer(gdin, gdout, nn, dtype, allocate)
  File "/usr/lib64/python2.6/site-packages/gpaw/transformers.py", line
49, in __init__
GPAW CLEANUP (node 17): <type 'exceptions.AssertionError'> occurred.
Calling MPI_Abort!
    gdout.n_c[0] * gdout.n_c[1])
AssertionError
GPAW CLEANUP (node 2): <type 'exceptions.AssertionError'> occurred.
Calling MPI_Abort!
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 4 in communicator MPI_COMM_WORLD
with errorcode 42.

NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
mpirun has exited due to process rank 0 with PID 5881 on
node DFT030 exiting improperly. There are two reasons this could occur:

1. this process did not call "init" before exiting, but others in
the job did. This can cause a job to hang indefinitely while it waits
for all processes to call "init". By rule, if one process calls "init",
then ALL processes must call "init" prior to termination.

2. this process called "init", but exited without calling "finalize".
By rule, all processes that call "init" MUST call "finalize" prior to
exiting or it will be considered an "abnormal termination"

This may have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------
[DFT030:05880] 10 more processes have sent help message
help-mpi-api.txt / mpi-abort
[DFT030:05880] Set MCA parameter "orte_base_help_aggregate" to 0 to
see all help / error messages




2012/4/24 Jesper Rude Selknæs <jesperrude at gmail.com>:
> Hi LIst
>
> I have recently installed GPAW, ASE3 and DACAPO on a cluster,
> consisting of 42 nodes of 12 cores each.
>
> I am using Redhat Enterprise 6.0 as my OS, OpenMPI 1.5.3 and TOPQUE.
>
> I have absoutely no problems running parallel GPAW aplications, as
> long as alle the porcesses are on the same node. However, as soon as i
> spread the processes over more than one node, i get into trouble. The
> stderr file from the TORQUE job give me:
>
>
> "
> [DFT046][[35378,1],5][btl_tcp_frag.c:215:mca_btl_tcp_frag_recv]
> mca_btl_tcp_frag_recv: readv failed: Connection reset by peer (104)
> [DFT046][[35378,1],10][btl_tcp_frag.c:215:mca_btl_tcp_frag_recv]
> mca_btl_tcp_frag_recv: readv failed: Connection reset by peer (104)
> [DFT046][[35378,1],4][btl_tcp_frag.c:215:mca_btl_tcp_frag_recv]
> mca_btl_tcp_frag_recv: readv failed: Connection reset by peer (104)
> [DFT046][[35378,1],7][btl_tcp_frag.c:215:mca_btl_tcp_frag_recv]
> mca_btl_tcp_frag_recv: readv failed: Connection reset by peer (104)
> [DFT046][[35378,1],1][btl_tcp_frag.c:215:mca_btl_tcp_frag_recv]
> mca_btl_tcp_frag_recv: readv failed: Connection reset by peer (104)
> [DFT046][[35378,1],6][btl_tcp_frag.c:215:mca_btl_tcp_frag_recv]
> mca_btl_tcp_frag_recv: readv failed: Connection reset by peer (104)
> =>> PBS: job killed: node 1 (DFT045) requested job terminate, 'EOF'
> (code 1099) - received SISTER_EOF attempting to communicate with
> sister MOM's
> mpirun: killing job...
>
> "
>
>
>
> And from the mom logs, i simply get a "sister process died" kinf of message.
>
> I have tried to run a simple "MPI HELLO WORLD" test on as much af 10
> nodes, utilizing all cores, and i have no trouble doing that.
>
> Does anybody has expieiences like this?
>
> Regards
>
> Jesper Selknæs



More information about the gpaw-users mailing list