[gpaw-users] Fwd: can't using gpaw-python
謝其軒
z955018 at gmail.com
Thu Oct 24 10:59:05 CEST 2013
I've compile gpaw without any reference to mkl library.
I export $PATH and $LD_LIBRARY_PATH myself and I got gpaw well compiled.
Then I gpaw-test it in the node with 8 cores. All tests passed but with
some parallel part acting strange:
===================
[z955018 at node05 test]$ mpiexec -np 4 gpaw-python `which gpaw-test` 2>&1 |
tee testgpaw.log
--------------------------------------------------------------------------
An MPI process has executed an operation involving a call to the
"fork()" system call to create a child process. Open MPI is currently
operating in a condition that could result in memory corruption or
other system errors; your MPI job may hang, crash, or produce silent
data corruption. The use of fork() (or system() or other calls that
create child processes) is strongly discouraged.
The process that invoked fork was:
Local host: node05 (PID 7485)
MPI_COMM_WORLD rank: 3
If you are *absolutely sure* that your application will successfully
and correctly survive a call to fork(), you may disable this warning
by setting the mpi_warn_on_fork MCA parameter to 0.
--------------------------------------------------------------------------
python 2.7.5 GCC 4.1.2 20080704 (Red Hat 4.1.2-52) 64bit ELF on Linux
x86_64 redhat 5.8 Final
Running tests in /tmp/gpaw-test-ujm_vy
Jobs: 1, Cores: 4, debug-mode: False
=============================================================================
gemm_complex.py 0.014 OK
mpicomm.py 0.010 OK
ase3k_version.py 0.007 OK
....................
integral4.py 0.025 OK
parallel/ut_parallel.py [node05:07477] 3 more processes have
sent help message help-mpi-runtime.txt / mpi_init:warn-fork
[node05:07477] Set MCA parameter "orte_base_help_aggregate" to 0 to see all
help / error messages
1.074 OK
transformations.py 0.017 OK
parallel/parallel_eigh.py 0.010 OK
spectrum.py 0.037 OK
xc.py 0.064 OK
......................
parallel/realspace_blacs.py 0.018 OK
AA_exx_enthalpy.py 139.244 OK
cmrtest/cmr_test.py 0.023 SKIPPED
cmrtest/cmr_test3.py 0.012 SKIPPED
cmrtest/cmr_test4.py 0.018 SKIPPED
cmrtest/cmr_append.py 0.015 SKIPPED
cmrtest/Li2_atomize.py 0.015 SKIPPED
=============================================================================
Ran 223 tests out of 230 in 2511.9 seconds
Tests skipped: 7
All tests passed!
=============================================================================
The ground state calculation of parallel version is fine.
But when I run the DF calculations, it seems that something familiar
emerges up:
python2.7: symbol lookup error:
/opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
undefined symbol: omp_get_num_procs
(this is run with serial version.....and it pissed me off)
===============================================================================
--------------------------------------------------------------------------
An MPI process has executed an operation involving a call to the
"fork()" system call to create a child process. Open MPI is currently
operating in a condition that could result in memory corruption or
other system errors; your MPI job may hang, crash, or produce silent
data corruption. The use of fork() (or system() or other calls that
create child processes) is strongly discouraged.
The process that invoked fork was:
Local host: node03 (PID 24987)
MPI_COMM_WORLD rank: 3
If you are *absolutely sure* that your application will successfully
and correctly survive a call to fork(), you may disable this warning
by setting the mpi_warn_on_fork MCA parameter to 0.
--------------------------------------------------------------------------
[node03:24983] 7 more processes have sent help message help-mpi-runtime.txt
/ mpi_init:warn-fork
[node03:24983] Set MCA parameter "orte_base_help_aggregate" to 0 to see all
help / error messages
gpaw-python: symbol lookup error:
/opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
undefined symbol: omp_get_num_procs
gpaw-python: symbol lookup error:
/opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
undefined symbol: omp_get_num_procs
gpaw-python: symbol lookup error:
/opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
undefined symbol: omp_get_num_procs
gpaw-python: symbol lookup error:
/opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
undefined symbol: omp_get_num_procs
gpaw-python: symbol lookup error:
/opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
undefined symbol: omp_get_num_procs
gpaw-python: symbol lookup error:
/opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
undefined symbol: omp_get_num_procs
gpaw-python: symbol lookup error:
/opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
undefined symbol: omp_get_num_procs
gpaw-python: symbol lookup error:
/opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
undefined symbol: omp_get_num_procs
--------------------------------------------------------------------------
mpirun has exited due to process rank 6 with PID 24990 on
node node03 exiting without calling "finalize". This may
have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------
=============================================================================
And this is the parallel version.
I think right now I might try to build with mkl library or I'd like to kill
myself....
BR,
chi-hsuan
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listserv.fysik.dtu.dk/pipermail/gpaw-users/attachments/20131024/1ffde104/attachment-0001.html>
More information about the gpaw-users
mailing list