[gpaw-users] Fwd: can't using gpaw-python
謝其軒
z955018 at gmail.com
Tue Oct 22 13:35:30 CEST 2013
Dear Marcin,
Although I may realize what the problem is, I still don't know how to solve
it.
Could you demonstrate step by step?
BR,
chi-hsuan
2013/10/19 Marcin Dulak <Marcin.Dulak at fysik.dtu.dk>
> Hi,
>
>
> On 10/18/2013 11:28 AM, 謝其軒 wrote:
>
> Dear everyone,
> I am one step close to compile the parallel version!
> here's my question.
> when I start the test :
> ----
> mpiexec -np 4 gpaw-python `which gpaw-test` 2>&1 | tee testgpaw.log
> ----
> and here's the file : testgpaw.log
>
> you can find my suggestion here:
> http://listserv.fysik.dtu.dk/pipermail/gpaw-users/2013-October/002408.html
>
> Best regards,
>
> Marcin
>
> ----
> --------------------------------------------------------------------------
> An MPI process has executed an operation involving a call to the
> "fork()" system call to create a child process. Open MPI is currently
> operating in a condition that could result in memory corruption or
> other system errors; your MPI job may hang, crash, or produce silent
> data corruption. The use of fork() (or system() or other calls that
> create child processes) is strongly discouraged.
>
> The process that invoked fork was:
>
> Local host: node04 (PID 26324)
> MPI_COMM_WORLD rank: 0
>
> If you are *absolutely sure* that your application will successfully
> and correctly survive a call to fork(), you may disable this warning
> by setting the mpi_warn_on_fork MCA parameter to 0.
> --------------------------------------------------------------------------
> python 2.7.5 GCC 4.1.2 20080704 (Red Hat 4.1.2-52) 64bit ELF on Linux
> x86_64 redhat 5.7 Final
> Running tests in /tmp/gpaw-test-4xzJUA
> Jobs: 1, Cores: 4, debug-mode: False
>
> =============================================================================
> gemm_complex.py 0.011 OK
> mpicomm.py 0.011 OK
> ase3k_version.py 0.009 OK
> numpy_core_multiarray_dot.py 0.008 OK
> eigh.py 0.011 OK
> lapack.py 0.012 OK
> dot.py 0.012 OK
> lxc_fxc.py 0.011 OK
> blas.py 0.013 OK
> erf.py 0.011 OK
> gp2.py 0.014 OK
> kptpar.py 4.205 OK
> non_periodic.py 0.040 OK
> parallel/blacsdist.py 0.014 OK
> gradient.py 0.017 OK
> cg2.py 0.022 OK
> kpt.py 0.028 OK
> lf.py 0.032 OK
> gd.py 0.013 OK
> parallel/compare.py 0.046 OK
> pbe_pw91.py 0.012 OK
> fsbt.py 0.014 OK
> derivatives.py 0.018 OK
> Gauss.py 0.024 OK
> second_derivative.py 0.022 OK
> integral4.py 0.038 OK
> parallel/ut_parallel.py [node04:26319] 3 more processes have
> sent help message help-mpi-runtime.txt / mpi_init:warn-fork
> [node04:26319] Set MCA parameter "orte_base_help_aggregate" to 0 to see
> all help / error messages
> 1.080 OK
> transformations.py 0.023 OK
> parallel/parallel_eigh.py 0.010 OK
> spectrum.py 0.080 OK
> xc.py 0.047 OK
> zher.py 0.046 OK
> pbc.py 0.041 OK
> lebedev.py 0.030 OK
> parallel/ut_hsblacs.py 0.065 OK
> occupations.py 0.070 OK
> dump_chi0.py 0.046 OK
> cluster.py 0.128 OK
> pw/interpol.py 0.013 OK
> poisson.py 0.054 OK
> pw/lfc.py gpaw-python: symbol lookup error:
> /opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
> undefined symbol: omp_get_num_procs
> gpaw-python: symbol lookup error:
> /opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
> undefined symbol: omp_get_num_procs
> gpaw-python: symbol lookup error:
> /opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
> undefined symbol: omp_get_num_procs
> gpaw-python: symbol lookup error:
> /opt/intel/composerxe-2011.3.174/mkl/lib/intel64/libmkl_intel_thread.so:
> undefined symbol: omp_get_num_procs
> --------------------------------------------------------------------------
> mpiexec has exited due to process rank 0 with PID 26324 on
> node node04 exiting without calling "finalize". This may
> have caused other processes in the application to be
> terminated by signals sent by mpiexec (as reported here).
> --------------------------------------------------------------------------
>
> what happened?
>
>
>
>
> 2013/10/11 Marcin Dulak <Marcin.Dulak at fysik.dtu.dk>
>
>> On 10/11/2013 12:13 PM, 謝其軒 wrote:
>>
>>
>>
>> ---------- Forwarded message ----------
>> From: 謝其軒 <z955018 at gmail.com>
>> Date: 2013/10/11
>> Subject: Re: [gpaw-users] can't using gpaw-python
>> To: Marcin Dulak <Marcin.Dulak at fysik.dtu.dk>
>>
>>
>> Hi again,
>> Is Mr. Marcin there?
>> I've tried to compile all the package myself and failed.
>> I recover the system to the original one that numpy and scipy and others
>> belongs to Canopy.
>> And of course, the error remain. I doubt that the numpy gpaw-python uses
>> is not the one in Canopy but in the root's directory.
>> But how can I change that for serial version works fine?
>>
>> please attach the customize.py and all the steps you perform.
>> After your last email it looked like we were close to get the working a
>> gpaw-python,
>> you just need to replace the intel threading libraries as described in
>> the link i sent.
>>
>> Best regards,
>>
>> Marcin
>>
>>
>
>
> --
> ***********************************
>
> Marcin Dulak
> Technical University of Denmark
> Department of Physics
> Building 307, Room 229
> DK-2800 Kongens Lyngby
> Denmark
> Tel.: (+45) 4525 3157
> Fax.: (+45) 4593 2399
> email: Marcin.Dulak at fysik.dtu.dk
>
> ***********************************
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listserv.fysik.dtu.dk/pipermail/gpaw-users/attachments/20131022/ee8b41fe/attachment.html>
More information about the gpaw-users
mailing list