Welcome to the forums. Please post in English or French.

You are not logged in.

#1 2019-06-04 12:51:01

m_golbs
Member
From: Germany
Registered: 2009-11-09
Posts: 118

[beo-01:9838 :0:9838] Caught signal 11 (Segmentation fault: address no

Hello,

I have by different analysis the same error, STAT_NON_LINE.

ERREUR A L'INTERPRETATION DANS ACCAS - INTERRUPTION
>> JDC.py : DEBUT RAPPORT
CR phase d'initialisation
   !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
   ! <S> Exception utilisateur levee mais pas interceptee.                            !
   ! Les bases sont fermees.                                                          !
   ! Type de l'exception : MatriceSinguliereError                                     !
   !                                                                                  !
   !    Arr?t pour cause de matrice non inversible.                                   !
   !    La base globale est sauvegard?e. Elle contient les pas archiv?s avant         !
   ! l'arr?t.                                                                         !
   !                                                                                  !
   !    Conseils :                                                                    !
   !       - V?rifiez vos conditions aux limites.                                     !
   !       - V?rifiez votre mod?le, la coh?rence des unit?s.                          !
   !       - Si vous faites du contact, il ne faut pas que la structure ne "tienne"   !
   ! que par le contact.                                                              !
   !                                                                                  !
   !       - Parfois, en parall?le, le crit?re de d?tection de singularit? de MUMPS   !
   ! est trop pessimiste ! Il reste n?anmoins souvent                                 !
   !         possible de faire passer le calcul complet en relaxant ce crit?re        !
   ! (augmenter de 1 ou 2 la valeur du mot-cl? NPREC) ou                              !
   !         en le d?branchant (valeur du mot-cl? NPREC=-1) ou en relan?ant le calcul !
   ! sur moins de processeurs.                                                        !
   !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
fin CR phase d'initialisation

>> JDC.py : FIN RAPPORT
End of the Code_Aster execution
[beo-01:9838 :0:9838] Caught signal 11 (Segmentation fault: address not mapped to object at address 0x48)
==== backtrace ====
    0  /usr/lib/x86_64-linux-gnu/libucs.so.0(+0x1a780) [0x14cd65877780]
    1  /usr/lib/x86_64-linux-gnu/libucs.so.0(+0x1a932) [0x14cd65877932]
    2  /usr/lib/x86_64-linux-gnu/libpython2.7.so.1.0(PyErr_Occurred+0xa) [0x14cd784e3eba]
    3  /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/bin/aster(utprin_+0x6f) [0x55d24966f1bf]
    4  /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/bin/aster(utmess_core_+0x469) [0x55d24a9e3789]
    5  /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/bin/aster(utmess_+0x881) [0x55d24a9e30f1]
    6  /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/bin/aster(asmpi_check_+0x787) [0x55d24a57b2f7]
    7  /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/bin/aster(terminate+0x3d) [0x55d249675f2d]
    8  /lib/x86_64-linux-gnu/libc.so.6(+0x43041) [0x14cd77a1b041]
    9  /lib/x86_64-linux-gnu/libc.so.6(+0x4313a) [0x14cd77a1b13a]
   10  /usr/lib/x86_64-linux-gnu/libpython2.7.so.1.0(+0x1bdd7f) [0x14cd784d4d7f]
   11  /usr/lib/x86_64-linux-gnu/libpython2.7.so.1.0(+0x1bde6e) [0x14cd784d4e6e]
   12  /usr/lib/x86_64-linux-gnu/libpython2.7.so.1.0(PyErr_PrintEx+0x175) [0x14cd7845fd15]
   13  /usr/lib/x86_64-linux-gnu/libpython2.7.so.1.0(PyRun_SimpleFileExFlags+0x398) [0x14cd78466618]
   14  /usr/lib/x86_64-linux-gnu/libpython2.7.so.1.0(Py_Main+0xb92) [0x14cd784dad32]
   15  /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xe7) [0x14cd779f9b97]
   16  /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/bin/aster(_start+0x2a) [0x55d24966b68a]
===================
/data/home/userfe/solve/edb_99A/9/global/mpi_script.sh: line 47:  9838 Segmentation fault      (core dumped) /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/bin/aster /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/lib/aster/Execution/E_SUPERV.py -commandes fort.1 -max_base 500000 --num_job=9768 --mode=interactif --rep_outils=/usr/lib/opse/apps/astk/2018/outils --rep_mat=/usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/share/aster/materiau --rep_dex=/usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/share/aster/datg --numthreads=2 --suivi_batch --memjeveux=31738.28125 --tpmax=12000000.0
EXECUTION_CODE_ASTER_EXIT_9768=139
PROC=0 INFO_CPU=  11689.86   11481.27     963.66 11987410.14
Content after execution of /tmp/slurm-userfe-289/proc.0 :
.:
total 15757895
drwx------ 3 userfe ad-domain          22 Jun  3 18:54 .
drwxr-xr-x 5 userfe root                5 Jun  3 15:39 ..
-rw-r--r-- 1 userfe ad-domain        2209 Jun  3 15:39 9768.export
drwxr-xr-x 2 userfe ad-domain           2 Jun  3 15:39 REPE_OUT
-rw-r--r-- 1 userfe ad-domain        2354 Jun  3 15:39 config.txt
-rw------- 1 userfe ad-domain   226877440 Jun  3 18:54 core
-rw-r--r-- 1 userfe ad-domain       11963 Jun  3 15:39 fort.1
-rw-r--r-- 1 userfe ad-domain       11963 Jun  3 15:39 fort.1.1
-rw-r--r-- 1 userfe ad-domain           0 Jun  3 15:39 fort.15
-rwxr-xr-x 1 userfe ad-domain     3022064 Jun  3 15:39 fort.19
-rw-r--r-- 1 userfe ad-domain     1233019 Jun  3 15:39 fort.20
-rw-r--r-- 1 userfe ad-domain     2042900 Jun  3 18:54 fort.6
-rw-r--r-- 1 userfe ad-domain           0 Jun  3 15:39 fort.8
-rw-r--r-- 1 userfe ad-domain           0 Jun  3 15:39 fort.9
-rw-r--r-- 1 userfe ad-domain 12884377608 Jun  3 18:54 glob.1
-rw-r--r-- 1 userfe ad-domain 12884377608 Jun  3 18:54 glob.2
-rw-r--r-- 1 userfe ad-domain 12884377608 Jun  3 18:54 glob.3
-rw-r--r-- 1 userfe ad-domain  3167027208 Jun  3 18:54 glob.4
-rw-r--r-- 1 userfe ad-domain          45 Jun  3 18:54 info_cpu
-rwxr-xr-x 1 userfe ad-domain        2599 Jun  3 15:39 mpi_script.sh
-rw-r--r-- 1 userfe ad-domain     4676178 Jun  3 18:54 pick.1
-rw-r--r-- 1 userfe ad-domain   305561608 Jun  3 18:54 vola.1

REPE_OUT:
total 2
drwxr-xr-x 2 userfe ad-domain  2 Jun  3 15:39 .
drwx------ 3 userfe ad-domain 22 Jun  3 18:54 ..

/data/home/userfe/solve/edb_99A/9/global/mpi_script.sh: line 47 >>>

cd $ASRUN_WRKDIR
( . /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/share/aster/profile.sh ; /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/bin/aster /usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/lib/aster/Execution/E_SUPERV.py -commandes fort.1 -max_base 500000 --num_job=9768 --mode=interactif --rep_outils=/usr/lib/opse/apps/astk/2018/outils --rep_mat=/usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/share/aster/materiau --rep_dex=/usr/lib/opse/apps/code_aster/13.6/openmpi-gcc8-4.0/share/aster/datg --numthreads=2 --suivi_batch --memjeveux=31738.28125 --tpmax=12000000.0 ; echo EXECUTION_CODE_ASTER_EXIT_9768=$? ) | tee fort.6
iret=$?

What can be the problem? About information I'm happy.

Thanks and greeting Markus


PS: The *.comm file + export file + mpi run generate however the files under base

Last edited by m_golbs (2019-06-06 13:00:49)


The superfluous chase, is to miss the essential.
Jules Saliège

Offline

Board footer