Page 1 of 1

2604 Run 0 CL 546 Gen 23 crash on clean restart

Posted: Sat Dec 15, 2007 3:28 am
by dschief
Launch directory: /home/jim/fold
Executable: ./fah6
Arguments: -smp -verbosity 9

[10:45:16] - Ask before connecting: No
[10:45:16] - User name: dschief (Team 13761)
[10:45:16] - User ID: D9053AF4A9A20EE
[10:45:16] - Machine ID: 1
[10:45:16]
[10:45:16] Loaded queue successfully.
[10:45:16]
[10:45:16] + Processing work unit
[10:45:16] Core required: FahCore_a1.exe
[10:45:16] Core found.
[10:45:16] - Autosending finished units...
[10:45:16] Trying to send all finished work units
[10:45:16] + No unsent completed units remaining.
[10:45:16] - Autosend completed
[10:45:16] Working on Unit 08 [December 15 10:45:16]
[10:45:16] + Working ...
[10:45:16] - Calling './mpiexec -np 4 -host 127.0.0.1 ./FahCore_a1.exe -dir work/ -suffix 08 -checkpoint 15 -verbose -lifeline 12389 -version 600'

[10:45:16]
[10:45:16] *------------------------------*
[10:45:16] Folding@Home Gromacs SMP Core
[10:45:16] Version 1.74 (November 27, 2006)
[10:45:16]
[10:45:16] Preparing to commence simulation
[10:45:16] - Ensuring status. Please wait.
[10:45:17]
[10:45:17] Project: 2604 (Run 0, Clone 546, Gen 23)
[10:45:17]
[10:45:17] Assembly optimizations on if available.
[10:45:17] Entering M.D.
[10:45:34] - Expanded 2943729 -> 15158509 (decompressed 514.9 percent)
[10:45:34]
[10:45:34] Project: 2604 (Run 0, Clone 546, Gen 23)
[10:45:34]
[10:45:34] Entering M.D.
NNODES=4, MYRANK=0, HOSTNAME=localhost.localdomain
NNODES=4, MYRANK=3, HOSTNAME=localhost.localdomain
NNODES=4, MYRANK=2, HOSTNAME=localhost.localdomain
NNODES=4, MYRANK=1, HOSTNAME=localhost.localdomain
NODEID=3 argc=15
NODEID=0 argc=15
NODEID=1 argc=15
NODEID=2 argc=15
Written by David van der Spoel, Erik Lindahl, Berk Hess, and others.
Copyright (c) 1991-2000, University of Groningen, The Netherlands.
Copyright (c) 2001-2004, The GROMACS development team,
check out http://www.gromacs.org for more information.

This inclusion of Gromacs code in the Folding@Home Core is under
a special license (see http://folding.stanford.edu/gromacs.html)
specially granted to Stanford by the copyright holders. If you
are interested in using Gromacs, visit http://www.gromacs.org where
you can download a free version of Gromacs under
the terms of the GNU General Public License (GPL) as published
by the Free Software Foundation; either version 2 of the License,
or (at your option) any later version.

(single precision)
[10:45:41] Protein: Protein in POPC
[10:45:41] Writing local files
starting mdrun 'Protein in POPC'
500000 steps, 1000.0 ps.

[10:45:42] Extra SSE boost OK.
[10:45:42] Finalizing output
[10:45:42] nt)
[10:45:42]
[10:45:42] Folding@home Core Shutdown: INTERRUPTED
[cli_3]: aborting job:
Fatal error in MPI_Sendrecv: Error message texts are not available
[0]0:Return code = 0, signaled with Segmentation fault
[0]1:Return code = 0, signaled with Segmentation fault
[0]2:Return code = 0, signaled with Segmentation fault
[0]3:Return code = 1
[10:45:46] CoreStatus = 1 (1)
[10:45:46] Client-core communications error: ERROR 0x1
[10:45:46] Deleting current work unit & continuing...




//

here's another crash with-in minutes same wu run ,core ,gen

[10:52:05] Protein: Protein in POPC
[10:52:05] Writing local files
starting mdrun 'Protein in POPC'
500000 steps, 1000.0 ps.

[10:52:06] Extra SSE boost OK.
[10:52:06] Finalizing output
[10:52:06] UPTED
[cli_3]: aborting job:
Fatal error in MPI_Sendrecv: Error message texts are not available
[0]0:Return code = 0, signaled with Segmentation fault
[0]1:Return code = 0, signaled with Segmentation fault
[0]2:Return code = 0, signaled with Segmentation fault
[0]3:Return code = 1
[10:52:10] CoreStatus = 1 (1)
[10:52:10] Client-core communications error: ERROR 0x1
[10:52:10] Deleting current work unit & continuing...

Re: 2604 Run 0 CL 546 Gen 23 crash on clean restart

Posted: Sat Dec 15, 2007 6:32 am
by 7im
No need to post in separate threads about the same work unit. Please continue the discussion in the other thread.
http://foldingforum.org/viewtopic.php?t=315