Back Midas Rome Roody Rootana
  Root Display GUI  Not logged in ELOG logo
Entry  02 Mar 2008, Exaos Lee, Bug Report, ROODY kills analyzer too!! 
    Reply  03 Mar 2008, Randolf Pohl, Bug Report, ROODY kills analyzer too!! crash-analyzer.txtcrash-roody.txtroody-tarball-dir.txt
       Reply  04 Mar 2008, Exaos Lee, Bug Report, Revision 221 works! But not remotely.  
          Reply  07 Mar 2008, Randolf Pohl, Bug Report, Revision 221 works! But not remotely.  
             Reply  07 Mar 2008, Stefan Ritt, Bug Report, Revision 221 works! But not remotely.  
    Reply  17 May 2009, Randolf Pohl, Bug Fix, [PATCH] ROODY kills analyzer too!! 
       Reply  23 Jul 2009, Exaos Lee, Bug Fix, [PATCH] ROODY kills analyzer too!! 
Message ID: 135     Entry time: 03 Mar 2008     In reply to: 134     Reply to this: 137
Author: Randolf Pohl 
Topic: Bug Report 
Subject: ROODY kills analyzer too!! 
Hi,

I had the very same problem last summer during our beam time. Don't know the revision, but it was the tarball I downloaded around mid-July.
Symptom: The MIDAS (online) analyzer crashed as soon as roody was started. Started, i.e. before any user interaction became even possible. And the crash happened always. 100.0%.

My workaround was: Use roody rev. 220.
No crashes, beam time saved.

I attach 2 outputs I saved at that time.

Oh, yes: I still have the sources and such. Last modified date is July 9 and earlier (see attachment 3). I can send them around if you want to find the revision and bracket the offending commits. (I wish you were using "git". Ever read about git-bisect?)

Cheers,

Randolf

PS: Konstantin, I had promised you to try and pin down the problem after the run. Obviously I forgot, sorry. Please tell me if I can be of any help now.
Attachment 1: crash-analyzer.txt  2 kB  | Hide | Hide all | Show all
Lamb_be>./bin/analyzer -e lamb07
Connect to experiment lamb07...OK
Loading previous online histos from /data/lamb/2007/data/last.root
Root server listening on port 9090...
Running analyzer online. Stop with "!"

 *** Break *** segmentation violation
Using host libthread_db library "/lib64/libthread_db.so.1".
Attaching to program: /proc/2785/exe, process 2785
[Thread debugging using libthread_db enabled]
[New Thread 46971434529216 (LWP 2785)]
[New Thread 1090525504 (LWP 2789)]
[New Thread 1082132800 (LWP 2786)]
0x00002ab8635f5302 in __select_nocancel () from /lib64/libc.so.6
Thread 3 (Thread 1082132800 (LWP 2786)):
#0  0x00002ab85ecde0ab in __accept_nocancel () from /lib64/libpthread.so.0
#1  0x00002ab85f82d510 in TUnixSystem::AcceptConnection ()
   from /usr/local/root/lib/root/libCore.so.5.14
#2  0x00002ab85f822592 in TServerSocket::Accept ()
   from /usr/local/root/lib/root/libCore.so.5.14
#3  0x000000000041a9ff in root_socket_server (arg=<value optimized out>)
    at src/mana.c:5454
#4  0x00002ab862df728a in TThread::Function ()
   from /usr/local/root/lib/root/libThread.so.5.14
#5  0x00002ab85ecd709e in start_thread () from /lib64/libpthread.so.0
#6  0x00002ab8635fb4cd in clone () from /lib64/libc.so.6
#7  0x0000000000000000 in ?? ()

Thread 2 (Thread 1090525504 (LWP 2789)):
#0  0x00002ab8635cce1f in waitpid () from /lib64/libc.so.6
#1  0x00002ab863575491 in do_system () from /lib64/libc.so.6
#2  0x00002ab863575817 in system () from /lib64/libc.so.6
#3  0x00002ab85f831851 in TUnixSystem::StackTrace ()
   from /usr/local/root/lib/root/libCore.so.5.14
#4  0x00002ab85f830a4a in TUnixSystem::DispatchSignals ()
   from /usr/local/root/lib/root/libCore.so.5.14
#5  <signal handler called>
#6  0x000000000041b339 in root_server_thread (arg=<value optimized out>)
    at src/mana.c:5333
#7  0x00002ab862df728a in TThread::Function ()
   from /usr/local/root/lib/root/libThread.so.5.14
#8  0x00002ab85ecd709e in start_thread () from /lib64/libpthread.so.0
#9  0x00002ab8635fb4cd in clone () from /lib64/libc.so.6
#10 0x0000000000000000 in ?? ()

Thread 1 (Thread 46971434529216 (LWP 2785)):
#0  0x00002ab8635f5302 in __select_nocancel () from /lib64/libc.so.6
#1  0x00000000004449e7 in ss_suspend (millisec=1000, msg=0)
    at src/system.c:3285
#2  0x000000000043acf3 in cm_yield (millisec=1000) at src/midas.c:3607
#3  0x000000000041afea in loop_online () at src/mana.c:3785
#4  0x000000000041dfba in main (argc=<value optimized out>, 
    argv=<value optimized out>) at src/mana.c:5756
#0  0x00002ab8635f5302 in __select_nocancel () from /lib64/libc.so.6
[midas.c:1592:] cm_disconnect_experiment not called at end of program
Lamb_be>
Attachment 2: crash-roody.txt  698 Bytes  | Hide | Hide all | Show all
Lamb_be>roody -h localhost
main.cxx: Re-executing 'roody' with compiled-in ROOTSYS='/usr/local/root'
SetOptStat("nemri") got 1001111
Unable to open file "default.xml": No such file or directory
ReadRemoteObject trig_start from folder histos
Folder histos pointer 0
ReadRemoteObject trig_pattern from folder histos
ReadRemoteObject trig_tdc from folder histos
SysError in <TUnixSystem::UnixSend>: send (Broken pipe)
ReadRemoteObject trig_adc from folder histos
ReadRemoteObject trig_detector from folder histos
ReadRemoteObject trig_fadc from folder histos
ReadRemoteObject trig_cool from folder histos
ReadRemoteObject trig_display from folder histos
ReadRemoteObject lambdameter from folder histos
Attachment 3: roody-tarball-dir.txt  1 kB  | Show | Hide all | Show all
ELOG V3.1.4-2e1708b5