UP | HOME

D-Hadronic Analysis - 7.06

Table of Contents

Overview

  • CLEOG: 20050525_MCGEN
  • MCPass2: 20041104_MCP2
  • HadronicDNTupleProc: 20050417_FULL
  • Reproduce the 281/pb result
    • PHOTOS v2.0 (dhad2.0, regular2), PHOTOS v2.15 (dhad2.1, regular3)
    • Signal MC comparison table
    • Data comparison table
  • Run on 818/pb data

Plots

Data - regular3

dhad-2.2 web plots regular3 -t d  

Signal - regular

dhad-2.2 web plots regular -t s

Signal - regular3

dhad-2.2 web plots regular3 -t s

Data - 818ipb

dhad-2.2 web plots 818ipb -t d  

Tables

Data 818/pb

Modedata yield (281ipb)data yield (818ipb)818ipb/281ipb
D0→ K-π+ 25761 ± 16475440 ± 2802.93 ± 0.02
D0B→ K+ π-26259 ± 16675870 ± 2812.89 ± 0.02
D0→ K-π+ π050276 ± 258153449 ± 4543.05 ± 0.02
D0B→ K+ π-π050537 ± 259154724 ± 4573.06 ± 0.02
D0→ Kπ+ ππ+ 39711 ± 216115067 ± 3682.90 ± 0.02
D0B→ K+ ππ+ π39608 ± 216115625 ± 3692.92 ± 0.02
D+ → K-π+ π+ 40249 ± 207116996 ± 3542.91 ± 0.02
D→ K+ ππ-40734 ± 209118309 ± 3562.90 ± 0.02
D+ → K-π+ π+ π012844 ± 15339103 ± 2713.04 ± 0.04
D→ K+ ππ-π012756 ± 15239315 ± 2733.08 ± 0.04
D+ → K0Sπ+ 5789 ± 8217502 ± 1453.02 ± 0.05
D→ K0Sπ5868 ± 8217709 ± 1463.02 ± 0.05
D+ → K0Sπ+ π013275 ± 15842233 ± 3123.18 ± 0.04
D→ K0Sππ013125 ± 15742686 ± 3133.25 ± 0.05
D+ → K0Sπ+ π+ π-8274 ± 13424728 ± 2502.99 ± 0.06
D→ K0Sππ-π+ 8284 ± 13424680 ± 2502.98 ± 0.06
D+ → K+ K-π+ 3519 ± 7310212 ± 1252.90 ± 0.07
D→ KK+π-3501 ± 7310154 ± 1252.90 ± 0.07

818/281 = 2.91

dhad-2.2 table compare yields data divide 281ipb 818ipb

Data 537/pb

Compare the 537ipb yield with 281ipb 537ipb in Data

Modedata yield (281ipb)data yield (537ipb)537ipb/281ipb
D0→ K-π+ 25761 ± 16449675 ± 2271.93 ± 0.02
D0B→ K+ π-26259 ± 16649604 ± 2271.89 ± 0.01
D0→ K-π+ π050276 ± 258103199 ± 3742.05 ± 0.01
D0B→ K+ π-π050537 ± 259104209 ± 3772.06 ± 0.01
D0→ Kπ+ ππ+ 39711 ± 21675356 ± 2971.90 ± 0.01
D0B→ K+ ππ+ π39608 ± 21676017 ± 2991.92 ± 0.01
D+ → K-π+ π+ 40249 ± 20776751 ± 2861.91 ± 0.01
D→ K+ ππ-40734 ± 20977581 ± 2881.90 ± 0.01
D+ → K-π+ π+ π012844 ± 15326280 ± 2242.05 ± 0.03
D→ K+ ππ-π012756 ± 15226584 ± 2272.08 ± 0.03
D+ → K0Sπ+ 5789 ± 8211723 ± 1202.03 ± 0.04
D→ K0Sπ5868 ± 8211845 ± 1212.02 ± 0.03
D+ → K0Sπ+ π013275 ± 15829051 ± 2622.19 ± 0.03
D→ K0Sππ013125 ± 15729639 ± 2652.26 ± 0.03
D+ → K0Sπ+ π+ π-8274 ± 13416506 ± 2141.99 ± 0.04
D→ K0Sππ-π+ 8284 ± 13416449 ± 2151.99 ± 0.04
D+ → K+ K-π+ 3519 ± 736692 ± 1011.90 ± 0.05
D→ KK+π-3501 ± 736652 ± 1011.90 ± 0.05

537/281 = 1.91

copied from: dhad-2.2 table compare yields data divide 281ipb 537ipb

537/281 = 1.91

dhad-2.2 table compare yields data divide 281ipb 537ipb

Compare parameter "sigmap1" data 281ipb 537ipb

Mode281ipb537ipbdiff(%)
D0→ K-π+ 0.003964 ± 0.0000500.004114 ± 0.000035+3.8(+3.0σ)
D0→ K-π+ π00.006716 ± 0.0000520.007049 ± 0.000037+5.0(+6.4σ)
D0→ Kπ+ ππ+ 0.004114 ± 0.0000490.004334 ± 0.000034+5.3(+4.5σ)
D+ → K-π+ π+ 0.004051 ± 0.0000490.004266 ± 0.000035+5.3(+4.4σ)
D+ → K-π+ π+ π00.004579 ± 0.0001020.004929 ± 0.000078+7.6(+3.4σ)
D+ → K0Sπ+ 0.002228 ± 0.0000690.002458 ± 0.000061+10.3(+3.3σ)
D+ → K0Sπ+ π00.006869 ± 0.0001440.007814 ± 0.000122+13.8(+6.6σ)
D+ → K0Sπ+ π+ π-0.003918 ± 0.0001870.004565 ± 0.000143+16.5(+3.5σ)
D+ → K+ K-π+ 0.004948 ± 0.0002480.004978 ± 0.000192+0.6(+0.1σ)
dhad-2.2 table compare parameter sigmap1 data 281ipb 537ipb --set rnd=0.000001

Compare parameter "md" data 281ipb 537ipb

Mode281ipb537ipbdiff(%)
D0→ K-π+ 1.864257 ± 0.0000071.864275 ± 0.0000050.0(+2.6σ)
D0→ K-π+ π01.864276 ± 0.0000081.864284 ± 0.0000050.0(+1.0σ)
D0→ Kπ+ ππ+ 1.864262 ± 0.0000061.864272 ± 0.0000040.0(+1.7σ)
D+ → K-π+ π+ 1.869080 ± 0.0000061.869081 ± 0.0000040.0(+0.2σ)
D+ → K-π+ π+ π01.869092 ± 0.0000151.869076 ± 0.000011-0.0(-1.1σ)
D+ → K0Sπ+ 1.869083 ± 0.0000151.869077 ± 0.000011-0.0(-0.4σ)
D+ → K0Sπ+ π01.869111 ± 0.0000161.869081 ± 0.000013-0.0(-1.9σ)
D+ → K0Sπ+ π+ π-1.869070 ± 0.0000181.869062 ± 0.000015-0.0(-0.4σ)
D+ → K+ K-π+ 1.869091 ± 0.0000231.869097 ± 0.0000170.0(+0.3σ)
dhad-2.2 table compare parameter md data 281ipb 537ipb --set rnd=0.000001

Compare parameter "p" data 281ipb 537ipb

Mode281ipb537ipbdiff(%)
D0→ K-π+ 0.176 ± 0.0590.203 ± 0.028+15.3(+0.5σ)
D0→ K-π+ π00.577 ± 0.0180.552 ± 0.011-4.3(-1.4σ)
D0→ Kπ+ ππ+ 0.482 ± 0.0200.505 ± 0.015+4.8(+1.2σ)
D+ → K-π+ π+ 0.656 ± 0.0410.558 ± 0.027-14.9(-2.4σ)
D+ → K-π+ π+ π00.513 ± 0.0150.500 ± 0.010-2.5(-0.9σ)
D+ → K0Sπ+ 0.624 ± 0.0710.543 ± 0.033-13.0(-1.1σ)
D+ → K0Sπ+ π00.524 ± 0.0150.506 ± 0.007-3.4(-1.2σ)
D+ → K0Sπ+ π+ π-0.509 ± 0.0130.504 ± 0.007-1.0(-0.4σ)
D+ → K+ K-π+ 0.541 ± 0.0310.486 ± 0.021-10.2(-1.8σ)
dhad-2.2 table compare parameter p data 281ipb 537ipb --set rnd=0.001

Compare parameter "xi" data 281ipb 537ipb

Mode281ipb537ipbdiff(%)
D0→ K-π+ -0.100 ± 1.655-0.100 ± 1.981-0.0(0.0σ)
D0→ K-π+ π0-16.797 ± 0.891-15.963 ± 0.572-5.0(+0.9σ)
D0→ Kπ+ ππ+ -8.350 ± 0.994-10.048 ± 0.733+20.3(-1.7σ)
D+ → K-π+ π+ -16.961 ± 1.808-13.944 ± 1.255-17.8(+1.7σ)
D+ → K-π+ π+ π0-10.883 ± 0.746-9.908 ± 0.483-9.0(+1.3σ)
D+ → K0Sπ+ -11.510 ± 3.290-9.438 ± 1.622-18.0(+0.6σ)
D+ → K0Sπ+ π0-10.552 ± 0.751-8.850 ± 0.362-16.1(+2.3σ)
D+ → K0Sπ+ π+ π--9.944 ± 0.605-9.348 ± 0.325-6.0(+1.0σ)
D+ → K+ K-π+ -10.487 ± 1.459-9.123 ± 1.041-13.0(+0.9σ)
dhad-2.2 table compare parameter xi data 281ipb 537ipb --set rnd=0.001

Compare parameter "chisq1" data 281ipb 537ipb

Mode281ipb537ipbdiff(%)
D0→ K-π+ 79.473 ± 0.00094.482 ± 0.00018.9
D0→ K-π+ π049.782 ± 0.00048.453 ± 0.000-2.7
D0→ Kπ+ ππ+ 78.229 ± 0.000118.298 ± 0.00051.2
D+ → K-π+ π+ 100.756 ± 0.000144.695 ± 0.00043.6
D+ → K-π+ π+ π050.059 ± 0.00065.742 ± 0.00031.3
D+ → K0Sπ+ 44.374 ± 0.00059.182 ± 0.00033.4
D+ → K0Sπ+ π059.143 ± 0.00059.199 ± 0.0000.1
D+ → K0Sπ+ π+ π-51.269 ± 0.00053.187 ± 0.0003.7
D+ → K+ K-π+ 63.001 ± 0.00052.884 ± 0.000-16.1
dhad-2.2 table compare parameter chisq1 data 281ipb 537ipb --set rnd=0.001

Compare parameter "chisq2" data 281ipb 537ipb

Mode281ipb537ipbdiff(%)
D0B→ K+ π-81.193 ± 0.00060.059 ± 0.000-26.0
D0B→ K+ π-π060.430 ± 0.00087.035 ± 0.00044.0
D0B→ K+ ππ+ π71.091 ± 0.00079.385 ± 0.00011.7
D→ K+ ππ-110.211 ± 0.000122.951 ± 0.00011.6
D→ K+ ππ-π069.903 ± 0.00049.937 ± 0.000-28.6
D→ K0Sπ61.790 ± 0.00081.452 ± 0.00031.8
D→ K0Sππ066.779 ± 0.00052.488 ± 0.000-21.4
D→ K0Sππ-π+ 48.635 ± 0.00069.994 ± 0.00043.9
D→ KK+π-55.973 ± 0.00068.069 ± 0.00021.6
dhad-2.2 table compare parameter chisq2 data 281ipb 537ipb --set rnd=0.001,namestyle=fnamebar

Details

Software

  • CLEOG : 20050525_MCGEN
  • PASS2 : 20041104_MCP2
  • NTuple : 20050316_FULL

Datasets

281/pb :

datasetslumi (/pb)
data3121.5
data3232.4
data336.5
data3552.0
data3670.7
data37111.1

Extened (818/pb):

datasetslumi (/pb)
data43125.154
data44187.754
data45118.153
data46148.579

DHad-281 CBX

  • Update CBX
    1. Read the last document for updating on page.
    2. Try: dhad update cbx table

      Add local package to setup.sh.

      Script is working.

    3. Need to check the source and result.
      • Table 36: tab:fitResultsData

        From the DHadAttr.py found cbx_data_results_pdg

        Reproduce this table

        dhad table cbx_data_results_pdg
        

        Found error: +-0.0-0.1sigma

        Fix it … done.

        Update to K0S, Reference: table prd_data_results K0S

        dhad table cbx_data_results_pdg K0S 
        

        Done.

      • Table 37: tab:fitResultsRatiosData

        cbx_data_results_bf_ratio_pdg

        dhad table cbx_data_results_bf_ratio_pdg K0S
        

        Done.

      • Table 38: tab:correlationMatrixData
        dhad table cbx_data_correlation_matrix K0S
        

        Done.

      • Table 39: tab:yieldSTResidualsData
        dhad table cbx_residual_single K0S
        

        Done.

      • Table 40: tab:yieldDTResidualsData
        dhad table cbx_residual_double K0S
        

        Done.

      • Table 41: tab:fitResultsDataVariations
        dhad table cbx_data_results_variations K0S
        

        Done.

    4. Update the CBX tables.

      Edit the tables in the DHadAttr.py

      dhad update cbx table
      

      Done.

      PS file is OK.

      Translate all the eps figure to pdf.

      dhad update cbx epstopdf ... OK.
      

      PDF file is OK.

    5. Check in the latex file and figures

      Done. Email to Anders and David. OK.

    6. Release the document

      From page https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/DocumentDatabase

      CBX2009-056

      Edit the CBX.

      Submitted. In the link:

      https://redms.classe.cornell.edu/record/1965

      Edit the DHad page … OK.

      Edit the Twiki page … done.

Start up

List the Software status

  • What I have
    • Yield extraction from DHad NTuple
    • Fits for tag yields
    • Branching Fraction fitter
    • Scripts creating most of the tables for CBX and PRD
  • What I don't have (know)
    • How to run the DHad code to produce the DHad NTuple
    • How to produce the root squre scaled plots in ROOT

Retrieve Source code

  • Ask Peter … sent.

    Reply from Peter:

    The directories in question were archived as per ticket #16977.  They
    are about a terabyte worth of (primarily) MC datasets and data/MC
    ntuples.  I don't think much code was involved (possibly some of my
    stuff for estimating backgrounds etc, but that should be redone for
    the full dataset anyway).
    
  • Ask DLK … sent.

    Reply from David:

    Bill Brangan is the keeper of backups.  Easiest way to reach him is
    to send an email request to service-lepp.  Be sure to specify the
    file names (or subdirectory) and what directory it was in.  I hope
    this was in someone's personal home disk area so it is readily
    available on backup.  Otherwise you may be in trouble.  Are you
    sure that it has migrated to tape?  Was it ever committed to the
    CLEO cvs repository?  dlk
    

    Message from Werner:

    Can you say what exactly is missing?  Was this
    on a common tem disk? Private code/data/etc in a user's area is not
    usually moved to tape without notification.
    

    Need to explore the source code!

Check out source code

Start from Software Doc page

  1. https://www.lepp.cornell.edu/restricted/CLEO/CLEO3/soft/intro/Tutorial/index.html
  2. Peter's talk http://www.lns.cornell.edu/public/CLEO/CLEO101/2007/Day5a/070614_cleo101.pdf

    Day 1. Richard Gray:

    Check the current release:

    c3rel
    

    Use this 20080228_FULL in the setdhad.sh

    Day 3. Peter's talk:

    D0toKpi example.

  3. Ask help from Werner

    Sent email to Werner.

    Reply from Werner:

    To get Peter's ntuple-making processor, do:
    
    setenv CVSROOT /nfs/cleo3/cvsroot
    cvs co HadronicDNtupleProc
    
    You should be able to compile this like any other CLEO-III code.
    
    For the suez job, it looks like you should use
    HadronicDNtupleProc/Test/loadHadronicDNtupleProc.tcl (see the
    top of this file for environment variables you need to set), but
    I would check with Peter to be sure.
    

    Check out the code :

    cvs -d /nfs/cleo3/cvsroot co HadronicDNtupleProc
    

    Error message:

    cvs checkout: warning: cannot write to history file /nfs/cleo3/cvsroot/CVSROOT/history: Permission denied
    cvs checkout: Updating HadronicDNtupleProc
    cvs checkout: failed to create lock directory for `/nfs/cleo3/cvsroot/Offline/src/HadronicDNtupleProc' (/nfs/cleo3/cvsroot/Offline/src/HadronicDNtupleProc/#cvs.lock): Permission denied
    cvs checkout: failed to obtain dir lock in repository `/nfs/cleo3/cvsroot/Offline/src/HadronicDNtupleProc'
    cvs [checkout aborted]: read lock failed - giving up
    

    Ask Werner … sent.

    Reply:

    Try using "cleo3cvs" instead of "cvs". 
    

    Check out the code :

    export CVSROOT=/nfs/cleo3/cvsroot 
    cleo3cvs co HadronicDNtupleProc ... OK. 
    

    Need to make the build area … see page.

    Re-set the variables in the setup.sh.

    c3make => make.log

    Check out the module CWNFramework

    cleo3cvs co CWNFramework 
    

    c3make => make.log.1

    Read Peter's webpage: http://www.lepp.cornell.edu/~ponyisi/cwnframework.html

    From the MyProc/MyProc directory, run the script CWNFramework/Test/genDefinitions.py . This will autogenerate the header files for the tuple.

    From the MyProc directory, run the script CWNFramework/Test/genSkels.py

    Still the same compile error.

    Ask Peter … sent.

    Reply from Peter:

    These appear to be template problems with compiling DChain.  Are you
    somehow getting the wrong version of g++?
    

    Try to uncomment the cms environment.

    Still the same problem.

    Raise question on the CLEO computing HN. sent.

    Message from Paras:

    I am having a very similar error right now with the same
    release, which I also can't seem to explain.
    
    I just replaced my current piece of code with a backup piece of code
    that I knew to be working and I have still have the same problem.
    

    Follow up from Paras:

    OOPS, I need to compile my code more often...
    compiling on the proper machine, lnx134, solved the problem for me
    hopefully that is also Xin's problem?
    

    Messge from Dan:

    Are you compiling on lnx134?  If not, what do you get on that system?
    Remember, building CLEO code on SL4+ systems is not supported, and most
    of our desktops are now SL4+.
    

    Log on to lnx134. Compile OK.

Test Run on data

Get the example tcl file.

suez -f /home/xs32/work/CLEO/analysis/DHad/script/tcl/test.tcl

Error message:

Suez.test> #Load in your very own processor.
Suez.test> proc sel HadronicDNtupleProc
%% ERROR-DynamicLoader.DLSharedObjectHandler: /home/xs32/tmp/build/20080228_FULL/Linux/shlib/HadronicDNtupleProc.so_20080228_FULL: undefined symbol: _ZNSs4_Rep20_S_empty_rep_storageE
%% SYSTEM-Interpreter.TclInterpreter: Tcl_Eval error: ERROR: cannot load HadronicDNtupleProc.

Found another example from Peter's dir:

cp /nfs/cor/user/ponyisi/daf9/simple-dtuple-data31.sh $dhad/script/bash/

Change it to test_data31.sh, can use the loadHadronicDNtupleProc.tcl

suez -f /home/xs32/work/CLEO/analysis/DHad/script/tcl/test.tcl
%% SYSTEM-Interpreter.TclInterpreter: Tcl_Eval error: can't read "env(INPUTDATA)": no such variable

Use the test_data31.sh

;Suez.test> proc sel HadronicDNtupleProc
%% ERROR-DynamicLoader.DLSharedObjectHandler: /home/xs32/tmp/build/20080228_FULL/Linux/shlib/HadronicDNtupleProc.so_20080228_FULL: undefined symbol: _ZNSs4_Rep20_S_empty_rep_storageE
%% SYSTEM-Interpreter.TclInterpreter: Tcl_Eval error: ERROR: cannot load HadronicDNtupleProc.

Use everything …

cd /home/xs32/work/CLEO/analysis/DHad/tmp
cp /nfs/cor/user/ponyisi/daf9/2005-3/local_setup . 

Edit the localsetup … Notice in the end c3rel 20050316_FULL, use c3rel 20080228_FULL.

%% ERROR-DynamicLoader.DLSharedObjectHandler: /home/xs32/tmp/build/20080228_FULL/Linux/shlib/HadronicDNtupleProc.so_20080228_FULL: undefined symbol: _ZNSs4_Rep20_S_empty_rep_storageE
%% SYSTEM-Interpreter.TclInterpreter: Tcl_Eval error: ERROR: cannot load HadronicDNtupleProc.

Ask questions on Dhad HN … sent.

Message from Fan:

You can check your Makefile to make sure that you have already
included all necessary Libs in that file.

From Werner:

Do you have the full log file?

From Sourik:

You could use "c++filt" to digest error message.

From Peter:

You're sure you compiled this on lnx134?  This is some kind of C++
string problem (as you can see by running c++filt
_ZNSs4_Rep20_S_empty_rep_storageE) so it sounds possibly like a
compiler incompatibility.

Check the CBX "code" section:

Monte Carlo was generated with the following prescription:
  [Cleog] was run using release 20050525_MCGEN.
  [MCPass2] was run using release 20041104_MCP2

The creation of ntuples was done in the 20050316_FULL release,
with the following additional package tags:
   MCCCTagger                             v02_00_04
   HadronicDNtupleProc                    ponyisi060929

Check out the additional packages :

export CVSROOT=/nfs/cleo3/cvsroot 
cleo3cvs co -rv02_00_04  MCCCTagger
cleo3cvs co -rponyisi060929 HadronicDNtupleProc 

Compile with 20050316_FULL release at lnx134:

c3rel 20050316_FULL
export USER_BUILD=/home/xs32/tmp/build/$C3LIB
export USER_SHLIB=${USER_BUILD}/Linux/shlib
c3make 

Need to check the CWNFramework:

cleo3cvs co CWNFramework

Compile OK.

Test :

. ./test_data31.sh >& $dhad/log/2009/0210/build.txt 

Error build.txt:

Compile again, and test, same result.

Ask on HN… sent.

Message from Surik:

Try to use debugger to find the cause of "abort signal." 
There is  well documented CLEO note, which will help
you to DEBUG your code.

https://www.lepp.cornell.edu/restricted/CLEO/CLEO3/soft/howto/howto_debugSharedProcessor.html

Hope this helps.

Sent reply.

Message from Werner:

You're running in 20050316_FULL, but somehow, you linked against the 
20080228_FULL version of ROOT:

[ Look for
/nfs/cleo3/Offline/rel/20080228_FULL/other_sources/Root/lib/libTree.so
in your build.txt ]

This may not be the cause of the problem, but certainly looks suspicious.

Trace the ROOT lib …

which root
/nfs/cleo3/Offline/rel/20080228_FULL/other_sources/Root/bin/root

Comment out the ROOT and Python part in the setup.sh.

Log off the terminal and restart on lnx224: => build.txt

setdhad
cd /home/xs32/work/CLEO/analysis/DHad/test/script
. ./test_data31.sh >& $dhad/log/2009/0211/build.txt 
%% ERROR-DynamicLoader.DLSharedObjectHandler: HadronicDNtupleProc.so_20050316_FULL: cannot open shared object file: No such file or directory
%% SYSTEM-Interpreter.TclInterpreter: Tcl_Eval error: ERROR: cannot load HadronicDNtupleProc.

Use c3rel 20050316_FULL for general now. On lnx224: => build.txt.1 OK.

setdhad
. /home/xs32/work/CLEO/analysis/DHad/test/script/test_data31.sh >& $dhad/log/2009/0211/build.txt.1 
***** Summary Info *****
Stream beginrun : 1
Stream event : 11
Stream startrun : 1
Processed 13 stops.

Try not use the setdhad in lnx224: => build.txt.2 same problem.

Change the sequence of the settings in the test_data31.sh same as the defult one. => build.txt.3 same problem.

Use all the same as defult one. (i.e. must use the c3rel $C3LIB again!) => build.txt.4 OK!

Find the output

ll /home/xs32/work/CLEO/analysis/DHad/tmp
-rw-r--r--  1 xs32 cms   30K Feb 11 10:21 data31_dskim_evtstore.root

Open with ROOT, OK.

Reply to HN … sent.

Ready to process 281/pb.

Generate Signal MC (2.0)

First Try

  1. Check the webpage .

    https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/Private/SW/CLEOcMC

    https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/Private/SW/CLEOcSignalMC

  2. Generate the signal MC D0 > K pi+ against generic D0-bar.
  3. Create runmc.scr
  4. Create cleogddbar.tcl
  5. Create mcpass2ddbar.tcl
  6. Create User decay file (ddbar.dec) used in the cleog script above

    Put the decay file in the tcsh dir for now.

  7. Run it.
    ./run_mc.src  ... 
    

    Connection to lnx7228.lns.cornell.edu closed. (exit in the last line of the script)

    Need to ask which machine to use for mcgen. … sent.

    Use lnx224 for now.

    No events processed => run1.log

    Ask on CLEOG HN… OK. (not found.)

    Ask on BugFix HN and cc to DHad group … sent.

    Got help from Laura, need to specify the directory of the decay file.

    Change it and run again …

    This time, there is the output run_200978_mcp2.pds, but still error in the output.

    Change the run_$MCRUN_cg.log to run_$env(MCRUN)_cg.log

    Change the run_$MCRUN_mcp2.log to run_$env(MCRUN)_mcp2.log

    Run again. … Not work. change to back.

    Register the log file => run.log

    Report on HN … sent.

    Message from Laura:

    So my MC generation code is a bit complicated. I submit files using:

    /home/ljf26/analysis/signalMC/submit_mc.scr
    
    The script that actually calls cleog and mcpass2 is
    
    /home/ljf26/workshop/DtoPiLNuProc/Class/gen_anyMC.scr
    
    These are probably a lot more complicated than you need, but they will
    give you a general idea of what you need to do, and they did work a
    few months ago at least.  Also, my analysis processor is in
    /home/ljf26/workshop/DtoPilnuProc/class/DtoPilnuProc.cc.  It is very
    expansive and messy, but the variable isMC is the boolean that extract
    to find out if I'm using data or MC.
    

    Message from Pete:

    /cdat/rd5/cleo3/Offline/rel/20071023_MCP2/src/include/StorageManagement/SMStorageHelper.h:70:
    T* SMStorageHelper<T>::deliver(SMSourceStream&, unsigned int) [with T =
    MCParticle]: Assertion `(iVersion-kFirstVersionNumber) <
    m_deliverers.size()' failed.
    
    Are you using the MCGEN release 20071207_MCGEN or newer MCGEN?
    
    If you are (and I think you are), you need to use a MCP2 release with
    the suffix "_A_1". See the end of the "Library release used" section on
    the followinf wikipage:
    https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/Private/SW/CLEOcMCstatus
    

    Check with c3rel, saw 20071023_MCP2_A.

    Try this one…. => run1.log

    Sent reply.

    Message from Surik:

    Try to use absolute path when you are redirecting the suez
    output. At this point your are using just file name.
    

    Message from Sean:

    To get those log files, you might try specifying the absolute path of
    the log file you're writing to, e.g. in your run_mc.scr, doing
    
    suez -f /home/xs32/work/CLEO/analysis/DHad/script/tcl/cleog_ddbar.tcl >&
    /home/xs32/work/CLEO/analysis/DHad/script/tcl/run_$MCRUN_cg.log
    
    (or whatever directory you want to output into)
    

    Try the absolute path … same problem.

    Try to echo the MCRUN variable before suez…

    echo "test" > run_$MCRUN_cg.log => not work.
    echo "test" > $MCRUN_cg.log => not work.
    echo "test" > $MCRUN.log => OK.
    

    Reply to HN… sent.

    Message from Sean:

    Yes, try using
    /cdat/tem/xs32/cleo/run_${MCRUN}_cg.log
    instead - I think that will interpret the variable correctly.
    (Though any wandering shell-scripting gurus can correct me if I'm wrong).
    

    Try this one …

    echo "test" > run_${MCRUN}_cg.log OK.
    

    Try the run file … done.

    Save the log file and pds file … OK.

    run200978cg.log, run200978mcp2.log

    Send back, mention the change of the Twiki page and raise the question of using two different output dir for the pds file. … sent.

    Message from Surik:

    Since you find the real cause, I am wondering if you do not 
    specify absolute path for your log files in suez command lines,
    do you find your log files in the current directory. You have 
    "cd" command line before the suez command line.
    
    I am just curious,
    

    Try the relative path … OK.

    Reply to Surik. …

  8. Ask Peter for the exact Release number for previous Signal MC generation

    Look at the dir /nfs/cor/user/ponyisi/daf9/ first … not able to tell.

    Send email … sent.

    Message from Peter:

    You can find my MC generation scripts in ~ponyisi/cleog/ ; in particular
    ~ponyisi/cleog/scripts-summerconf-photosint contains scripts for
    generating MC with the PHOTOS interference on.  The top-level script is
    fire_up_jobs_new.sh; it should be fairly clear what it does if you read
    it (makes more sense than most of my code!)
    
    The releases are defined in that file; they appear to have been
    20050525_MCGEN and 20041104_MCP2.  This *should* be mentioned in the
    CBX, but I might have forgot to put this in.
    

    Need to reply … sent.

Reproduce DtoKpi MC

CLEOG

From the top file fire_up_jobs_new.sh

mkdir $dhad/test
mkdir $dhad/test/script
cp  ~ponyisi/cleog/scripts-summerconf-photosint/fire_up_jobs_new.sh $dhad/test/script 
cp  ~ponyisi/cleog/scripts-summerconf-photosint/test_mode_list $dhad/test/script 
cp  ~ponyisi/cleog/scripts-summerconf-photosint/simple-cleog-generic-array.sh $dhad/test/script 
mkdir $dhad/test/script/tag_numbers
cp  ~ponyisi/cleog/scripts-summerconf-photosint/tag_numbers/number_of_jobs $dhad/test/script/tag_numbers 
cp  ~ponyisi/cleog/scripts-summerconf-photosint/tag_numbers/Single_D0_to_Kpi $dhad/test/script/tag_numbers
mkdir $dhad/test/script/tag_decfiles
cp  ~ponyisi/cleog/scripts-summerconf-photosint/tag_decfiles/Single_D0_to_Kpi.dec $dhad/test/script/tag_decfiles
cp  ~ponyisi/cleog/genmc_anders.tcl $dhad/test/script 

Edit the fire_up_jobs_new.sh as fire_up_job_DtoKpi.sh

Try the cleog first.

. ./fire_up_job_DtoKpi.sh
Your job-array 1185697.1-10:1 ("cleog-Single_D0_to_Kpi Single_D0_to_Kpi") has been submitted

Check the qsub status … done.

Change the Email address.

. ./fire_up_job_DtoKpi.sh
Your job-array 1185698.1-10:1 ("cleog-Single_D0_to_Kpi Single_D0_to_Kpi") has been submitted

From the output:

nl: /home/xs32/work/CLEO/analysis/DHad/test/script/runlist: No such file or directory
ls: /cdat/lns150/disk2/c3mc/RandomTriggerEvents/data*/: Input/output error

Fix the runlist:

cp ~ponyisi/cleog/scripts-summerconf-photosint/runlist $dhad/test/script/

Try again:

. ./fire_up_job_DtoKpi.sh
Your job-array 1185699.1-10:1 ("cleog-Single_D0_to_Kpi Single_D0_to_Kpi") has been submitted

Same output.

Remove the output and try again.

EvtGen:Could not open /home/xs32/work/CLEO/analysis/DHad/test/script/tag_decfiles/Single_D0_to_Kpi.dec

Copy the dec file again:

mkdir $dhad/test/script/tag_decfiles
cp  ~ponyisi/cleog/scripts-summerconf-photosint/tag_decfiles/Single_D0_to_Kpi.dec $dhad/test/script/tag_decfiles

Try again:

. ./fire_up_job_DtoKpi.sh
Your job-array 1185702.1-10:1 ("cleog-Single_D0_to_Kpi Single_D0_to_Kpi") has been submitted

Check the output message … noticed:

==========================                      
 PHOTOS, Version:  2. 0                       
 Released at:  16/11/93                       
==========================                   
%% ERROR-DataStorage.dataStringTagsToDataKeysUsingRecordContents: unknown TypeTag( FATable<CalibratedSVRphiHit> )
%% WARNING-MCInfo.MCDecayMode: Potential Energy nonconservation: vpho --> psi(3770) gamma 
%% WARNING-MCInfo.MCParticleProperty: Adding a decay mode with a branching fraction of zero!!!!
mode 41: frac =        0 D+ --> anti-K0 pi+ pi+ pi-  (Model 0)
%% WARNING-MCInfo.MCDecayMode: Potential Energy nonconservation: B_c+ --> anti-B_s0 tau+ anti-nu_tau 
%% WARNING-MCInfo.MCDecayMode: Potential Energy nonconservation: B_c- --> B_s0 tau- nu_tau 
%% WARNING-MCInfo.MCParticleProperty: Adding a decay mode with a branching fraction of zero!!!!
mode 12: frac =        0 psi(4040) --> D0 anti-D0 pi0  (Model 0)
%% WARNING-wtk_drift_stereo:  disc <0. Axial geometry is used. Drift dist, cm =  0.567896426
%% WARNING-check_geant_hit:  Illegal Ion Distance =   0.0212069992
%% WARNING-MCResponse.MCCathodesResponseProxy: Too many anode responses
%% WARNING-CDOffCal.DataDriftFunction: inverse velocity has illegal value
%% WARNING-check_geant_hit:  Illegal sinXangle =   1.02591395 restricted
***** Summary Info *****

Stream physics : 1
Stream beginrun : 1
Stream endrun : 1
Stream event : 3137
Stream startrun : 1
Processed 3141 stops.

Find out the output pds file … OK.

Need to ask the line … OK.

ll /home/xs32/work/CLEO/analysis/DHad/test/cleog_0214
total 1.1G
-rw-r--r--  1 xs32 cms 113M Feb 10 17:14 cleog_Single_D0_to_Kpi_10.pds
-rw-r--r--  1 xs32 cms 104M Feb 10 17:21 cleog_Single_D0_to_Kpi_1.pds
-rw-r--r--  1 xs32 cms 104M Feb 10 17:23 cleog_Single_D0_to_Kpi_2.pds
-rw-r--r--  1 xs32 cms 104M Feb 10 17:24 cleog_Single_D0_to_Kpi_3.pds
-rw-r--r--  1 xs32 cms 109M Feb 10 17:21 cleog_Single_D0_to_Kpi_4.pds
-rw-r--r--  1 xs32 cms 110M Feb 10 17:21 cleog_Single_D0_to_Kpi_5.pds
-rw-r--r--  1 xs32 cms 111M Feb 10 17:21 cleog_Single_D0_to_Kpi_6.pds
-rw-r--r--  1 xs32 cms 114M Feb 10 17:13 cleog_Single_D0_to_Kpi_7.pds
-rw-r--r--  1 xs32 cms 115M Feb 10 17:13 cleog_Single_D0_to_Kpi_8.pds
-rw-r--r--  1 xs32 cms 119M Feb 10 17:14 cleog_Single_D0_to_Kpi_9.pds

Ask Peter about the noticed error and the below reason … sent.

ls -R /cdat/lns150/disk2/c3mc/RandomTriggerEvents/data*/ > /dev/null 

Message from Peter:

It looks fine.  The strange line you noticed was to ensure that the
random events files were available (if there's an NFS glitch they may
not appear and the job will fail).  However they have probably moved
them to a different directory and the path I use is wrong.  You can
remove that line.

PASS2

cp  ~ponyisi/cleog/scripts-summerconf-photosint/simple-pass2-generic-array.sh $dhad/test/script 
cp  ~ponyisi/cleog/mcp2_anders.tcl  $dhad/test/script 
cd /home/xs32/work/CLEO/analysis/DHad/test/script/
. ./fire_up_job_DtoKpi.sh

Change the Email !!!

Change the mcp2_anders.tcl:

set histout /home/xs32/work/CLEO/analysis/DHad/test/hist$env(BATCH).rzn

Change the simple-pass2-generic-array.sh, comment out #cd ~/cleog

Add a log output for pass2: => Not work.

#$ -o /home/xs32/work/CLEO/analysis/DHad/test/pass2.txt

Use log as command line.

suez -f mcp2_anders.tcl >& /home/xs32/work/CLEO/analysis/DHad/test/log/pass2_${SGE_TASK_ID}.log

Test … => pass21.log :

 %% SYSTEM-Interpreter.TclInterpreter: Error opening file mcp2_anders.tcl

Need to use cd $SCRIPTDIR !

Test … => pass21.log.1 Noticed parts:

%% NOTICE-ConstantsPhase2Delivery.DBCP2Proxy: DBMUHomogenousValues using version 1.1
%% NOTICE-ConstantsPhase2Delivery.DBCP2Proxy: DBMUPoissonValues using version 1.1
*** S/R ERPROP   IERR =    2
*** Error in subr. TRPROP   2  called bysubr. ERPROP
*** S/R ERPROP   IERR =    2
*** Error in subr. TRPROP   2  called bysubr. ERPROP

NTUPLE

CLOCK: 2009-02-11 Wed 15:34 2009-02-11 Wed 15:50 => 0:16

cp  ~ponyisi/cleog/scripts-summerconf-photosint/simple-dtuple-generic-array.sh  $dhad/test/script 

Change:

  • Email
  • Edit the content of the local_setup into the same sh file.
  • Add log output from the suez .
  • Set the event to be 10 in the tcl.

Test … => signalD0toKpi.log

%% SYSTEM-Interpreter.TclInterpreter: Error opening file HadronicDNtupleProc/Test/loadHadronicDNtupleProc.tcl

Use the absolute path… => signalD0toKpi.log.1

 %% SYSTEM-Interpreter.TclInterpreter: Error opening file /home/xs32/work/CLEO/analysis/DHad/test/script/tmp/dt-Single_D0_to_Kpi

Make the tmp dir and try again … => signalD0toKpi.log.2

%% ERROR-DynamicLoader.DLSharedObjectHandler: /home/xs32/tmp/build/20050316_FULL/Linux/shlib/MCBeamEnergyFromMCBeamParametersProd.so_20050316_FULL: cannot open shared object file: No such file or directory
%% SYSTEM-Interpreter.TclInterpreter: Tcl_Eval error: ERROR: cannot load /home/xs32/tmp/build/20050316_FULL/Linux/shlib/MCBeamEnergyFromMCBeamParametersProd.

Need the lib: MCBeamEnergyFromMCBeamParametersProd

Check the CBX … Not found.

Ask Peter which version to check out and which release to compile. … sent.

Message from Peter:

It's probably fine for you to copy it from
/nfs/cor/user/ponyisi/daf9/2005-3/ and compile it in 20050316_FULL.

Copy :

cp -r /nfs/cor/user/ponyisi/daf9/2005-3/MCBeamEnergyFromMCBeamParametersProd $dhad/src/

On lnx134:

Change the setup.sh : c3rel 20050316_FULL

setdhad
cd $dhad/src/
c3make 

Compile OK.

Try again … => signalD0toKpi.log.3 OK.

Do the full sample (go) … done.

Produce D0BtoKpi

  1. CLEOG
    cp fire_up_job_DtoKpi.sh gen_Single_D0B_to_Kpi.sh 
    

    Change

    FILELISTINGS="Single_D0B_to_Kpi"
    

    Provide Single_D0B_to_Kpi file.

    Start job … => cleog-SingleD0BtoKpi-1.txt Not work.

    Error:

    cat: /home/xs32/work/CLEO/analysis/DHad/test/script/tag_numbers/Single_D0B_to_Kpi: No such file or directory
    

    Copy the needed file:

    cp  ~ponyisi/cleog/scripts-summerconf-photosint/tag_numbers/Single_D0B_to_Kpi $dhad/test/script/tag_numbers
    cp  ~ponyisi/cleog/scripts-summerconf-photosint/tag_decfiles/Single_D0B_to_Kpi.dec $dhad/test/script/tag_decfiles
    

    Try again … => cleog-SingleD0BtoKpi-1.txt OK.

  2. PASS2

    Edit gen_Single_D0B_to_Kpi.sh : DOPASS2=1

    Run … done.

  3. NTUPLE

    Edit gen_Single_D0B_to_Kpi.sh : DONTUPLE=1

    Run … done.

Gen all Siganl MC

  1. Ask Brian where to put the big files … sent.

    Got dir at : /nfs/cor/an2/xs32

    Make softlink

    ln -s /nfs/cor/an2/xs32 ~/disk/2
    mkdir ~/disk/2/cleo
    mkdir ~/disk/2/cleo/dhad
    mv dat ~/disk/2/cleo/dhad/
    ln -s  ~/disk/2/cleo/dhad/dat $dhad
    
  2. Start the script for CLEOG

    Goal:

    dhad gen cleog
    

    Start the script in 2.0

    cd /home/xs32/work/CLEO/analysis/DHad/src/2.0/python
    

    With initial tag: V02-00-00

    cvs tag V02-00-00
    

    Re-direct the dhad command. Linked to dhad-2.0.

    Do one mode first:

    dhad gen cleog -m 0
    

    Pause for now. Get the generation running first!

  3. Generate all of the signal modes
    cp  ~ponyisi/cleog/scripts-summerconf-photosint/tag_numbers/* $dhad/src/2.0/gen/tag_numbers
    cp  ~ponyisi/cleog/scripts-summerconf-photosint/tag_decfiles/* $dhad/src/2.0/gen/tag_decfiles
    

    Cleog … done.

    cd  $dhad/src/2.0/gen
    . ./generate_mc.sh
    

    Qstat with one exception:

    1187664 0.55009 cleog-Sing xs32         Eqw   02/12/2009 17:26:31       1 4
    

    Ask on HN for how to re-submit …

    Message from Surik:

    To learn why it is in Eqw state, you could do 
    "qstat -j JobId" (I in your case JobId is 1187664).
    
    If you find that that there is no error from user side,
    then you could clear error state and resubmit it 
    by saying: qmod -cj JobId.
    

    qstat -j 1187664 => jobstat.txt

    Message from Chul Su:

    qmod -c 1187664
    man qmod
    
    Job-array task 1187664.4 is not in error state
    

    Message from Laura (forward from dlk):

    Because of the &%#@* "network glitch", it is fairly common to
    have a batch job put into Eqw status through no fault of your own:
    
    1060079 0.56000 drawdists_ ljf26        Eqw   10/05/2008 22:37:58
    
    This note is about how to clear that status without deleting
    and resubmitting that job.  Do a
    
    qstat -j 1060079   (Queue status for the job 1060079)
    
    and look through the voluminous output for the line describing the
    cause of the error:
    
    error reason    1:          10/05/2008 22:38:05 [1648:17407]: error: can't chdir to /home/ljf26: No such file or directory
    
    This obviously absurd contention is because there was a
    temporary dropout of the network connection between the queuing master
    server and the /home disk.  (Sometimes if you write the output to the
    /tem disk, batch can't find that.)  You can clear the error and get a
    retry by doing
    
    qmod -c 1060079   (Queue modify job 1060079 by clearing error status)
    

    Pass2 … done.

    Ntuple … done.

Fix the three channels

Edit fixed_single_mode_list. with three channels.

Edit generate_mc.sh for cleog.

CLEOg …

cat: /home/xs32/work/CLEO/analysis/DHad/src/2.0/gen/tag_numbers/Dp_to_Kspipipi: No such file or directory

Add "Single" in the mode list. Re-run. … done.

PASS2 … done.

Still one left:

xs32_lnx224% ll /home/xs32/work/CLEO/analysis/DHad/dat/signal/2.0/cleog_0214/
total 191M
-rw-r--r--  1 xs32 cms 172M Feb 12 20:09 cleog_Single_Dm_to_Kspipipi_1.pds
-rw-r--r--  1 xs32 cms  20M Feb 18 10:59 cleog_Single_Dp_to_Kspipi0_4.pds
xs32_lnx224% tail  /home/xs32/work/CLEO/analysis/DHad/dat/signal/2.0/log_c_0525_p2_1104/cleog-Single_Dp_to_Kspipi0-4.txt
>> Wed Feb 18 10:59:39 2009 Run:  206080 Event:     460 Stop: event <<
%% NOTICE-Processor.MCRunEvtNumberProc:  RandomGenerator seeds=546597654, 176706065
%% NOTICE-Processor.RunEventNumberProc: Run: 206080 Event: 460
%% NOTICE-Processor.RunEventNumberProc: Run: 206080 Event: 25718

>> Wed Feb 18 10:59:40 2009 Run:  206080 Event:     461 Stop: event <<
%% NOTICE-Processor.MCRunEvtNumberProc:  RandomGenerator seeds=1605634292, 349285818
%% NOTICE-Processor.RunEventNumberProc: Run: 206080 Event: 461
/nfs/cleo3/Offline/rel/20050525_MCGEN/bin/Linux/g++/suez: line 317:  4465 Floating point exception(core dumped) $ECHO $exe $SCRIPT $options
Wed Feb 18 10:59:42 EST 2009

Notice the cleog_Single_Dm_to_Kspipipi_1.pds, which cause the 4% difference. The cleog looked OK. Need to re-run the pass2.

Need to address these two later … Proceed with Ntuple first.

Ntuple … done.

Get yield:

Kpipi0: mode 1
Kspipi0: mode 203
Kspipipi: mode 204
dhad yield regular2 -t s -m 1
dhad yield regular2 -t s -m 203
dhad yield regular2 -t s -m 204
dhad fit regular2 -t s -m 1 --qsub                                                                                       

log-2009-02-18 21:02:07

dhad fit regular2 -t s -m 203 --qsub                                                                                      $

log-2009-02-18 21:02:52

dhad fit regular2 -t s -m 204 --qsub                                                                                    $

log-2009-02-18 21:06:32

Move on to new sample and fix them.

Generate Signal MC with PHOTOS-2.15 (2.1)

First Try

Use 20050417_FULL

dhad yield test -t s --sign 1

Need to specify the root file.

cp $dhad/test/dtuple_c_0525_p2_1104/Single_D0_to_Kpi.root $dhad/dat/signal/2.0/test 

Now, just backup the old link and use the new as test.

dhad yield test -t s -m 1 --sign 1

Error:

Error in <TFile::TFile>: file /home/xs32/work/CLEO/analysis/DHad/7.06/dat/signal/test/Single_D0_to_Kpipi0.root does not exist

Need to produce more signal MC … Used the wrong mode number!

dhad yield dtuple_c_0525_p2_1104 -t s -m 0  OK.

Fit for yield:

dhad fit dtuple_c_0525_p2_1104 -t s -m 0

Compare the yield with previous signal MC:

dhad table compare yields dtuple_c_0525_p2_1104  -t s -m 0 

Consider using the previous framework to reduce non-essential work …

cd /home/xs32/work/CLEO/analysis/DHad/7.06/dat/signal
ln -s /home/xs32/work/CLEO/analysis/DHad/dat/signal/2.0/dtuple_c_0525_p2_1104/ regular2

Use the $dhad/scripy/python as the main dhad.

dhad yield regular2 -t s -m 0  OK.
dhad fit regular2 -t s -m 0 ... processing OK.

Need the qsub.

dhad yield regular2 -t s ... done.
dhad fit regular2 -t s --qsub

Set the complete qsub env.

log-2009-02-16 20:24:23 … done.

dhad table compare yields regular2 -t s

=> page.

Need to compare efficiency, or generate same number of events as in CBX.

List the generated numbers:

ModeCBX 281/pbRegular 2Regular 3
D0toKpi627403137x106274x10
D0BtoKpi627403137x106274x10
D0toKpipi019735019735x10No change
D0BtoKpipi019735019735x10No change
D0toKpipipi994809948x10No change
D0BtoKpipipi994809948x10No change
DptoKpipi799807998x10No change
DmtoKpipi799807998x10No change
DptoKpipipi0733807338x10No change
DmtoKpipipi0733807338x10No change
DptoKspi800008000x10No change
DmtoKspi800008000x10No change
DptoKspipi0571605716x10No change
DmtoKspipi0571605716x10No change
DptoKspipipi393103931x10No change
DmtoKspipipi393103931x10No change
DptoKKpi200002000x10No change
DmtoKKpi200002000x10No change

Only need to re-generate the first one.

Create fixed_single_mode_list. with only one mode.

CLEOg , PASS2, Ntuple … done.

dhad yield regular2 -t s -m 0  OK.
dhad fit regular2 -t s -m 0 --qsub

log-2009-02-17 12:21:37

Make the compare page:

dhad table compare yields regular2 -t s
Modesignal diff(%)
D0toKpi1.08
D0BtoKpi1.57
D0toKpipi00.07
D0BtoKpipi0-19.95
D0toKpipipi0.72
D0BtoKpipipi0.70
DptoKpipi-0.44
DmtoKpipi0.42
DptoKpipipi0-0.05
DmtoKpipipi0-0.00
DptoKspi-0.24
DmtoKspi-0.42
DptoKspipi0-12.65
DmtoKspipi0-4.19
DptoKspipipi-10.17
DmtoKspipipi0.72
DptoKKpi-0.94
DmtoKKpi0.90

Message from Peter:

Are you sure that it's not because some of your jobs crashed? 
You see different effects in D and Dbar.
Eh, are they all on lnx303?  Then there's probably something wrong
with the node... sometimes the jobs just don't run and the batch
queue kills them for lasting too long.  Usually in this case I'd
just fully rerun those three channels (i.e. including cleog).

Check :

cleog-SingleDmtoKspipipi-1.txtlnx311
cleog-SingleD0BtoKpipi0-5.txtlnx324
cleog-SingleD0BtoKpipi0-6.txtlnx65110
pass2-SingleD0BtoKpipi0-5.txtlnx303
pass2-SingleD0BtoKpipi0-6.txtlnx303
pass2-SingleDmtoKspipipi-1.txtlnx323

Move on to the new sample.

Produce Latest Photos/Dalitz MC

Find the location of the right code

Email Peter for the production process … sent.

Message from Peter:

> > Could you tell me how to generate the MC with "UPDDALITZ" ?
What do you mean?
> >   Also, where is the place to look for the version of the photos?
> >
Do you mean 'how can I tell what version of PHOTOS is run in a job'? 
Look in the logfile from the cleog step; there is a lot of output when
PHOTOS is initialized and one of the things it prints is the code version.

From the output of cleog:

*******************************************************************************
*                                                                              
*                          ==========================                          
*                            PHOTOS, Version:  2. 0                            
*                            Released at:  16/11/93                            
*                          ==========================                          
*                                                                              
*                  PHOTOS QED Corrections in Particle Decays                   
*                                                                              
*         Monte Carlo Program - by E. Barberio, B. van Eijk and Z. Was         
*         From version 2.0 on - by E.B. and Z.W.                               
*                                                                              
*******************************************************************************

Message from Peter:

The directory is /home/ponyisi/cleog/scripts-summerconf, the script
is fire_up_jobs_new_photosnew.sh . This should give you both the new
photos and the updated Dalitz plot for Ks pi pi0.

Run on the code

Copy the script:

cp /home/ponyisi/cleog/scripts-summerconf/fire_up_jobs_new_photosnew.sh /home/xs32/work/CLEO/analysis/DHad/src/2.1/gen
cp /home/ponyisi/cleog/scripts-summerconf/simple-cleog-generic-array-photosnew.sh /home/xs32/work/CLEO/analysis/DHad/src/2.1/gen

The difference between the photos new and old one is:

cd /nfs/cor/an1/ponyisi/2007-4/
. local_setup

Ask Peter what's the correct way to use that, copy dir or re-build … sent.

Message from Peter:

You're probably ok leaving that directory where it is, but if you run
into problems you can do that.

Change the file name (email address).

Copy tag numbers, tag decay files, tcl file, mode list.

cp -r $dhad/src/2.0/gen/tag_numbers $dhad/src/2.1/gen/
cp -r $dhad/src/2.0/gen/tag_decfiles $dhad/src/2.1/gen/
cp $dhad/src/2.0/gen/genmc_anders.tcl $dhad/src/2.1/gen/
cp $dhad/src/2.0/gen/single_mode_list $dhad/src/2.1/gen/

Ran cleog on one channel … Error from CLEOG:

=> cleog-SingleDptoKspipi0-1.txt

no files matched glob pattern "/nfs/cleoc/mc1/RandomTriggerEvents/Links/data1/RandomTriggerEvents_*.pds"

Ask Peter … sent.

Peter pointed out:

nl: /home/xs32/work/CLEO/analysis/DHad/src/2.1/gen/runlist: No such file or directory

Copy that file:

cp $dhad/src/2.0/gen/runlist $dhad/src/2.1/gen/

OK now. PHOTOS, Version: 2.15

Run all CLEOG … OK.

Pass2 …

cp /home/ponyisi/cleog/scripts-summerconf/simple-pass2-generic-array.sh /home/xs32/work/CLEO/analysis/DHad/src/2.1/gen

Save as pass2-generic-array.sh

Change the submit job sh.

Change email address.

cp $dhad/src/2.0/gen/mcp2_anders.tcl $dhad/src/2.1/gen

Found leftover pds from pass2.

xs32_lnx570> pwd
/home/xs32/work/CLEO/analysis/DHad/dat/signal/2.1/cleog_0214_photosnew
xs32_lnx570> ll
total 1.6G
-rw-r--r--  1 xs32 cms 494M Feb 19 20:51 cleog_Single_D0B_to_Kpipipi_6.pds
-rw-r--r--  1 xs32 cms 612M Feb 19 18:04 cleog_Single_D0_to_Kpipi0_8.pds
-rw-r--r--  1 xs32 cms  58M Feb 19 18:06 cleog_Single_Dm_to_Kpipipi0_10.pds
-rw-r--r--  1 xs32 cms 161M Feb 19 20:56 cleog_Single_Dm_to_Kspi_3.pds
-rw-r--r--  1 xs32 cms 298M Feb 19 18:22 cleog_Single_Dp_to_Kspi_8.pds
-rw-r--r--  1 xs32 cms 209K Feb 19 18:56 cleog_Single_Dp_to_Kspipipi_5.pds

Fix the bad pass2 runs

cleog_Single_D0B_to_Kpipipi_6.pds

Check the cleog output log …

cd $dhad/dat/signal/2.1/log_c_0525_photosnew_p2_1104
tail -20 cleog-Single_D0B_to_Kpipipi-6.txt 
RandomModule Status:  Engine = RanecuEngine, seeds = 356693554, 1339856888

%% INFO-JobControl.SummaryModule:
***** Summary Info *****

Stream physics : 1
Stream beginrun : 1
Stream endrun : 1
Stream event : 9948
Stream startrun : 1
Processed 9952 stops.
tail -20 pass2-Single_D0B_to_Kpipipi-6.txt 
***** Summary Info *****

Stream physics : 1
Stream beginrun : 1
Stream endrun : 1
Stream event : 9948
Stream startrun : 1
Processed 9952 stops.

Looks OK.

Need to generate: 9948.

Re-do the pass2 with fix task ID = 6. … done.

The cleog pds file has been removed, fix done.

cleog_Single_D0_to_Kpipi0_8.pds

CLEOG:

tail -20 cleog-Single_D0_to_Kpipi0-8.txt 
## new death vertex info:
## interaction type  : 31
## geant Parent Track: 69
## position          : (-132.112471,179.447715,-115.464307)
## geant vtx number  : 0
%% ERROR-MCInfo.MCParticle: attempt to add fatal interaction vertex to particle that is already dead.
Particle:
60   1 gamma (-0.007845, 0.010756,-0.006829; 0.014962) ---- 42  52   2   43  1 61
1  0
previous decay vertex:
43   1   6 (-132.112471,179.447715,-115.464307) 8.823860e+05: gamma -->   e+
new decay vertex:
1   0  31 (-132.112471,179.447715,-115.464307) 8.823860e+05: gamma -->
suez.exe: /nfs/cleo3/Offline/rel/20050525_MCGEN/src/MCInfo/Class/MCDecayTree/MCParticle.cc:639: void MCParticle::addVertex(MCVertex*): Assertion `false' failed.

Pass2:

tail -20 pass2-Single_D0_to_Kpipi0-8.txt 
***** ERROR in HROUT : Current Directory must be a RZ file :
%% INFO-RandomModule.RandomModule:

RandomModule Status:  Engine = RanecuEngine, seeds = 399089476, 1571312518

%% INFO-JobControl.SummaryModule:
***** Summary Info *****

Stream physics : 1
Stream beginrun : 1
Stream event : 15844
Stream startrun : 1
Processed 15847 stops.

File size 40M smaller.

xs32_lnx570> ls *D0_to_Kpipi0_* -lh
-rw-r--r--  1 xs32 cms 196M Feb 19 23:55 pass2_Single_D0_to_Kpipi0_10.pds
-rw-r--r--  1 xs32 cms 194M Feb 20 00:14 pass2_Single_D0_to_Kpipi0_1.pds
-rw-r--r--  1 xs32 cms 192M Feb 20 00:43 pass2_Single_D0_to_Kpipi0_2.pds
-rw-r--r--  1 xs32 cms 191M Feb 20 00:42 pass2_Single_D0_to_Kpipi0_3.pds
-rw-r--r--  1 xs32 cms 192M Feb 19 23:54 pass2_Single_D0_to_Kpipi0_4.pds
-rw-r--r--  1 xs32 cms 193M Feb 19 23:52 pass2_Single_D0_to_Kpipi0_5.pds
-rw-r--r--  1 xs32 cms 194M Feb 20 00:55 pass2_Single_D0_to_Kpipi0_6.pds
-rw-r--r--  1 xs32 cms 193M Feb 20 00:46 pass2_Single_D0_to_Kpipi0_7.pds
-rw-r--r--  1 xs32 cms 155M Feb 20 00:00 pass2_Single_D0_to_Kpipi0_8.pds
-rw-r--r--  1 xs32 cms 193M Feb 19 23:54 pass2_Single_D0_to_Kpipi0_9.pds

Consult Peter for generating single cleog, efficiency… sent.

Message from Peter:

If the cleog output is still there it means the *pass2* job failed. 
What is the pass2 log?

Sent reply.

Message from Peter:

Well, in this case it looks like the pass2
file should have been created (though with fewer events than you
requested); you should check the actual pass2 pds files.

Sent reply.

Find the supposed producing number: 19735

fix_mc_job.sh … done.

Check the tail:

tail -20 cleog-Single_D0_to_Kpipi0-8.txt 

Got the wrong number! Need to be 8!

Re-submit job with Task ID = 8. … done.

tail -20 cleog-Single_D0_to_Kpipi0-8.txt  ... OK.

Run pass2 …

Fix the rest

cleog_Single_Dm_to_Kpipipi0_10.pds
cleog_Single_Dm_to_Kspi_3.pds
cleog_Single_Dp_to_Kspi_8.pds
cleog_Single_Dp_to_Kspipipi_5.pds
  • Single_Dm_to_Kpipipi0_10
    tail -20 cleog-Single_Dm_to_Kpipipi0-10.txt 
    Floating point exception(core dumped) $
    

    Need to redo cleog:

    1. Change fix_mc_job.sh

      DOCLEOG=1, DOPASS2=0, DONTUPLE=0, FIXTASKID=10

    2. Change fix_single_mode_list

      Use Single_Dm_to_Kpipipi0

    3. Run script

      . ./fix_mc_job.sh … done.

    4. Pass2 …
      tail -20 cleog-Single_Dm_to_Kpipipi0-10.txt  OK. 
      
      dhad gen pass2 Single_Dm_to_Kpipipi0 task 10 ... OK. 
      
  • cleog_Single_Dm_to_Kspi_3
    tail -20 cleog-Single_Dm_to_Kspi-3.txt 
    
    STOP BIMPCT statement executed
    

    Re-run cleog:

    dhad gen cleog Single_Dm_to_Kspi task 3  ... OK. 
    

    pass2:

    dhad gen pass2 Single_Dm_to_Kspi task 3  ... done.
    
  • The rest two
    tail -20 cleog-Single_Dp_to_Kspi-8.txt 
    dhad gen cleog Single_Dp_to_Kspi task 8 ... OK
    dhad gen pass2 Single_Dp_to_Kspi task 8 ... done. 
    
    tail -20 cleog-Single_Dp_to_Kspipipi-5.txt 
    File FLUKAAF.DAT not found 
    
    dhad gen cleog Single_Dp_to_Kspipipi task 5 ... OK
    dhad gen pass2 Single_Dp_to_Kspipipi task 5 ... done.  
    

Run Dtuple

cp /home/ponyisi/cleog/scripts-summerconf/simple-dtuple-generic-array.sh $dhad/src/2.1/gen

Save as dtuple-generic-array.sh.

Change email, add mkdir for tmp, run … done.

Compare yields with photos 2.15

Link the root files.

cd $dhad/7.06/dat/signal
ln -s /home/xs32/work/CLEO/analysis/DHad/dat/signal/2.1/dtuple_c_0525_photosnew_p2_1104/ regular3

Extract yields.

dhad yield regular3 -t s ... done. 

Fit … done.

dhad fit regular3 -t s --qsub

log-2009-02-24 15:45:42

dhad table compare yields regular3 -t s
Modesignal diff(%)
D0toKpi0.31
D0BtoKpi-0.19
D0toKpipi00.06
D0BtoKpipi0-0.69
D0toKpipipi0.56
D0BtoKpipipi0.27
DptoKpipi-0.57
DmtoKpipi0.31
DptoKpipipi00.46
DmtoKpipipi0-9.51
DptoKspi-0.87
DmtoKspi-0.90
DptoKspipi0-0.42
DmtoKspipi00.44
DptoKspipipi1.63
DmtoKspipipi-0.61
DptoKKpi-0.34
DmtoKKpi1.56

Check the source of three modes

DmtoKpipipi0, DptoKspipipi, DmtoKKpi

  1. DmtoKpipipi0 => fixed.

    CLEOg:

    tail -20 $dhad/dat/signal/2.1/log_c_0525_photosnew_p2_1104/cleog-Single_Dm_to_Kpipipi0-1.txt
    
    Job IDProcessedStream event
    173427338
    273427338
    300

    Need tool:

    dhad check cleog Single_Dm_to_Kpipipi0 
    
    JobIDProcessedStream event
    173427338
    273427338
    3N/AN/A
    473427338
    573427338
    673427338
    773427338
    873427338
    973427338
    1073427338

    Rerun cleog:

    dhad gen cleog Single_Dm_to_Kpipipi0 task 3 ... done. 
    

    Check log:

    dhad check cleog Single_Dm_to_Kpipipi0  ...OK now. 
    

    Ran pass2:

    dhad gen pass2 Single_Dm_to_Kpipipi0 task 3 ... done.
    

    Check log:

    dhad check pass2 Single_Dm_to_Kpipipi0  ... OK.
    

    Run dtuple …

    dhad gen dtuple Single_Dm_to_Kpipipi0  ... done. 
    

    Extact yield:

    dhad yield regular3 -t s -m 201
    

    Fit … done.

    dhad fit regular3 -t s -m 201 --qsub
    

    log-2009-02-25 13:47:19

    Compare again:

    dhad table compare yields regular3 -t s
    
    Modesignal diff(%)
    D0toKpi0.31
    D0BtoKpi-0.19
    D0toKpipi00.06
    D0BtoKpipi0-0.69
    D0toKpipipi0.56
    D0BtoKpipipi0.27
    DptoKpipi-0.57
    DmtoKpipi0.31
    DptoKpipipi00.42
    DmtoKpipipi00.48
    DptoKspi-0.87
    DmtoKspi-0.90
    DptoKspipi0-0.42
    DmtoKspipi00.44
    DptoKspipipi1.63
    DmtoKspipipi-0.61
    DptoKKpi-0.34
    DmtoKKpi1.56

    Fixed the difference.

    http://www.lepp.cornell.edu/~xs32/private/DHad/7.06/tables/compare_yields_regular3.html

    Create the DHad Meeting page … https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/DHadGroupMeeting

  2. DptoKspipipi
    dhad check cleog Single_Dp_to_Kspipipi
    
    JobIDProcessedStream event
    139353931
    239353931
    339353931
    439353931
    539353931
    639353931
    739353931
    839353931
    939353931
    1039353931

    check pass2:

    tail -20 $dhad/dat/signal/2.1/log_c_0525_photosnew_p2_1104/pass2-Single_Dp_to_Kspipipi-1.txt
    
    dhad check pass2 Single_Dp_to_Kspipipi
    
    JobIDProcessedStream event
    139353931
    239353931
    339353931
    439353931
    539353931
    639353931
    739353931
    839353931
    939353931
    1039353931

    Looks OK, what might be the reason?

  3. DmtoKKpi
    dhad check cleog Single_Dm_to_KKpi
    
    JobIDProcessedStream event
    120042000
    220042000
    320042000
    420042000
    520042000
    620042000
    720042000
    820042000
    920042000
    1020042000
    dhad check pass2 Single_Dm_to_KKpi
    
    JobIDProcessedStream event
    120042000
    220042000
    320042000
    420042000
    520042000
    620042000
    720042000
    820042000
    920042000
    1020042000

    Look OK. Need to discuss.

Add error to the difference

Refer to table double_generic_eff

dhad-2.1 table compare yields signal regular3

Where σ = (a-b)/σb

Also add the original columns OK. => Table link.

Add more digits to the table

dhad-2.1 table compare yields signal regular3 --set rnd=0.01

fix in the tabletools.py:

c_err  = a_val/float(b_err)

Now, one digit is OK.

Update the table with two digits. => Signal Table

Update the data compare table:=> Data Table

dhad-2.1 table compare yields data regular3 --set rnd=0.001

Generate double tag signal MC (2.1) [%]

Setup env

cd $dhad/src/2.1/gen
cp ~ponyisi/cleog/scripts-summerconf-photosint/double_mode_list . 
dhad-2.1 gen cleog Single_Dm_to_Kspi task 1 --test OK.
dhad-2.1 gen cleog Double_Dp_to_Kspipipi__Dm_to_KKpi task 1 --test OK.

Test 10 events. (cleog-generic-array.sh)

dhad-2.1 gen cleog Double_Dp_to_Kspipipi__Dm_to_KKpi task 1 OK.

Use the standard number of events… OK.

dhad-2.1 gen cleog Double_Dp_to_Kspipipi__Dm_to_KKpi task 1-10 --test  OK.

Generate the events

dhad-2.1 gen cleog Double_Dp_to_Kspipipi__Dm_to_KKpi task 1-10 
dhad-2.1 check cleog Double_Dp_to_Kspipipi__Dm_to_KKpi OK.
dhad-2.1 gen pass2 Double_Dp_to_Kspipipi__Dm_to_KKpi task 1-5  --test OK.
dhad-2.1 gen pass2 Double_Dp_to_Kspipipi__Dm_to_KKpi task 1-10
dhad-2.1 check pass2 Double_Dp_to_Kspipipi__Dm_to_KKpi OK.
dhad-2.1 gen dtuple Double_Dp_to_Kspipipi__Dm_to_KKpi
dhad-2.1 check dtuple Double_Dp_to_Kspipipi__Dm_to_KKpi OK.

Process 281 data (2.1)

list the needed data set

ls /home/xs32/work/CLEO/analysis/DHad/7.06/dat/data/
data31_dskim_evtstore.root
data32_dskim_evtstore.root 
data33_dskim_evtstore.root 
data35_dskim_evtstore.root 
data36_dskim_evtstore.root 
data37_dskim_evtstore_1.root
data37_dskim_evtstore.root 

data 31

cp $dhad/test/script/test_data31.sh $dhad/src/2.1/gen

Edit it according to dtuple-generic-array.sh.

Test run for 10 evts … OK.

Run dtuple-data31.sh. … done.

Comment out the WIDEKS=1, save them in the dir wideks.

Then re-run , submitted …

data 32

cp /nfs/cor/user/ponyisi/daf9/simple-dtuple-data32.sh $dhad/src/2.1/gen

Edit according to dtuple-data31.sh and save as dtuple-data32.sh .

Submit … OK.

data 33

cp /nfs/cor/user/ponyisi/daf9/simple-dtuple-data33.sh $dhad/src/2.1/gen

Copy 32 to 33, ask Peter about the local setup duplication … OK to comment out.

Submit …

data 35

cp /nfs/cor/user/ponyisi/daf9/simple-dtuple-data35.sh $dhad/src/2.1/gen

Copy 33 to 35.

submit …

data 36

Copy 35 to 36, edit.

cp /nfs/cor/user/ponyisi/daf9/simple-dtuple-data36.sh $dhad/src/2.1/gen

submit…

data 37

Copy 36 to 37, edit.

cp /nfs/cor/user/ponyisi/daf9/simple-dtuple-data37.sh $dhad/src/2.1/gen

submit …

Extract yields

Link the root files to 7.06/dat/data/regular3

dhad yield regular3 -t d -m 0  OK.
dhad yield regular3 -t d --qsub ... done. 

log-2009-02-26 09:00:18

Fit data

Set up the 2.1 env:

cd /home/xs32/work/CLEO/analysis/DHad/src/2.1
cvs co -r V02-01-00 -d python dhad/src/python
. /home/xs32/work/CLEO/analysis/DHad/src/2.1/bash/dhadenv.sh
dhad fit regular3 -t d -m 0 Not working!

Use the script/python to cover the head of src/python :

cvs co -d python dhad/src/python
cp /home/xs32/work/CLEO/analysis/DHad/script/python/* /home/xs32/work/CLEO/analysis/DHad/src/2.1/python
cvs up
dhad fit regular3 -t d -m 0 OK.

Check in the code with tag V02-01-01 OK.

Edit the qsub with one line:

. /home/xs32/work/CLEO/analysis/DHad/src/2.1/bash/dhadenv.sh

Try one mode:

dhad fit regular3 -t d  -m 0 --qsub

log-2009-03-03 14:18:50

Not work.

Add one line : export ROOTLIB=$ROOTSYS/lib

dhad fit regular3 -t d  -m 0 --qsub

log-2009-03-03 14:24:41 Same problem.

Use the full setup

dhad fit regular3 -t d  -m 0 --qsub

log-2009-03-03 15:35:00 same problem.

Try signal fit:

dhad fit regular3 -t s -m 0 --qsub

log-2009-03-03 15:41:39 same problem. This should be caused by the script.

Use the script/python code:

dhad-2.1 fit regular3 -t d  -m 0 --qsub

log-2009-03-03 16:26:33

For other modes:

setdhad
dhadrel 2.1
dhad-2.1 fit regular3 -t d -m 1   OK.
dhad-2.1 fit regular3 -t d -m 1 --qsub

log-2009-03-04 11:27:13 working.

dhad-2.1 fit regular3 -t d -m 3 --qsub

log-2009-03-04 11:30:02

dhad-2.1 fit regular3 -t d -m 200 --qsub

log-2009-03-04 11:30:52

dhad-2.1 fit regular3 -t d -m 201 --qsub

log-2009-03-04 11:31:06

dhad-2.1 fit regular3 -t d -m 202 --qsub

log-2009-03-04 11:31:17

dhad-2.1 fit regular3 -t d -m 203 --qsub

log-2009-03-04 11:31:28

dhad-2.1 fit regular3 -t d -m 204 --qsub

log-2009-03-04 11:31:37

dhad-2.1 fit regular3 -t d -m 205 --qsub

log-2009-03-04 11:31:51

Compare yields

dhadrel 2.1
dhad-2 table compare yields regular3 -t d -m 0 

Need to finish the other modes. OK.

dhad-2.1 table compare yields data regular3

Table Link

Process 818 data (2.2) [88%]

First Try on data 43,44,45,46

https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/Private/SW/DatasetSummary

data 43,44,45,46

  1. Data 43 Create dtuple-data43.sh
    qsub dtuple-data43.sh ... 
    
  2. Data 44
    qsub dtuple-data44.sh ...
    
  3. Data 45
    qsub dtuple-data45.sh ...
    
  4. Data 46
    qsub dtuple-data46.sh ...
    

Error :

%% SYSTEM-Interpreter.TclInterpreter: Tcl_Eval error: can't read "preliminaryPass2": no such variable

Save the output => data43.txt

Ask Peter … sent.

Message from Peter:

You must edit $env(USER_SRC)/HadronicDNtupleProc/Test/dataselection.tcl to include a
section for 'data43_dskim_evtstore'.

Set up src 2.2

  1. Branch out SRC python B02-01
    cd /home/xs32/work/CLEO/analysis/DHad/src/2.1/python
    cvs tag V02-01-02
    cvs tag -b -r V02-01-02 B02-01
    cvs up -r B02-01 OK.
    
  2. Create src/2.2 based on V02-01-02
    cd $dhad/src
    cvs co -d 2.2 -r V02-01-02 dhad/src 
    cp ../2.1/* -r .
    cd python 
    cvs tag -r V02-01-02 -b B02-02
    cvs up -r B02-02
    
  3. Check out the cleo code
    cd $dhad/src/2.2/cleo
    export CVSROOT=/nfs/cleo3/cvsroot 
    cleo3cvs co -rv02_00_04  MCCCTagger
    cleo3cvs co -rponyisi060929 HadronicDNtupleProc 
    cleo3cvs co CWNFramework 
    

    Log on lnx134

    setdhad 2.2
    cd $dhad/src/2.2/cleo
    c3make ... OK.
    
  4. Test run on data31

    Edit dtuple-data31.sh

    mkdir /home/xs32/work/CLEO/analysis/DHad/dat/data/$rel
    cd $dhad/src/2.2/gen
    . dtuple-data31.sh 
    

    Error:

    %% ERROR-DynamicLoader.DLSharedObjectHandler: /nfs/cleo3/Offline/rel/20050417_FULL/other_sources/lib/Linux/g++
    /libXrdPosix.so: undefined symbol: XrdSecGetProtocol                                                          
    

    Compile again … Use the previous env setup… OK

Second Try : Encounter the coredump

  1. Edit dataselection.tcl

    Compare with the new one

    cd src/2.2/cleo
    mkdir new
    cd new
    export CVSROOT=/nfs/cleo3/cvsroot 
    cleo3cvs co  HadronicDNtupleProc 
    

    No data43 info in the tcl.

    Add section in the tcl:

    if { ( $env(INPUTDATA) == "data43_dskim_evtstore" ) } {
        set skim yes
        set preliminaryPass2 no
        set millionMC no
        set mc no
        set use_setup_analysis yes
        module sel EventStoreModule
        eventstore in 20050429 dtag all dataset data43
    }
    

    Ask Peter for comfirm with the date … sent.

    Message from Peter:

    Don't know if you should keep the same timestamp - in fact, probably
    not, I don't think the data was even collected then?
    
  2. Try it for 1 event

    Edit the data43 sh file

    Edit the tcl file for one event. Error:

     %% SYSTEM-Module: added command "eventstore" from EventStoreModule
     %% SYSTEM-MySQLQuerier.MySQLQuerier: Connecting to EventStore@lnx150.lns.cornell.edu
     %% ERROR-EventStoreModuleBase.EventStoreModuleBase: no runs found in required range
    
  3. Use Dskim 20070822

    From page:

    https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/Private/SW/FederationDetails

    Change the tcl.

    eventstore in 20070822 dtag all dataset data43
    

    OK. Got output.

    Reset the tcl to run for all events.

    qsub dtuple-data43.sh 
    

    Notice error in the data31 (dtuple-data31.txt)

    Suez.setup_analysis_command...> error "setup_analysis: unknown option: $subCommand"
    
    %% ERROR-Processor.RunEventNumberProc: No data of type "DBRunHeader" "" "" in Record beginrun 
    Please add a Source or Producer to your job which can deliver this data.
    

    Try to checkout:

    cd src/2.2/cleo
    export CVSROOT=/nfs/cleo3/cvsroot 
    cleo3cvs co DBRunHeaderProd
    c3make 
    

    Still has the same problem.

    Ask Peter … sent.

    Re-check the previous data 281/pb process, it had the same error message in the log file.

    Sent to Peter.

    Finished processing the data43 => dtuple-data43.txt

    Notice lines:

     >> Wed Jun  3 15:48:45 2009 Run:  221242 Event:    4599 Stop: event << 
     %% NOTICE-Processor.HadronicDNtupleProc: Hard limit on number of DD candidates exceeded
    
     >> Wed Jun  3 16:12:07 2009 Run:  221242 Event:    6470 Stop: event << 
     %% NOTICE-Processor.HadronicDNtupleProc: To many missing masses
     %% NOTICE-Processor.HadronicDNtupleProc: To many missing masses
    

    No summary info at the bottom either.

    Post on DHad HN … sent.

    Message from Anders:

     Have you looked at why this message gets produced? It means
    that you try to produce to many missing mass candidates which
    are used for tracking efficiency studies etc. I would have hoped that
    the code would have handled this gracefully, but there seems to
    be some problem as you don't get any further output. Do you get
    any error output from the job?
    

    Respond with the core dump.

    Message from Anders:

    are they from the jobs you submitted? (e.g. by looking at the timestamps?)
    If so can you get a trace back and see where it failed.
    

    Check the date stamp.

    ls -l /home/xs32/work/CLEO/analysis/DHad/dat/data/2.2/
    Jun  3 15:47 data43_dskim_evtstore.root
    Jun  4 15:26 dtuple-data43.txt  => this time is coused by the latest data44-46 run. Resubmited. 
    
    ls -l /home/xs32/work/CLEO/analysis/DHad/src/2.2/cleo/
    Jun  3 17:02 core.2694
    Jun  3 17:02 core.2718
    Jun  3 17:02 core.2738
    

    Open another section for tracing the core dump …

  4. Run on data44-46

    Add sections in the tcl file.

    Edit dtuple-data44.sh, 46, 46. Remember to change the log file!!

    qsub dtuple-data44.sh 
    qsub dtuple-data45.sh 
    qsub dtuple-data46.sh 
    

    Data44 finished, but the log file has the same error at the end => dtuple-data44.txt

    Trace the coredump.

    Ask Peter about running the 818 data… later…

Trace the coredump: CPU time limit exceeded

Start from page: http://cs.baylor.edu/~donahoo/tools/gdb/tutorial.html

cd /home/xs32/work/CLEO/analysis/DHad/src/2.2/cleo/
gdb core.2694 
"/a/lnx112/nfs/cor/user/xs32/local/work/CLEO/analysis/DHad
/src/2.2/cleo/core.2694": not in executable format: File format not recognized                                

Respond to Anders … sent.

Message from Anders:

to use the debugger you need to do:
gdb <path to the suez executable> core.2694

From the cleo3defs:

suez()        { ${C3_LIB}/bin/${OS_NAME}/${C3CXXTYPE}/suez "$@"; }

Redo:

setdhad 2.2
echo ${C3_LIB}/bin/${OS_NAME}/${C3CXXTYPE}
/nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++
gdb /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez core.2694 
This GDB was configured as "i386-redhat-linux-gnu"..."/a/lnx134/cdat/rd5/cleo3/Offline/rel/20050417_FULL/bin/Li
nux/g++/suez": not in executable format: File format not recognized                                           

Core was generated by `bash /nfs/sge/root/default/spool/lnx1622/job_scripts/1275637'.
Program terminated with signal 24, CPU time limit exceeded.
#0  0x00b3e7a2 in ?? ()
gdb /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez.exe core.2694
This GDB was configured as "i386-redhat-linux-gnu"...Using host libthread_db library "/lib/tls/libthread_db.so.
1".                                                                                                           
warning: core file may not match specified executable file.
Core was generated by `bash /nfs/sge/root/default/spool/lnx1622/job_scripts/1275637'.
Program terminated with signal 24, CPU time limit exceeded.
#0  0x00b3e7a2 in ?? ()
gdb /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez.exe core.2718
warning: core file may not match specified executable file.
Core was generated by `bash /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez -f HadronicDNtuple'.
Program terminated with signal 24, CPU time limit exceeded.
#0  0x00b3e7a2 in ?? ()
gdb /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez.exe core.2738
   
Core was generated by `/nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez.exe -f HadronicDNtupleP'.
Program terminated with signal 24, CPU time limit exceeded.
Cannot access memory at address 0x48ca6000
#0  0x002739ce in ?? ()

Respond to Anders … sent.

Message from Anders:

Are you actually running out of CPU?

Check the new finished data44 coredump.

gdb /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez.exe core.5477
warning: core file may not match specified executable file.
Core was generated by `bash /nfs/sge/root/default/spool/lnx65111/job_scripts/1277459'.
Program terminated with signal 24, CPU time limit exceeded.
#0  0x003e17a2 in ?? ()
gdb bash /nfs/sge/root/default/spool/lnx65111/job_scripts/1277459 core.5477
/nfs/sge/root/default/spool/lnx65111/job_scripts/1277459: No such file or directory.
gdb /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez core.5499
Core was generated by `bash /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez -f HadronicDNtuple'.
gdb bash /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez -f HadronicDNtuple core.5499
"/nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez" is not a core dump: File format not recognized
gdb bash core.5499
Core was generated by `bash /nfs/cleo3/Offline/rel/20050417_FULL/bin/Linux/g++/suez -f HadronicDNtuple'.
Program terminated with signal 24, CPU time limit exceeded.

warning: .dynamic section for "/lib/libtermcap.so.2" is not at the expected address

warning: difference appears to be caused by prelink, adjusting expectations

warning: .dynamic section for "/lib/libdl.so.2" is not at the expected address

warning: difference appears to be caused by prelink, adjusting expectations

warning: .dynamic section for "/lib/tls/libc.so.6" is not at the expected address

warning: difference appears to be caused by prelink, adjusting expectations

warning: .dynamic section for "/lib/ld-linux.so.2" is not at the expected address

warning: difference appears to be caused by prelink, adjusting expectations
Reading symbols from /lib/libtermcap.so.2...(no debugging symbols found)...done.
Loaded symbols for /lib/libtermcap.so.2
Reading symbols from /lib/libdl.so.2...(no debugging symbols found)...done.
Loaded symbols for /lib/libdl.so.2
Reading symbols from /lib/tls/libc.so.6...(no debugging symbols found)...done.
Loaded symbols for /lib/tls/libc.so.6
Reading symbols from /lib/ld-linux.so.2...
(no debugging symbols found)...done.
Loaded symbols for /lib/ld-linux.so.2
#0  0x003e17a2 in _dl_sysinfo_int80 () from /lib/ld-linux.so.2

Respond to Anders … sent.

Message from Anders:

can you send me the code that generated the printouts about
to many missing mass candidates.

Found in bottom of HadronicDNtupleProcmissingMass.cc:

   if (tuple.nmiss>=10000) {
      report( NOTICE, kFacilityString ) 
         <<"To many missing masses"<<endl;
      tuple.nmiss=9999;
   }

Message from Peter:

Depending on whether or not you want to redo the relevant systematics
studies, you may be able to comment out the calls to the missing mass
code - this saves a lot of CPU and disk space...

Message from Fan:

  Are you running your jobs on the data before DSkim or After DSkim? I
  guess such kind of problem maybe due to the array exceed. You can
  check some particular event region to see whether there are some
  strange events or not. Such as there are many soft photons in one
  event (over 300 or something like that). This kind of event sometime
  can cause serious multi-candidate problem and reach your array
  limit.

Trace the "To many missing masses" problem

Print debug info according to Anders:

$dhad/src/2.2/cleo/HadronicDNtupleProc/Class/HadronicDNtupleProcmissingMass.cc

Line 460:

cout << "Number of recoil cands:"<<RecoilList.size()<<endl;

Compile … OK.

cd $dhad/src/2.2/gen
. dtuple-data43.sh

Error:

 %% ERROR-DynamicLoader.DLSharedObjectHandler: /nfs/cleo3/Offline/rel/20050417_FULL/other_sources/lib/Linux/g++
/libXrdPosix.so: undefined symbol: XrdSecGetProtocol                                                          
 %% SYSTEM-Interpreter.TclInterpreter: Tcl_Eval error: ERROR: cannot load EventStoreModule.
 %% SYSTEM-Interpreter.TclInterpreter: Tcl_Eval error: ERROR: cannot load EventStoreModule.

Try again.

Same error.

Re-compile … still the same.

Comment out the print out … same error.

Ask on Prelimbugs HN … sent.

Sent service request.

Message from Dan:

https://hypernews.lepp.cornell.edu/HyperNews/get/Prelimbugs/26/1.html:

   echo $LD_LIBRARY_PATH

print?  That error usually means that you have a version of ROOT on
your LD_LIBRARY_PATH that is incompatible with the version of the
xrootd libraries EventStoreModule now uses.

We encourage people to *not* put ROOT on LD_LIBRARY_PATH; use the
"c3root" command to run root instead.

Remove the ROOT from the LD lib in setup.sh:

#export LD_LIBRARY_PATH=$C3PYTHONPATH/lib:$ROOTSYS/lib
export LD_LIBRARY_PATH=$C3PYTHONPATH/lib

Works.

Submit qsub … => data43.txt

Send to Anders … sent.

Edit from Anders :

Move the print out into line 676:

      report( NOTICE, kFacilityString ) << "Number of recoil cands:"<<RecoilList.size()<<endl;
      break;

Compile … OK.

Need to break down the data43 into two pieces.

Break down data43 into two

From page: https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/Private/SW/DatasetSummary

data setbeginendrunsfirst 1000 runs
data432208982232712373221898

In this page: https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/Private/SW/EventStore

eventstore in 20090402 runs 202126 203103

Edit the dataselection.tcl:

if { ( $env(INPUTDATA) == "data43_1_dskim_evtstore" ) } {
    set skim yes
    set preliminaryPass2 no
    set millionMC no
    set mc no
    set use_setup_analysis yes
    module sel EventStoreModule
    eventstore in 20070822 dtag all runs 220898 221898
}
if { ( $env(INPUTDATA) == "data43_2_dskim_evtstore" ) } {
    set skim yes
    set preliminaryPass2 no
    set millionMC no
    set mc no
    set use_setup_analysis yes
    module sel EventStoreModule
    eventstore in 20070822 dtag all runs 221899 223271
}

Create dtuple-data43_1.sh

Test run … OK.

qsub dtuple-data43_1.sh 

Log file => dtuple-data43_1.txt

Error happend at:

 >> Wed Jun 10 14:14:59 2009 Run:  221242 Event:    4599 Stop: event << 
 %% NOTICE-Processor.HadronicDNtupleProc: Hard limit on number of DD candidates exceeded

 >> Wed Jun 10 14:32:54 2009 Run:  221242 Event:    6470 Stop: event << 
 %% NOTICE-Processor.HadronicDNtupleProc: To many missing masses
 %% NOTICE-Processor.HadronicDNtupleProc: Number of recoil cands:12611

Find ways to get over this event …

Create dtuple-data43_2.sh

qsub dtuple-data43_2.sh 

Log file => dtuple-data43_2.txt

Just stopped at :

 >> Wed Jun 10 12:12:52 2009 Run:  221988 Event:   22938 Stop: event << 
 %% NOTICE-Processor.RunEventNumberProc: Run: 221988 Event: 22938

Break the data from here for another sub sample …

Read the log info about the DHad code …

Event Display the trouble event

(Run 221242 Event: 6470)

EventDisplay :

Refer to Suez talk.

Try the example.tcl:

suez -f example.tcl

Can open the display frame. But when click on the "continue", it showed this error message:

Fail to open /cdat/cleo/detector/event/daq/226000/226000/r226000.bin: No such file or directory
 %% WARNING-Processor.SpExtractShowerAttributesProc.ShowerAttributesList: Fail to open /cdat/cleo/detector/event/daq/226000
/226000/r226000.bin                                                                                                       
Please immediately send request to service@mail.lns.cornell.edu
Either a disk or system serving EventStore data is down

Display the run 221242:

In file HadronicDNtupleProc/Test/dataselection.tcl, eventstore 20070822.

Run the display.tcl, get the same error:

 %% INFO-Processor.SpExtractDRHitsProc: setting up CalibratedCathodeHits for display
 %% WARNING-UInt32StreamsFormat.FDIStream: Attempt to open /cdat/cleo/detector/event/daq/221200/221242/r221242.bin
 %% WARNING-UInt32StreamsFormat.FDIStream: Attempt to open /cdat/cleo/detector/event/daq/221200/221242/r221242.bin
Fail to open /cdat/cleo/detector/event/daq/221200/221242/r221242.bin: No such file or directory
 %% WARNING-Processor.SpExtractDRHitsProc.CalibratedDRHitList: Fail to open /cdat/cleo/detector/event/daq/221200/221242/r22
1242.bin                                                                                                                  
Please immediately send request to service@mail.lns.cornell.edu
Either a disk or system serving EventStore data is down    

Ask on Preliminary bugs HN … link.

Reply from Dan:

We only have a fraction of the raw data spinning on disk, usually the
data that were most recently processed through pass2.  Since we never
implemented on-demand staging in eventstore, any raw data has go be
pre-staged for use.  If there's a short list of runs you want to look
at, I can stage them fairly quickly; if there's a long list, you may
to negotiate with Anders and Souvik.  If you just want to look at
something, most runs in data48 are currently staged, along with the
first 100 runs of data47.

Ask for this particular Run 221242 Event: 6470 sent.

Reply from Dan:

I've initiated staging of 221242; you can

ls /cdat/sol404/disk1/objy/cleo/detector/event/daq/221200/221242/r221242.bin

to see when it's ready (I'd expect it within an hour or two unless
there are problems with the tape robot).

ls OK.

Suez> goto hlep
goto 221242 6470
go

Save the picture:

./log/2009/1013/evt_6470.png

Inplement code to bypass the trouble events

Re-run the test to identify trouble events

In HadronicDNtupleProc_missingMass.cc:

  if (tuple.nmiss>=10000) {
     report( NOTICE, kFacilityString ) 
        <<"To many missing masses"<<endl;
     tuple.nmiss=9999;
     report( NOTICE, kFacilityString ) << "Number of recoil cands:"<<RecoilList.size()<<endl;
     break;
  }

Test data43, run 221242

Create $dhad/src/2.2/gendtuple-data43-test.sh :

#$ -o /home/xs32/work/CLEO/analysis/DHad/dat/data/2.2/dtuple-data43-test.txt
export INPUTDATA=data43_test_dskim_evtstore
export FNAME=data43_test_dskim 

Add lines in dataselection.tcl

if { ( $env(INPUTDATA) == "data43_test_dskim_evtstore" ) } {
    set skim yes
    set preliminaryPass2 no
    set millionMC no
    set mc no
    set use_setup_analysis yes
    module sel EventStoreModule
    eventstore in 20070822 dtag all runs 221242 221242

Job complete: Log file dtuple-data43-test.txt

Job 1545587 (dtuple-data43-test.sh) Complete
 User             = xs32
 Queue            = all.q@lnx187.lns.cornell.edu
 Host             = lnx187.lns.cornell.edu
 Start Time       = 10/13/2009 10:44:11
 End Time         = 10/13/2009 11:12:41
 User Time        = 00:27:09
 System Time      = 00:00:08
 Wallclock Time   = 00:28:30
 CPU              = 00:27:17
 Max vmem         = 1.798G
 Exit Status      = 0

Event 794:

./log/2009/1013/evt_794.png

In the log file:

 >> Tue Oct 13 10:46:43 2009 Run:  221242 Event:    4599 Stop: event << 
%% NOTICE-Processor.HadronicDNtupleProc: Hard limit on number of DD candidates exceeded

Event 4599:

./log/2009/1013/evt_4599.png

Fix the Hard limit on number of DD candidates

Change the code:

  if( tuple.nddcand >= MAXNDDCAND )
  {
     report( NOTICE, kFacilityString ) 
 << "Hard limit on number of DD candidates exceeded" << endl;
 continue; // added by xs32
  }

Compile …

sl lc
setdhad 2.2
cd $dhad/src/2.2/cleo
c3make

Error message:

/home/xs32/work/CLEO/analysis/DHad/src/2.2/cleo/HadronicDNtupleProc/Class/HadronicDNtupleProc.cc: In
   member function `virtual ActionBase::ActionResult 
   HadronicDNtupleProc::event(Frame&)':
/home/xs32/work/CLEO/analysis/DHad/src/2.2/cleo/HadronicDNtupleProc/Class/HadronicDNtupleProc.cc:921: continue
   statement not within a loop
gmake[1]: *** [HadronicDNtupleProc.o] Error 1

Read basics of event loop in CLEO101.

Read the example code : MyFirstProc.cc

Read Writing a Processor. In slide:

./log/2009/1013/actionbase.png

Test on the kFalaiure and then ask …

  if( tuple.nddcand >= MAXNDDCAND )
  {
     report( NOTICE, kFacilityString ) 
 << "Hard limit on number of DD candidates exceeded" << endl;
     return ActionBase::kFailed; // added by xs32
  }

Compile … OK.

Clear old log file:

rm /home/xs32/work/CLEO/analysis/DHad/dat/data/2.2/dtuple-data43-test.txt

Test again:

cd $dhad/src/2.2/gen
qsub dtuple-data43-test.sh

Save log file: dtuple-data43-test.txt

Job summary:

Job 1546917 (dtuple-data43-test.sh) Complete
User             = xs32
Queue            = all.q@lnx189.lns.cornell.edu
Host             = lnx189.lns.cornell.edu
Start Time       = 10/14/2009 08:10:06
End Time         = 10/14/2009 08:29:40
User Time        = 00:19:18
System Time      = 00:00:03
Wallclock Time   = 00:19:34
CPU              = 00:19:21
Max vmem         = 1.794G
Exit Status      = 0

Time spend on event 4599:

TestJob startJob endDuration
FirstOct 13 10:46:43 2009Oct 13 11:10:34 20090@ 23' 51"
SecondOct 14 08:12:04 2009Oct 14 08:28:43 20090@ 16' 39"

Time spend on event 6470:

TestJob startJob endDuration
FirstOct 13 11:10:34Oct 13 11:12:410@ 2' 7"
SecondOct 14 08:28:43Oct 14 08:29:400@ 0' 57"

Ask Peter … sent.

Comments from David and Jim after the meeting: Use Bad Event List.

Get the Bad Event List
   if ( tuple.run == 221242 &&  tuple.event == 4599 ) {
     report( NOTICE, kFacilityString ) << "Bypassing ... Run: " << tuple.run
                               << "   Event: " << tuple.event << endl;
  
     return ActionBase::kFailed;
   }

OK.

Add event 6470 and continue … OK. => data43-test.txt

RunEvent
2212424599
2212426470
22129519061
22131122714

Test all the test runs in data43 (from 221243 to 223271)

Stuck on 221295 Event: 19061 … fixed.

Run: 221311 Event: 22714 : Spent 16 minutes DD limit … fixed.

Run: 221384 Event: 17807 : Spent 7 minutes DD limit … fixed.

Run: 221413 Event: 47962 : 6 min … not this one.

 >> Mon Oct 19 14:04:06 2009 Run:  221413 Event:   47962 Stop: event << 
 %% NOTICE-Processor.HadronicDNtupleProc: Bypassing ... Run: 221413 Event: 47962

 >> Mon Oct 19 14:09:20 2009 Run:  221413 Event:   54964 Stop: event << 
 %% NOTICE-Processor.HadronicDNtupleProc: To many missing masses

 ...

 %% NOTICE-Processor.HadronicDNtupleProc: Hard limit on number of DD candidates exceeded

 >> Mon Oct 19 14:10:01 2009 Run:  221413 Event:   59012 Stop: event << 
 %% NOTICE-Processor.RunEventNumberProc: Run: 221413 Event: 59012

Run: 221413 Event: 54964 : 6 min … fixed.

Run: 221438 Event: 37151 : 5 min … fixed.

Run: 221477 Event: 57536 : 6 min … fixed.

Run: 221492 Event: 62595 : 4 min … fixed.

Job aborted at Run: 221493 Event: Event: 47146

Continue from 221493 to 221898 …

Job aborted at Run: 221493 Event: 47852 :

 Job 1553357 (dtuple-data43-test.sh) Aborted
 Exit Status      = 152
 Signal           = XCPU
 User             = xs32
 Queue            = all.q@lnx187.lns.cornell.edu
 Host             = lnx187.lns.cornell.edu
 Start Time       = 10/20/2009 10:54:22
 End Time         = 10/20/2009 10:56:28
 CPU              = 00:01:50
 Max vmem         = 1.866G
failed assumedly after job because:
job 1553357.1 died through signal XCPU (24)

Run: 221493 Event: 47852 : Aborted … bypassed.

Got error:

ERROR: Suez caught "bad_alloc" exception:
"St9bad_alloc"

Bypass the previous run … Run: 221493 Event: 47146 …

Still aborted:

Job 1553374 (dtuple-data43-test.sh) Aborted
 Exit Status      = 152
 Signal           = XCPU
 User             = xs32
 Queue            = all.q@lnx187.lns.cornell.edu
 Host             = lnx187.lns.cornell.edu
 Start Time       = 10/20/2009 11:36:06
 End Time         = 10/20/2009 11:38:11
 CPU              = 00:01:49
 Max vmem         = 1.868G
failed assumedly after job because:
job 1553374.1 died through signal XCPU (24)     

Bypass the next event : Run: 221493 Event: 47853 … still aborted.

Shrink the range for the trouble event [47853, 60000] … OK.

[60001, 75516 ]

Run: 221493 Event: 60003, 60017, 60067, 60096, 60102, 60198, 60221, 60282, 60301, 60465, 60479, 60628, 60649, 60653, 60705(not filled) … aborted …

Need to read in a bad run list file to avoid compile every time … Done.

  • Start the test2 for the second half of data43 (221899 223271)

    Run: 221929 Event: 16197 : 8 min …

    Run: 221951 Event: 9909 : 5 min …

    Run: 221963 Event: 8017 : 16 min …

    Run: 221963 Event: 49934 : 4 min…

    Aborted at Run: 221988 Event: 22938 …

Avoid missing mass calc

Comment out the if statement.

   if ( tuple.run == 221242 &&  tuple.event == 4599 ||
  tuple.run == 221242 &&  tuple.event == 6470 ||
  tuple.run == 221295 &&  tuple.event == 19061 ||
  tuple.run == 221384 &&  tuple.event == 17807 ||
  tuple.run == 221413 &&  tuple.event == 54964 ||
  tuple.run == 221438 &&  tuple.event == 37151 ||
  tuple.run == 221477 &&  tuple.event == 57536 ||
  tuple.run == 221492 &&  tuple.event == 62595 ||
  tuple.run == 221493 &&  tuple.event == 47146 ||
  tuple.run == 221493 &&  tuple.event == 47852 ||
  tuple.run == 221493 &&  tuple.event == 60003 ||
  tuple.run == 221493 &&  tuple.event == 60017 ||
  tuple.run == 221493 &&  tuple.event == 60067 ||
  tuple.run == 221493 &&  tuple.event == 60096 ||
  tuple.run == 221493 &&  tuple.event == 60102 ||
  tuple.run == 221493 &&  tuple.event == 60198 ||
  tuple.run == 221493 &&  tuple.event == 60221 ||
  tuple.run == 221493 &&  tuple.event == 60282 ||
  tuple.run == 221493 &&  tuple.event == 60301 ||
  tuple.run == 221493 &&  tuple.event == 60465 ||
  tuple.run == 221493 &&  tuple.event == 60479 ||
  tuple.run == 221493 &&  tuple.event == 60628 ||
  tuple.run == 221493 &&  tuple.event == 60649 ||
  tuple.run == 221493 &&  tuple.event == 60653 ||
  //      tuple.run == 221493 &&  ( tuple.event <= 47853 || tuple.event >= 60000) || 
  tuple.run == 221493 &&  ( tuple.event <= 60653 || tuple.event >= 75516) || 
  tuple.run > 221493  || 

  tuple.run == 221311 &&  tuple.event == 22714 
  ) {
     report( NOTICE, kFacilityString ) << "Bypassing ... Run: " << tuple.run
                                 << " Event: " << tuple.event << endl;
  
     return ActionBase::kFailed;
   }

Comment out the missingMassObjects. … done.

Test on run 221242, change line in dataselection.tcl.

qsub dtuple-data43-test.sh 

Save log file data43.txt.

"segmentation violation" error.

Comment out additional lines…

/home/xs32/work/CLEO/analysis/DHad/src/2.2/cleo/HadronicDNtupleProc/Class/HadronicDNtupleProc.cc:581: `
   pionListTmp' undeclared (first use this function)

Move on …

/home/xs32/work/CLEO/analysis/DHad/src/2.2/cleo/HadronicDNtupleProc/Class/HadronicDNtupleProc.cc: In
   member function `virtual ActionBase::ActionResult 
   HadronicDNtupleProc::event(Frame&)':
/home/xs32/work/CLEO/analysis/DHad/src/2.2/cleo/HadronicDNtupleProc/Class/HadronicDNtupleProc.cc:585: `
   kaonListTmp' undeclared (first use this function)
/home/xs32/work/CLEO/analysis/DHad/src/2.2/cleo/HadronicDNtupleProc/Class/HadronicDNtupleProc.cc:585: (Each
   undeclared identifier is reported only once for each function it appears 
   in.)

Uncomment more … OK.

Test run again … log file data43.txt.1 same error.

Comment the delete lines … not work.

Need to understand more about the code …

Suggestion from Peter:

Comment out the makeMissingMassStage

//     makeMissingMassStage1(*iDListItem, *labMomentum);
//     makeMissingMassStage2(tagDecay, *labMomentum);

Test on run 221242: change line in dataselection.tcl.

qsub dtuple-data43-test.sh 

Log file data43.txt.

Still spent 1:30 on Run: 221242 Event: 6470.

Move on from run 221243 to 221494 … OK.

mv $dhad/dat/data/2.2/dtuple-data43-test.txt $dhad/log/2009/1103/data43.txt.1

data43.txt.1

Check the trouble time, add to EventsToSkip.txt

221295 19061 1hour 221493 61358 1:30

Test 221295 using the EventsToSkip … OK. data43.txt.2

>> Tue Nov  3 13:48:19 2009 Run:  221295 Event:   19061 Stop: event << 
%% NOTICE-Processor.SkipBadEventsProc: Skipping event

Proceed on the rest of runs.

Use Skip Bad Event

Read the /home/dskim/DSkimPackage/tcl/skimD_SkipBadEventsProc.tcl

proc sel SkipBadEventsProc
param SkipBadEventsProc EventsToSkip "/home/dskim/DSkimPackage/tcl/EventsToSkip.txt"

Add lines in loadHadronicDNtupleProc.tcl after the SkipBadRunsProc:

# Skip events which consum too much CPU time 
proc sel SkipBadEventsProc
param SkipBadEventsProc EventsToSkip "/home/xs32/work/CLEO/analysis/DHad/src/2.2/gen/EventsToSkip.txt"

Add events in EventsToSkip.txt… works.

Proceed on the rest of the data43

From 221495 to 221898 … done. data43-test-1.txt

qsub  dtuple-data43-test.sh 

Run: 221504 Event: 5468 … 16 min. Run: 221699 Event: 31675 … 8 min. Run: 221705 Event: 29678 … 3:10 ! Run: 221715 Event: 65978 … 23 min.

From 221899 to 223271 … done. data43-test-2.txt

qsub  dtuple-data43-test2.sh 

Run: 221929 Event: 16197 … 8 min. Run: 221963 Event: 7013 … 17 min. Run: 221963 Event: 38551 … 4 min. Run: 221988 Event: 23888 … 22 min. Run: 222014 Event: 47972 … 3 min Run: 222019 Event: 62955 … 5 min. Run: 222131 Event: 8240 … 1hour! Run: 222153 Event: 69636 … 8 min. Run: 222207 Event: 32078 … 6 min. Run: 222279 Event: 65673 … 1:36 hour! Run: 222515 Event: 60769 … 6 min. Run: 222544 Event: 42858 … 38 min. Run: 222844 Event: 46989 … 12 min. Run: 223051 Event: 28048 … 17 min.

Skip events > 15 min…

RunEventTime
22124264701:30
221295190611:00
221493613581:30
221705296783:10
221715659780:23
22196370130:17
221988238880:22
22213182401:00
222279656731:36
222544428580:38
223051280480:17

Clean the data43 … OK.

qsub dtuple-data43.sh 

Job finished:

Job 1579825 (dtuple-data43.sh) Complete
 User             = xs32
 Queue            = all.q@lnx187.lns.cornell.edu
 Host             = lnx187.lns.cornell.edu
 Start Time       = 11/04/2009 09:55:51
 End Time         = 11/05/2009 05:48:58
 User Time        = 16:01:55
 System Time      = 00:15:03
 Wallclock Time   = 19:53:07
 CPU              = 16:16:58
 Max vmem         = 1.848G
 Exit Status      = 0    

Output root files:

-rw-r--r--  1 xs32 cms 941M Nov  5 05:48 data43_dskim_evtstore_1.root
-rw-r--r--  1 xs32 cms 1.8G Nov  4 23:49 data43_dskim_evtstore.root

Avoid the memeory leak for data43

Edit the code HadronicDNtupleProc.cc:

677  //   *missingMassObjects.loosePi0List = TurnPhdPi0sIntoNavPi0ToGGs(loosePi0table, showtable); 
678  //  xs32: avoide the memory leak, only used for the missing mass calculation
1120 } // Added this xs32, ryd
RunEventTime
22124264701:30
221705296783:10

Test on run 221242,

Compile in 2.2 … OK.

Change the tcl file … OK.

Remove the event in the EventToSkip … OK.

qsub dtuple-data43-test.sh ...

Save log: data43-test.txt

Test on run 221705:

Change the tcl file … OK.

qsub dtuple-data43-test2.sh ...

Looks still wasting time on that … log: data43-test2.txt

RunEventTime (old)New
22124264701:301:30
221705296783:103:33

Put Time Statement inside code

Ask Peter for advice … sent.

Proceed on data44, 45, 46

qsub  dtuple-data44.sh 
qsub  dtuple-data45.sh 
qsub  dtuple-data46.sh 

Data 44 job died:

Job 1579399 (dtuple-data44.sh) Aborted
 Exit Status      = 152
 Signal           = XCPU
 User             = xs32
 Queue            = all.q@lnx187.lns.cornell.edu
 Host             = lnx187.lns.cornell.edu
 Start Time       = 11/04/2009 09:25:06
 End Time         = 11/05/2009 02:55:35
 CPU              = 14:58:04
 Max vmem         = 1.869G
failed assumedly after job because:
job 1579399.1 died through signal XCPU (24)    

Log file: dtuple-data44.txt

Output files:

-rw-r--r--  1 xs32 cms  59M Nov  5 01:20 data44_dskim_evtstore_1.root
-rw-r--r--  1 xs32 cms 1.8G Nov  5 01:06 data44_dskim_evtstore.root

Need code to get the table of time consuming…

Data 45 job finished:

Job 1579400 (dtuple-data45.sh) Complete
 User             = xs32
 Queue            = all.q@lnx189.lns.cornell.edu
 Host             = lnx189.lns.cornell.edu
 Start Time       = 11/04/2009 09:26:06
 End Time         = 11/04/2009 23:00:39
 User Time        = 12:33:38
 System Time      = 00:05:00
 Wallclock Time   = 13:34:33
 CPU              = 12:38:38
 Max vmem         = 1.743G
 Exit Status      = 0

Output files:

-rw-r--r--  1 xs32 cms 680M Nov  4 23:00 data45_dskim_evtstore_1.root
-rw-r--r--  1 xs32 cms 1.8G Nov  4 19:40 data45_dskim_evtstore.root

Data 46 job aborted:

Job 1579401 (dtuple-data46.sh) Aborted
 Exit Status      = 152
 Signal           = XCPU
 User             = xs32
 Queue            = all.q@lnx65107.lns.cornell.edu
 Host             = lnx65107.lns.cornell.edu
 Start Time       = 11/04/2009 09:26:35
 End Time         = 11/04/2009 20:46:50
 CPU              = 10:00:03
 Max vmem         = 1.863G
failed assumedly after job because:
job 1579401.1 died through signal XCPU (24)

Log file: dtuple-data46.txt

Output files:

-rw-r--r--  1 xs32 cms 585M Nov  4 20:42 data46_dskim_evtstore_1.root
-rw-r--r--  1 xs32 cms 1.8G Nov  4 17:38 data46_dskim_evtstore.root

The possible solutions are :

  1. Break the job into two
  2. Skip the bad events (manual or with tools)

Process data44

Skip the bad events which take more than 20 minutes.

dhad-2.2 table time data44
RunEventDuration
223560213961:16:05
223763839530:27:06
223979773170:26:26
224185324950:34:37
224239237800:29:34
224512534921:04:57
224597418170:28:44

It stopped at Run: 224597 Event: 46933, test this run …

Skip the 41817 .

qsub  dtuple-data44-test.sh 

Still stuck at this event… data44.txt

 >> Sat Nov 14 11:29:37 2009 Run:  224597 Event:   46933 Stop: event << 
 %% NOTICE-Processor.HadronicDNtupleProc: Hard limit on number of D candidates exceeded
 %% NOTICE-Processor.HadronicDNtupleProc: Hard limit on number of DD candidates exceeded

 >> Sat Nov 14 14:43:29 2009 Run:  224597 Stop: none << 
 %% INFO-JobControl.JobControl: Reached end of an active source
 %% INFO-JobControl.SourceManager: Removed source-in-use "eventstore"

This is a must skip event. Add in the EventsToSkip.txt. … OK.

Get the last run on the first root file: getMaxRun.C.

cd $dhad/dat/data/2.2
root -l 
.L getMaxRun.C
getMaxRun("data44_dskim_evtstore.root")
Max Run: 224556 

Use this run as the first run in the second job.

Find the starting run in DataSetSummary … 223279 - 226042

data44_1:  223279 224555
data44_2:  224556 226042

Submit jobs.

qsub  dtuple-data44_1.sh
qsub  dtuple-data44_2.sh

Check the second job log:

 >> Tue Nov 17 23:20:06 2009 Run:  225760 Event:    4270 Stop: event << 
 %% NOTICE-Processor.RunEventNumberProc: Run: 225760 Event: 4270
Fill: Switching to new file: /home/xs32/work/CLEO/analysis/DHad/dat/data/2.2/data44_2_dskim_evtstore_1.root

Need to start the third job from run 225760.

Other long time runs …

cd $dhad/dat/data/2.2
dhad-2.2 table time dtuple-data44_2.txt
RunEventDuration
224597469333:23:14
22465088361:22:46
224664482500:39:38
224668190002:30:54
224689705230:30:49
224765625306:51:46
224867442420:42:09
224898339860:24:11
22490789980:25:26
224909109780:22:10
225080699740:24:48
225519766390:50:34
225599573560:38:39
225665343353:03:24
225718927580:23:34
225764132221:36:38
225785152220:47:01
225878578502:16:28
225991199291:55:02
2260319630:28:26

Wait for finish then update the above table … done.

Update the EventsToSkip … done.

Update the tcl … done.

Update the sh … done.

Change the third file name:

mv  data44_2_dskim_evtstore_1.root data44_3_dskim_evtstore.root

Process data46

Skip the bad events which take more than 20 minutes.

dhad-2.2 table time data46
RunEventDuration
228307330061:06:27
228677465210:27:29
229695565040:29:09

The job stopped at Run: 229724 Event: 68512

Test on this run …

qsub  dtuple-data46-test.sh 

Log file data46.txt:

 >> Sat Nov 14 11:42:04 2009 Run:  229724 Event:   68512 Stop: event << 
 %% NOTICE-Processor.HadronicDNtupleProc: Hard limit on number of D candidates exceeded
 %% NOTICE-Processor.HadronicDNtupleProc: Hard limit on number of DD candidates exceeded

 >> Sat Nov 14 13:51:31 2009 Run:  229724 Event:   69349 Stop: event << 
 %% NOTICE-Processor.RunEventNumberProc: Run: 229724 Event: 69349

 >> Sat Nov 14 13:51:34 2009 Run:  229724 Event:   78793 Stop: event << 
 %% INFO-TrackManProd.TMData: The curvature of track 8 is 0. Setting the various radii to zero.

 >> Sat Nov 14 13:51:35 2009 Run:  229724 Stop: none << 
 %% INFO-JobControl.JobControl: Reached end of an active source
 %% INFO-JobControl.SourceManager: Removed source-in-use "eventstore"
Suez.loadHadronicDNtupleProc> exit

Spend 2 hours on this event. Skip it and the previous 3 events.

Break the job into two:

Get the last run number:

cd $dhad/dat/data/2.2
root -l 
.L getMaxRun.C
getMaxRun("data46_dskim_evtstore.root")
Max Run: 229385

First job: 228285 229384 Second job: 229385 230130

qsub dtuple-data46_1.sh
qsub dtuple-data46_2.sh

Check the first time table:

dhad-2.2 table time dtuple-data46_1.txt
RunEventDuration
228360296260:28:20
229228447930:30:28
229268405810:21:13
dhad-2.2 table time dtuple-data46_2.txt
RunEventDuration
229632752590:21:44
229653726300:20:42
229778691640:26:30
229800390240:47:52

Update EventsToSkip … done.

Re-process data43

Update the skip events … OK.

qsub dtuple-data43.sh 

Done.

Update the time table:

dhad-2.2 table time $dhad/dat/data/2.2/dtuple-data43.txt
RunEventDuration
221238261930:37:26

Update this event to EventToSkip.txt … OK.

From the log file:

 >> Tue Nov 17 01:02:30 2009 Run:  222627 Event:   29127 Stop: event << 
 %% NOTICE-Processor.RunEventNumberProc: Run: 222627 Event: 29127
Fill: Switching to new file: /home/xs32/work/CLEO/analysis/DHad/dat/data/2.2/data43_dskim_evtstore_1.root

Confirm the number by script:

cd $dhad/dat/data/2.2
root -l 
.L getMaxRun.C
getMaxRun("data43_dskim_evtstore.root")
Max Run: 222627

First job: 220898 222626 Second job: 222627 223271

Update the tcl file and sh file … OK.

Use them next time.

Change the file names:

mv data43_dskim_evtstore.root data43_1_dskim_evtstore.root
mv data43_dskim_evtstore_1.root data43_2_dskim_evtstore.root

Re-process data45

Get the last run number:

cd $dhad/dat/data/2.2
root -l 
.L getMaxRun.C
getMaxRun("data45_dskim_evtstore.root")
Max Run: 227841

Break job into two:

  1. 226429 227840
  2. 227841 228271

Edit the dataselection.tcl.

Get the bad events:

dhad-2.2 table time data45
RunEventDuration
226804436670:24:34
227041529210:30:03
227805775720:21:16
2278138370:35:04
228247288370:37:37

Update the EventToSkip… done.

qsub dtuple-data45_1.sh
qsub dtuple-data45_2.sh

Update the time table … done.

dhad-2.2 table time dtuple-data45_1.txt 
dhad-2.2 table time dtuple-data45_2.txt 

All skipped.

Remove the old root files:

rm data45_dskim_evtstore*.root 

Extract Yields

Link the root files to 7.06/dat/data/537ipb

cd $dhad/7.06/dat/data
ln -s $dhad/dat/data/2.2 537ipb
dhad-2.2 yield 537ipb -t d -m 0  OK. 

dhad-2.2 yield 537ipb -t d -m 0 –qsub

Not work:

 User             = xs32
 Queue            = all.q@lnx326.lns.cornell.edu
 Host             = lnx326.lns.cornell.edu
 Start Time       = 11/19/2009 08:43:26
 End Time         = 11/19/2009 08:43:26
 User Time        = 00:00:00
 System Time      = 00:00:00
 Wallclock Time   = 00:00:00
 CPU              = NA
 Max vmem         = NA
 Exit Status      = 127    

Check the log file:

/nfs/sge/root/default/spool/lnx326/job_scripts/1602642: line 36: dhad-2.2: command not found

Simple try:

dhad yield 537ipb -t d -m 1 –qsub

/nfs/sge/root/default/spool/lnx302/job_scripts/1602647: line 36: dhad: command not found

Add setdhad 2.2 in the sh file … OK.

dhad-2.2 yield 537ipb -t d -m 1 –qsub log-2009-11-19 10:26:54

dhad-2.2 yield 537ipb -t d -m 0 –qsub log-2009-11-19 10:31:11

dhad-2.2 yield 537ipb -t d -m 3 –qsub log-2009-11-19 10:31:36

dhad-2.2 yield 537ipb -t d -m 200 –qsub log-2009-11-19 10:31:43

dhad-2.2 yield 537ipb -t d -m 201 –qsub log-2009-11-19 10:31:48

dhad-2.2 yield 537ipb -t d -m 202 –qsub log-2009-11-19 10:31:55

dhad-2.2 yield 537ipb -t d -m 203 –qsub log-2009-11-19 10:32:01

dhad-2.2 yield 537ipb -t d -m 204 –qsub log-2009-11-19 10:32:07

dhad-2.2 yield 537ipb -t d -m 205 –qsub log-2009-11-19 10:32:13

Clean the duplicated files … done.

dhad yield 537ipb -t d -m 1 OK.

dhad-2.2 yield 537ipb -t d -m 0 –qsub log-2009-12-10 11:20:19

dhad-2.2 yield 537ipb -t d -m 1 –qsub log-2009-12-10 11:20:03

dhad-2.2 yield 537ipb -t d -m 3 –qsub log-2009-12-10 11:20:27

dhad-2.2 yield 537ipb -t d -m 200 –qsub log-2009-12-10 11:20:33

dhad-2.2 yield 537ipb -t d -m 201 –qsub log-2009-12-10 11:20:38

dhad-2.2 yield 537ipb -t d -m 202 –qsub log-2009-12-10 11:20:43

dhad-2.2 yield 537ipb -t d -m 203 –qsub log-2009-12-10 11:20:48

dhad-2.2 yield 537ipb -t d -m 204 –qsub log-2009-12-10 11:20:54

dhad-2.2 yield 537ipb -t d -m 205 –qsub log-2009-12-10 11:20:59

Do the Fits

setdhad 2.2
dhad-2.2 fit 537ipb  -t d -m 0 OK.

dhad-2.2 fit 537ipb -t d -m 0 –qsub log-2009-11-23 15:29:29

dhad-2.2 fit 537ipb -t d -m 1 –qsub log-2009-11-23 15:29:39

dhad-2.2 fit 537ipb -t d -m 3 –qsub log-2009-11-23 15:29:47

dhad-2.2 fit 537ipb -t d -m 200 –qsub log-2009-11-23 15:29:52

dhad-2.2 fit 537ipb -t d -m 201 –qsub log-2009-11-23 15:29:57

dhad-2.2 fit 537ipb -t d -m 202 –qsub log-2009-11-23 15:30:02

dhad-2.2 fit 537ipb -t d -m 203 –qsub log-2009-11-23 15:30:07

dhad-2.2 fit 537ipb -t d -m 204 –qsub log-2009-11-23 15:30:12

dhad-2.2 fit 537ipb -t d -m 205 –qsub log-2009-11-23 15:30:17

After the clean up for the duplicated events, do the fit again:

dhad-2.2 fit 537ipb -t d -m 0 –qsub log-2009-12-10 14:12:28

dhad-2.2 fit 537ipb -t d -m 1 –qsub log-2009-12-10 14:12:33

dhad-2.2 fit 537ipb -t d -m 3 –qsub log-2009-12-10 14:12:42

dhad-2.2 fit 537ipb -t d -m 200 –qsub log-2009-12-10 14:12:49

dhad-2.2 fit 537ipb -t d -m 201 –qsub log-2009-12-10 14:12:54

dhad-2.2 fit 537ipb -t d -m 202 –qsub log-2009-12-10 14:12:59

dhad-2.2 fit 537ipb -t d -m 203 –qsub log-2009-12-10 14:13:04

dhad-2.2 fit 537ipb -t d -m 204 –qsub log-2009-12-10 14:13:09

dhad-2.2 fit 537ipb -t d -m 205 –qsub log-2009-12-10 14:13:14

Store the fitting plots

dhad-2.2 web plots 537ipb -t d  

Extract yields for 818 data

Link the root files to 7.06/dat/data/537ipb

cd $dhad/7.06/dat/data
mkdir 818ipb
cd 818ipb

Fix the duplicated data45 files, changed the file names … done.

ln -s $dhad/dat/data/2.1/*.root . 
ln -s $dhad/dat/data/2.2/*.root . 
dhad-2.2 yield 818ipb -t d -m 0  OK.

dhad-2.2 yield 818ipb -t d -m 0 –qsub log-2009-12-09 11:44:32

dhad-2.2 yield 818ipb -t d -m 1 –qsub log-2009-12-09 11:46:55

dhad-2.2 yield 818ipb -t d -m 3 –qsub log-2009-12-09 11:47:17

dhad-2.2 yield 818ipb -t d -m 200 –qsub log-2009-12-09 11:50:59

dhad-2.2 yield 818ipb -t d -m 201 –qsub log-2009-12-09 11:51:28

dhad-2.2 yield 818ipb -t d -m 202 –qsub log-2009-12-09 11:51:45

dhad-2.2 yield 818ipb -t d -m 203 –qsub log-2009-12-09 11:51:53

dhad-2.2 yield 818ipb -t d -m 204 –qsub log-2009-12-09 11:51:58

dhad-2.2 yield 818ipb -t d -m 205 –qsub log-2009-12-09 11:52:07

Fitting for 818 data

dhad-2.2 fit 818ipb  -t d -m 0 ... OK.

dhad-2.2 fit 818ipb -t d -m 0 –qsub log-2009-12-09 15:06:18

dhad-2.2 fit 818ipb -t d -m 1 –qsub log-2009-12-09 15:07:05

dhad-2.2 fit 818ipb -t d -m 3 –qsub log-2009-12-09 15:07:14

dhad-2.2 fit 818ipb -t d -m 200 –qsub log-2009-12-09 15:07:21

dhad-2.2 fit 818ipb -t d -m 201 –qsub log-2009-12-09 15:07:26

dhad-2.2 fit 818ipb -t d -m 202 –qsub log-2009-12-09 15:07:31

dhad-2.2 fit 818ipb -t d -m 203 –qsub log-2009-12-09 15:07:36

dhad-2.2 fit 818ipb -t d -m 204 –qsub log-2009-12-09 15:07:41

dhad-2.2 fit 818ipb -t d -m 205 –qsub log-2009-12-09 15:07:46

Store fitting plots

dhad-2.2 web plots 818ipb -t d  

Compare the yields

Refer to 200607 report table:

CategoryY(281)/Y(56)
D0 ST5.00 ± 0.02
D+ ST5.02 ± 0.03
D0D0B DT5.00 ± 0.10
D+ D- DT5.06 ± 0.12
281/55.85.04

And the efficiency table:

CategoryY(281)/Y(56)Y(281)/Y(56) (old)
D all ST5.16 ± 0.085.00 ± 0.02
--5.02 ± 0.03
D0 D0B DT4.90 ± 0.115.00 ± 0.10
D+ D- DT4.83 ± 0.145.06 ± 0.12
281/55.85.045.04

Need to do the fits for 818. … done.

Compare the yields with 281 …

dhad-2.2 table compare yields data divide 281ipb 818ipb

Compare the 537 yields with 281 …

dhad-2.2 table compare yields data divide 281ipb 537ipb

Update for the error … OK.

Need to talk with David for about the 818 case where the new data was included…

Compare the fitting parameters

Compare one parameter first:

dhad-2.2 table compare parameter data sigmap1  281ipb 537ipb --set rnd=0.000001 ... OK

Compare the others … OK.

Process 281 data with eventstore 20070822 (2.2)

Data 31

Check the sh file dtuple-data31.sh … OK.

Change the dataselection.tcl :

if { ( $env(INPUTDATA) == "data31_dskim_evtstore" ) } {
#    eventstore in 20050429 dtag all dataset data31
   eventstore in 20070822 dtag all dataset data31

qsub dtuple-data31.sh … No output due to the later timestamp.

Stop all the jobs … OK.

Data 32

Change the file dtuple-data32.sh … OK.

Change the dataselection.tcl :

if { ( $env(INPUTDATA) == "data32_dskim_evtstore" ) } {
#    eventstore in 20050429 dtag all dataset data32
   eventstore in 20070822 dtag all dataset data32

Also change the data33, data35, data36, data37… OK.

qsub dtuple-data32.sh …

Data33

Change the file dtuple-data33.sh … OK.

qsub dtuple-data33.sh …

Data35

Change the file dtuple-data35.sh … OK.

qsub dtuple-data35.sh …

Data36

Change the file dtuple-data36.sh … OK.

qsub dtuple-data36.sh …

Data37

Change the file dtuple-data37.sh … OK.

qsub dtuple-data37.sh …

Prepare four-vector for CLEOG (2.3)

Generate

Start from dhad 2.1

cd $dhad/src
cp -r 2.1 2.3 
cd 2.3/python
cvs tag -r V02-01-02 -b B02-03
cvs up -r  B02-03
  1. Remove the rel in setup.sh
  2. Edit the DHad.py src=2.3
  3. Edit fixmcjob.sh

    Suggestion from Laura, check the HN from Indiana group.

    Found page: https://wiki.lepp.cornell.edu/lepp/bin/view/CLEO/Private/SW/GeneratingMCDecayTree

  4. Create gendecaytree.tcl
    set fileout $env(OUTDIR)/cleog_decaytree_$env(DECAYTITLE)_$env(BATCH).pds
    
  5. Edit cleog-generic-array.sh
    suez -f gen_decaytree.tcl
    
  6. Test for one mode for 10 events
    setdhad 2.3 
    

    Edit cleog-generic-array.sh:

    export NUMEVT=10
    

    Test

    dhad-2.3 gen cleog Single_Dm_to_Kspi task 1
    

    Error: => cleog.txt

    can't read "env(MCRUN)": no such variable 
    

    Change to RUNNUMBER. … => cleog.txt.1 (this file corrupts by a later call) stops.

    Ask on HN… sent.

    Message from Surik,

    Suggestion from Dan, delete the commented line … => cleog.txt.2

    can't read "env(USERDEC)": no such variable 
    

    Change the USERDEC to UDECAY … => cleog.txt.3

    Try large number …

    Edit cleog-generic-array.sh:

    export NUMEVT=`cat $SCRIPTDIR/tag_numbers/$1`
    

    Run … => cleog.txt.4 OK. Output pdf 5.5M.

    Need mass production … hold.

    Try to continue cleog in this release.

    Re-name the process as 'decaytree':

    Edit fixmcjob.sh.template.

    Create decaytree-generic-array.sh.

    Test 10 events.

    dhad-2.3 gen decaytree Single_Dm_to_Kspi task 1
    

    Save log file => decaytree.txt

    Continue cleog in this release.

  7. Try to remove the lines in the decay tree gen stage:
    file out $fileout "beginrun {}" "startrun {}" "endrun {}"  physics "event $treeMC"
    
    dhad-2.3 gen decaytree Single_Dm_to_Kspi task 2
    

    Save log => decaytree.txt.1

    No output.

    Remove the line: => decaytree.txt.2

    sink rm $fileout;
    

    Has output.

    Consult with Dan, the syntax is correct.

  8. Generate one normal number mode

    Edit decaytree-generic-array.sh:

    export NUMEVT=`cat $SCRIPTDIR/tag_numbers/$1`
    

    Restore gendecaytree.tcl:

    file out $fileout "beginrun {}" "startrun {}" "endrun {}"  physics "event $treeMC"
    
    dhad-2.3 gen decaytree Single_Dm_to_Kspi task 3
    

    Output 5.5M.

    Continue test in 9.03 … OK.

    Clean up files … done.

  9. Generate all modes
    dhad-2.3 gen decaytree single ... problem.
    

    Move to dhad-3.0.5 to deal with this.

Use four-vector for CLEOG

Edit the cleog-generic-array.sh.

Add $INDIR.

Create cleogdecaytree.tcl.

Test 10 events.

dhad-2.3 gen cleog Single_Dm_to_Kspi task 1

Save log file => cleog.txt

Same error.

Reply to Surik and HN …

Try the removed pds file task 2: => cleog.txt.1

dhad-2.3 gen cleog Single_Dm_to_Kspi task 2

Same error:

can't read "argArray(generatorProd)": no such element in array 

Fix the bug of cleogcommand.tcl by Dan:

~dsr/clego_command.tcl 

Edit line in cleogdecaytree.tcl :

#run_file $env(C3_SCRIPTS)/cleog_command.tcl
run_file /home/dsr/cleog_command.tcl

Test 10 events.

dhad-2.3 gen cleog Single_Dm_to_Kspi task 1

Other error => cleog.txt.2

unknown cleoc pass2 tag "6" requested

Try it in the 9.03 … OK.

So, use this only at 9.03 for now.

Close this section.

Leave 9.03 for continue cleog and pass2.

Debug in the BeamEnergyProxy (2.4)

Setup 2.4

Based on 2.1.

cd $dhad/src
cp -r 2.1 2.4 
cp /nfs/cor/an1/ponyisi/2007-4/local_setup /home/xs32/work/CLEO/analysis/DHad/src/2.4/cleo

Backup the file. localsetup.peter

Change local_setup to use the local build.

Change the cleog-generic-array.sh … OK.

Change the python code DHad.py for 2.4 … OK.

Change the setup.sh … OK.

Change the fixmcjob.sh.template …

Test for 10 events …

dhad-2.4  gen cleog Single_Dm_to_Kpipipi0 task 3 

OK. cleog.txt

Check out the CesrBeamEnergyPord

Look up the pkg versions for the CesrBeamEnergyProd

/nfs/cleo3/Offline/rel/20050525_MCGEN/pkg_versions

Found:

CesrBeamEnergyProd   Infrastructure   v01_00_02   proc

Check out the CesrBeamEnergyProd:

cd $dhad/src/2.4/cleo/
export CVSROOT=/nfs/cleo3/cvsroot
cleo3cvs co -r v01_00_02 CesrBeamEnergyProd
. local_setup
c3make 

Edit the BeamEnergyProxy.cc:

std::cout << "xs23_debug >>> beam energy:" << p_BeamEnergy.value() << std::endl;

Test for 10 events :

dhad-2.4  gen cleog Single_Dm_to_Kpipipi0 task 3 

cleog.txt.1 no debug info.

But it is using the CesrBeamEnergyProd:

CesrBeamEnergyProd :   /home/xs32/work/CLEO/analysis/DHad/src/2.4/cleo/build/Linux/shlib/CesrBeamEnergyProd.so_20050525_MCGEN

Suggestion from Anders: Add print statement in the beginning…

Try the report (INFO) …

/home/xs32/work/CLEO/analysis/DHad/src/2.4/cleo/CesrBeamEnergyProd/Class/BeamEnergyProxy.cc:157: no
   matching function for call to `std::auto_ptr<BeamEnergy>::value()'
/home/xs32/work/CLEO/analysis/DHad/src/2.4/cleo/CesrBeamEnergyProd/Class/BeamEnergyProxy.cc:160: no
   matching function for call to `std::auto_ptr<BeamEnergy>::value()'

Check the method, put this before the reset.

std::cout << "xs23_debug >>> cesr beam energy:" << cesr->value() << std::endl;

Test …

dhad-2.4 gen cleog Single_Dm_to_Kpipipi0 task 3 
std::cout << "xs23_debug >>> beam energy:" << p_BeamEnergy->value() << std::endl;

Test … no output.

Use the INFO.

   report( INFO, kFacilityString ) 
     << "xs23_debug >>> beam energy:" << p_BeamEnergy->value() << endl ;

Set the DEBUG level in the tcl … OK.

Clean the previous log.

dhad-2.4 gen cleog Single_Dm_to_Kpipipi0 task 3

xs23!! -> Change to xshi! Std works too.

Need more digits:

#include <iostream>
#include <iomanip>
   std::cout << "xshi_debug >>> beam energy:" << std::setprecision(8) 
       << p_BeamEnergy->value() << std::endl;

Test …

xshi_debug >>> beam energy:1.8863001

Now find the number in the new release.

More precision in MCBeamParametersProxy

Look up the pkg versions for the MCSymmetricBeamProd

/nfs/cleo3/Offline/rel/20050525_MCGEN/pkg_versions
MCSymmetricBeamProd   MonteCarlo   v01_02_03   proc

Check out the code:

cd $dhad/src/2.4/cleo/
export CVSROOT=/nfs/cleo3/cvsroot
cleo3cvs co -r v01_02_03 MCSymmetricBeamProd
. local_setup
c3make 

Add precision in the code:

#include <iomanip>
report( INFO, kFacilityString ) << "Nominal beam energy = " << std::setprecision(8) 

Back up the previous log.

Test on 10 events …

dhad-2.4 gen cleog Single_Dm_to_Kpipipi0 task 3

Save log cleog.txt.1

 %% NOTICE-ConstantsPhase2Delivery.DBCP2Proxy: DBBeamEnergyShift using version 1.27
 %% DEBUG-CesrBeamEnergyProd.BeamEnergyProxy: Applying Beam Energy shift of -0.00102 GeV
xshi_debug >>> beam energy:1.8863001
 %% DEBUG-ConstantsPhase2Delivery.DBCP2Proxy: getConstants: returned successfully.

 %% NOTICE-ConstantsPhase2Delivery.DBCP2Proxy: DBBeamEnergySpread using version 1.11
 %% INFO-MCSymmetricBeamProd.MCBeamParametersProxy: Nominal beam energy = 1.8863001, spread = 0.0015, min=3.7619936, max=3.7832068

Fit 537 data with new parameters (2.5)

Setup 2.5

cd $dhad/src
mkdir 2.5
cd 2.5
cvs co -d python -r V02-02-02 dhad/src/python
mkdir bash
cp ../2.2/bash/setup.sh bash
ln -s /home/xs32/work/CLEO/analysis/DHad/lib/python/hep python 

Change the src in DHad.py to 2.5.

cd $dhad/7.06/dat/evt
ln -s 537ipb 537ipbv5

Test:

dhad-2.5 fit 537ipbv5 -t d -m 0  ... OK.

Change the parameters

Mass = 3.7718 GeV => 3.7724 Width = 0.0286 GeV => 0.0252 R = 12.3 => 12.7

dhad-2.5 fit 537ipbv5 -t d -m 0  ... OK.

Do the fits

dhad-2.5 fit 537ipbv5 -t d -m 0 –qsub log-2010-03-10 11:51:11

dhad-2.5 fit 537ipbv5 -t d -m 1 –qsub log-2010-03-10 11:52:15

dhad-2.5 fit 537ipbv5 -t d -m 3 –qsub log-2010-03-10 11:52:22

dhad-2.5 fit 537ipbv5 -t d -m 200 –qsub log-2010-03-10 11:52:27

dhad-2.5 fit 537ipbv5 -t d -m 201 –qsub log-2010-03-10 11:52:32

dhad-2.5 fit 537ipbv5 -t d -m 202 –qsub log-2010-03-10 11:52:38

dhad-2.5 fit 537ipbv5 -t d -m 203 –qsub log-2010-03-10 11:52:43

dhad-2.5 fit 537ipbv5 -t d -m 204 –qsub log-2010-03-10 11:52:48

dhad-2.5 fit 537ipbv5 -t d -m 205 –qsub log-2010-03-10 11:52:53

Fits are all OK.

Save the plots … figure

dhad-2.5 web plots 537ipbv5 -t d  

Date: 2010-03-10 16:34:47 EST