Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

P120 Reprocessing

status: In Progress Complete
last update: 30 17 August 20102011

This page is a record of the configuration for the P120 reprocessing project, event reclassification using Pass 7.3/7.4/7.6. This project involves reprocessing with Pass7 classification trees and (ultimately) new IRFs. The name "P120" derives from the word "processing" and the initial file version to be used for the output data products, e.g., r0123456789_v120_merit.root.

  • P120-MERIT - this task reads DIGI+RECON+MERIT and produces reprocessed MERIT + FILTEREDMERIT (photons) + ELECTRONMERIT
  • P120-FT1 - will (eventually) read FILTEREDMERIT and produce
    • reads MERIT and produces FT1 (photons) + LS1 (merit-like FITS file for photons) + electron FITS file
    + LS3 (live-time cube)
    • V2.1 of this task is modified such that FT1 and LS1 files are filtered for FSSC, while new EXTENDEDFT1 and EXTENDEDLS1 files are produced containing all photon event classes.
  • P120-LEO-MERIT - this task reads DIGI+RECON+MERIT and produces P120-LEO-MERIT - this task reads DIGI+RECON+MERIT and produces reprocessed MERIT + FILTEREDMERIT (photons) + ELECTRONMERIT for 200 runs of earth limb (L&EO) data

Datafile names, versions and locations

Data file version numbers for this reprocessing will begin with v120.

XROOT location and file naming

Location template:

Code Block
/glast/Data/Flight/Reprocess/<reprocessName>/<dataType>

...

Code Block
/glast/Data/Flight/Reprocess/P120/merit
/glast/Data/Flight/Reprocess/P120/filteredmerit
/glast/Data/Flight/Reprocess/P120/electronmerit
/glast/Data/Flight/Reprocess/P120/ft1
/glast/Data/Flight/Reprocess/P120/extendedft1
/glast/Data/Flight/Reprocess/P120/electronft1
/glast/Data/Flight/Reprocess/P120/ls1
/glast/Data/Flight/Reprocess/P120/ls3extendedls1

File naming:

Data Type

aka

Send to FSSC

Naming template

MERIT

 

No

r<run#>_<version>_<dataType>.root

FILTEREDMERIT

 

No

r<run#>_<version>_<dataType>.root

ELECTRONMERIT

 

No

r<run#>_<version>_<dataType>.root

ELECTRONFT1

 

No

gll_el_p<procVer>_r<run#>_<version>.fit

FT1 EXTENDEDFT1

LS-002  

YesNo

gll_phxp_p<procVer>_r<run#>_<version>.fit

LS1 FT1

LS-001 002

Yes

gll__ph_p<procVer>_r<run#>_<version>.fit

EXTENDEDLS1

 

No

gll_xe_p<procVer>_evp<procVer>r<run#>_<version>.fit

LS3 LS1

LS-003 001

MaybeYes

gll_ev_ltp<procVer>_r<run#>_<version>.fit

Note: 'procVer' is a field added to the file name (and the keyword "PROC_VER" in the primary header) added to the FFD 5/12/2010. Ref: http://fermi.gsfc.nasa.gov/ssc/dev/current_documents/Science_DP_ICDFFD_RevA.pdfImage Removed

Example:

Code Block
/glast/Data/Flight/Reprocess/P120/merit/r0239557414_v120_merit.root
/glast/Data/Flight/Reprocess/P120/filteredmerit/r0239557414_v120_filteredmerit.root
/glast/Data/Flight/Reprocess/P120/electronmerit/r0239557414_v120_electronmerit.root
/glast/Data/Flight/Reprocess/P120/ft1extendedft1/gll_phxp_p120_r0239559565_v120.fit
/glast/Data/Flight/Reprocess/P120/electronft1ft1/gll_elph_p120_r0239559565_v120.fit
/glast/Data/Flight/Reprocess/P120/ls1electronft1/gll_evel_p120_r0239559565_v120.fit
/glast/Data/Flight/Reprocess/P120/ls3extendedls1/gll_ltxe_p120_r0239559565_v120.fit
/glast/Data/Flight/Reprocess/P120/ls1/gll_ev_p120_r0239559565_v120.fit
DataCatalog location and naming

Logical directory and group template:

...

Code Block
Data/Flight/Reprocess/P120:MERIT r0239557414
Data/Flight/Reprocess/P120:FILTEREDMERIT r0239557414
Data/Flight/Reprocess/P120:EXTENDEDFT1 r0239557414
Data/Flight/Reprocess/P120:FT1 r0239557414
Data/Flight/Reprocess/P120:ELECTRONFT1 r0239557414
Data/Flight/Reprocess/P120:LS1EXTENDEDLS1 r0239557414
Data/Flight/Reprocess/P120:LS3LS1 r0239557414

Data Sample

The currently defined data sample for P120 reprocessing includes:

First run

239557414 (MET), 2008-08-04 15:43:34 (UTC)

beginning of Science

Last run

302313321 333880535 (MET), 20102011-0708-31 2301 08:5535:21 33 (UTC)

Official Pass7 release

Total runs

16,459

 10916

Total MERIT events

2335,758921,285,498

Total FT1 events

n/a

Bookkeeping

  1. (This page): Define ingredients of reprocessing (processing code/configuration changes)
  2. Processing History database: http://glast-ground.slac.stanford.edu/HistoryProcessing/HProcessingRuns.jsp?processingname=P120Image Removed
    1. List of all reprocessings
    2. List of all data runs reprocessed
    3. Pointers to all input data files (-> dataCatalog)
    4. Pointers to associated task processes (-> Pipeline II status)
  3. Data Catalog database: http://glast-ground.slac.stanford.edu/DataCatalog/folder.jspImage Removed
    1. Lists of and pointers to all output data files
    2. Meta data associated with each output data product

...

Status chronology

  • 8/29/2010 - Discovered three merge steps that silently failed (xroot file access). TASK complete.
  • 8/28/2010 - processing formally complete (10916 runs), but some discrepancy in # of events
  • 8/26/2010 - serious xroot problems. See initial distribution of files across xroot servers. From this report (courtesy Wilko) it is easy to see where problems are likely to arise - when the number of servers involved is small, e.g. two or three.
  • 8/19/2010 - production continues at a crawl due to xroot server difficulties
  • 8/16/2010 - resume full production, but at a slow trickle (max 350 simultaneous processClump jobs)
  • 8/8/2010 - block 2 reprocessing complete. Many xroot server problems. (5 days to process 2084 runs)
  • 8/3/2010 - begin block 2 reprocessing (through 255132033 MET), bringing the total runs reprocessed to 2721, about 5-1/2 months of data.
  • 7/28/2010 - block 1 re-reprocessing complete
  • 7/27/2010 - New GlastRelease (v17r35p10) containing new evtUtils, "to make the FT1EventClass bits compatible with the ScienceTools". Cleanup, including removing all files created last week during the first attempt.
  • 7/21/2010 - block 1 reprocessing complete
  • 7/20/2010 - agree upon 'pilot block' of runs (239557417 - 243220241), 637 runs. Begin...
  • 7/19/2010 - submit first test run. success. await feedback

Configuration

...

Task Location

...

/nfs/farm/g/glast/u38/Reprocess-tasks/P120-MERIT

666,747

all "events"

Total FILTEREDMERIT/EXTENDEDFT1/LS1 events

5,035,929,409

all photon event classes

Total ELECTRONMERIT/ELECTRONFT1 events

68,055,849

 

Total LS1 (FSSC selection) events

1,025,359,231

event classes (bits) 0,2,3,4 (transient, source, clean, ultraclean)

Total FT1 (FSSC selection) events

142,042,060

event classes (bits) 2,3,4 (source, clean, ultraclean)

Total disk space used

33.9 TB

 

Summary from DataCatalog as of 8/2/2011.

Name

Files

Events

Size

ELECTRONFT1

16459

68,055,849

6.4 GB

ELECTRONMERIT

16459

68,055,849

147.4 GB

EXTENDEDFT1

16459

5,035,929,409

441.3 GB

EXTENDEDLS1

16459

5,035,929,409

816.5 GB

FILTEREDMERIT

16459

5,035,929,409

4.0 TB

FT1

16459

142,042,060

12.9 GB

LS1

16459

1,025,359,231

166.6 GB

MERIT

16459

35,921,666,747

28.3 TB

NOTE: One run, 242429468, of type TrigTest was declared 'good for science' but long after this task got started, so it has been intentionally omitted.

8/17/2011 update: Four orphaned runs are being reprocessed, including one TrigTest run and four nadir-pointed runs.

Stream

run

type

16459

242429468

(TrigTest)

16460

333355876

(nadirOps)

16461

333358500

(nadirOps)

16462

333365716

(nadirOps)

Bookkeeping

...

Task Status

  1. (This page): Define ingredients of reprocessing (processing code/configuration changes)
  2. Processing History database:
  1. http://glast-ground.slac.stanford.edu/

...

  1. HistoryProcessing/

...

  1. HProcessingRuns.jsp

...

GlastRelease

...

v17r35p8 v17r35p10

  1. ?processingname=P120
    1. List of all reprocessings
    2. List of all data runs reprocessed
    3. Pointers to all input data files (-> dataCatalog)
    4. Pointers to associated task processes (-> Pipeline II status)
  2. Data Catalog database: http://glast-ground

...

Input Data Selection

"standard" from

...

  1. .slac.stanford.edu/

...

along with "&& (RunQuality != "Bad" || is_null ( RunQuality )"

...

s/c data

...

FT2 from P105 (runs 239557414 - 271844560), then from current Level 1 production

...

Input Run List

...

photonFilter

...

CTBParticleType==0 && CTBClassLevel>0

...

electronFilter

...

CTBParticleType==1

...

jobOpts

...

Output Data Products

...

MERIT, FILTEREDMERIT, ELECTRONMERIT

  1. DataCatalog/folder.jsp
    1. Lists of and pointers to all output data files
    2. Meta data associated with each output data product

...

P120-MERIT
Anchor
P120-MERIT
P120-MERIT

Status chronology

  • 8/17/2011 - Begin reprocessing four orphan runs.
  • 8/1/2011 - Begin and complete final backfill through run 333880535 (2011-08-01 08:35:33 UTC). For now, the three nadirOps runs are represented by dummy place-holder entries in the runFile.txt – their pipeline streams will fail.
  • 7/29/2011 - Modified selection criteria for reprocessing run selection (findRunsRepro.py) to allow nadir-pointed data. This means adding ' || sIntent=="nadirOps"' to the dataCatalog selection string. See https://confluence.slac.stanford.edu/display/ISOC/Nadir+Obs+Test+-+26+July+2011 for a list of runs affected by the nadir-pointed test.
  • 7/26/2011 - Recovered three missing runs (below).
  • 7/21/2011 - Begin and (mostly) complete reprocessing block 13, through run 332930755 (2011-07-21 08:45:53 UTC), 528 new runs - special GRB request. Note three runs failed skim/merge and are being investigated:

    stream

    run

    UTC

    Crash location

    16136

    332054399

    2011-07-11 05:19:57

    Filtered Merit skim failure

    16151

    332140182

    2011-07-12 05:09:40

    Electron Merit skim failure

    16156

    332169056

    2011-07-12 13:10:54

    Filtered Merit skim failure

  • 6/17/2011 - Begin and complete reprocessing block 11, through run 329923889 (2011-06-16 13:31:27 UTC), 889 new runs.
  • 4/19/2011 - Begin and complete reprocessing block 10, through run 324849509 (2011-04-18 19:58:27 UTC), 52 new runs.
  • 4/15/2011 - Begin and complete reprocessing block 9, through run 324551768 (2011-04-15 10:51:27 UTC)
  • 4/13/2011 - Begin and complete reprocessing block 8, through run 324368491 (2011-04-13 06:21:29 UTC)
  • 4/3/2011 - The three missing runs have been produced by Level 1. Runlist recreated and those runs rolled back. There are no missing runs at this point.
  • 3/29/2011 - Due to some hidden I/O problems, two changes have been made to this task, neither of which should affect data content. Note that this change will take effect on runs after 321756673, or any runs rolled back after this date.

    Package

    Old Version

    New Version

    Reason for change

    GPLtools

    v1r15p1

    GPLtools-02-00-00

    Checks size of files before and after move between disk and xroot

    skimmer

    07-07-00

    08-01-00

    detects failure to open input file

  • 3/17/2011 - Last block complete (one instance of skimmer failure in mergeClumps)
  • 3/15/2011 - Expand P120 to present (last run 321756673, 2011-03-14 00:51:11 UTC). Three runs continue to be dummied out. Once those runs have proper RECON files, their streams can be rolled back.

    run

    task stream

    # subStreams

    306353950

    11624

    10

    316611240

    13431

    8

    320850543

    14171

    7

  • 2/4/2011 - Expand P120 to present (last run 318211122, 2011-01-31 23:58:40 UTC). This include three runs for which there are no Recon files which have, for the moment, been supplied with 'dummy' entries in the runFile.:

    306353950

    'bad chunk' (known previously) - awaiting GR update

    313483912

    missing 700s, being worked on

    316611240

    'bad chunk' - awaiting GR update

  • 1/12/2011 - Expand P120 coverage to include Crab ToO
    • Crab ToO 9/23/2010 15:50:50 to 9/27/2010 19:49:38, corresponding to MET r0306949696-r0307308940
    • reconfig through end of Sep 2010: 10916 -> 11841 runs, increase of 925 runs
    • No recon file for run 306353950 (being worked on, for the moment, place dummy file in runFile.txt)
    • First run in new block completed, awaiting checkered flag to continue...
  • 8/29/2010 - Discovered three merge steps that silently failed (xroot file access). TASK complete.
  • 8/28/2010 - processing formally complete (10916 runs), but some discrepancy in # of events
  • 8/26/2010 - serious xroot problems. See initial distribution of files across xroot servers. From this report (courtesy Wilko) it is easy to see where problems are likely to arise - when the number of servers involved is small, e.g. two or three.
  • 8/19/2010 - production continues at a crawl due to xroot server difficulties
  • 8/16/2010 - resume full production, but at a slow trickle (max 350 simultaneous processClump jobs)
  • 8/8/2010 - block 2 reprocessing complete. Many xroot server problems. (5 days to process 2084 runs)
  • 8/3/2010 - begin block 2 reprocessing (through 255132033 MET), bringing the total runs reprocessed to 2721, about 5-1/2 months of data.
  • 7/28/2010 - block 1 re-reprocessing complete
  • 7/27/2010 - New GlastRelease (v17r35p10) containing new evtUtils, "to make the FT1EventClass bits compatible with the ScienceTools". Cleanup, including removing all files created last week during the first attempt.
  • 7/21/2010 - block 1 reprocessing complete
  • 7/20/2010 - agree upon 'pilot block' of runs (239557417 - 243220241), 637 runs. Begin...
  • 7/19/2010 - submit first test run. success. await feedback

Configuration

Task Location

/nfs/farm/g/glast/u38/Reprocess-tasks/P120-MERIT

Task Status

http://glast-ground.slac.stanford.edu/Pipeline-II/exp/Fermi/task.jsp?task=41146114

GlastRelease

v17r35p8 v17r35p10

Input Data Selection

"standard" from

https://confluence.slac.stanford.edu/display/SCIGRPS/LAT+Dataset+Definitions

along with "&& (RunQuality != "Bad" || is_null ( RunQuality )"

s/c data

FT2 from P105 (runs 239557414 - 271844560), then from current Level 1 production

Input Run List

ftp://ftp-glast.slac.stanford.edu/glast.u38/Reprocess-tasks/P120-MERIT/config/runFile.txt

photonFilter

CTBParticleType==0 && CTBClassLevel>0

electronFilter

CTBParticleType==1

jobOpts

ftp://ftp-glast.slac.stanford.edu/glast.u38/Reprocess-tasks/P120-MERIT/config/reClassify.txt

Output Data Products

MERIT, FILTEREDMERIT, ELECTRONMERIT

Timing and Scaling

  • (beyond block 2 results) Due to xroot problems (overstressing a small number of machines) the processing throughput dropped to 25-30 runs/hour (190-225 jobs/hour)
    • Wilko begins redistributing files around the xroot system in order to balance the load. This is only partially done by task completion.
    • Logs of job submission can be found here
  • (block 1 results) The processClump step is taking ~40 hequ-minutes (or ~65 fell-minutes). With >500 simultaneous jobs running, there is little noticeable strain on xroot. There are five servers in the yellow-orange load range and they are claiming ~110-130 MB/s I/O rate.
    Image Added
  • The mergeClumps step is taking ~5 hequ-minutes
  • It was observed that submitting 70 runs at once put a strain on /u30, home of GlastRelease. Some 93 of ~540 jobs failed with I/O error, but succeeded upon rollback.

Load balancing

Introduce new trickleStreams.py script to (partially) assess pipeline activity and only the number of jobs considered appropriate based on available data.
(block 1)

Code Block

maxProcessClumps = 600     ## prevent overload of xroot
maxMergeClumps = 20        ## prevent overload of xroot (inactive)
maxStreamsPerCycle = 20    ## prevent overload of /u30 on startup
timePerCycle = 900         ## 15 minutes:  allow time for dust to settle

With these parameters, it took ~ 5 hours to reach a point where fewer than 20 jobs per cycle were regularly submitted. Another 4.5 hours for the task to complete. On average, one run generated 7.5 processClump batch jobs.

For subsequent data (beyond block 2), xroot displayed such stress, that the maxProcessClumps limit was reduced to 250 or 300.

...

P120-FT1
Anchor
P120-FT1
P120-FT1

This task generates all desired FITS data products. An example of the code processing chain appears on a child page.

Status chronology

  • 8/1/2011 - Begin and complete final block of Pass7 reprocessing
  • 7/26/2011 - Recovered the three missing runs (see P120-MERIT chronology), and reran stream 1018 (run 245403855), which had a bogus tstart time in the datacatalog – leap second issue, and recovered 27 events in that run.
  • 7/22/2011 - Begin and (mostly) reprocessing block 13, through run 332930755 (2011-07-21 08:45:53 UTC), 525 new runs (+ 3 'dummy' runs due to skim crashes, see above) - special GRB request.
  • 7/3/2011 - Task complete through run 329923889 (2011-06-16 13:31:27 UTC), 15,763 runs
  • 6/30/2011 - New ST 09-24-00 (with gtdiffrsp fix), restart trials with task version 2.1
  • 6/14/2011 - Begin trials. Concern that gtdiffrsp is crashing often (20-25% of time)
  • 6/7/2011 - IMPORTANT UPDATE: a decision was made to rollback entire task with these changes:
    • calculate diffuse response for 'source' and 'clean' event classes
    • Produce new subset photon files for FSSC (FT1 with source and above, LS1 with transient and above)
    • Update various configurations (ScienceTools, evtClassDefs, etc.)
      This is being done by creating a whole new task, version 2.0, which from the pipeline perspective will overlay the older version 1.0.
  • 4/19/2011 - Begin and complete reprocessing block 10, through run 324849509 (2011-04-18 19:58:27 UTC), 52 new runs.
  • 4/15/2011 - Begin and complete reprocessing block 8, through run 324551768 (2011-04-15 10:51:27 UTC)
  • 4/14/2011 - Begin and complete reprocessing block 7, through run 324368491 (2011-04-13 06:21:29 UTC)
  • 4/3/2011 - The three missing runs have now been reprocessed. There are no missing runs at this point.
  • 3/17/2011 - Catch up with P120-MERIT (last run 321756673, 2011-03-14 00:51:11 UTC)
  • 3/15/2011 - Due to missing run, rollback runs 11624-11841. Bookkeeping is now correct.
  • 2/4/2011 - Catch up with merit production (through 30 Sep 2010), but with one missing run/stream
  • 1/28/2011 - Pass 7.4 reincarnation of this task complete through 31 Jul 2010
  • 1/24/2011 - Entire task, xroot files, dataCat entries deleted. Prepare to reprocess as Pass 7.4

Timing and Scaling

  • (beyond block 2 results) Due to xroot problems (overstressing a small number of machines) the processing throughput dropped to 25-30 runs/hour (190-225 jobs/hour)
    • Wilko begins redistributing files around the xroot system in order to balance the load. This is only partially done by task completion.
    • Logs of job submission can be found here
  • (block 1 results) The processClump step is taking ~40 hequ-minutes (or ~65 fell-minutes). With >500 simultaneous jobs running, there is little noticeable strain on xroot. There are five servers in the yellow-orange load range and they are claiming ~110-130 MB/s I/O rate.
    Image Removed
  • The mergeClumps step is taking ~5 hequ-minutes
  • It was observed that submitting 70 runs at once put a strain on /u30, home of GlastRelease. Some 93 of ~540 jobs failed with I/O error, but succeeded upon rollback.

Load balancing

Introduce new trickleStreams.py script to (partially) assess pipeline activity and only the number of jobs considered appropriate based on available data.
(block 1)

Code Block

maxProcessClumps = 600     ## prevent overload of xroot
maxMergeClumps = 20        ## prevent overload of xroot (inactive)
maxStreamsPerCycle = 20    ## prevent overload of /u30 on startup
timePerCycle = 900         ## 15 minutes:  allow time for dust to settle

With these parameters, it took ~ 5 hours to reach a point where fewer than 20 jobs per cycle were regularly submitted. Another 4.5 hours for the task to complete. On average, one run generated 7.5 processClump batch jobs.

For subsequent data (beyond block 2), xroot displayed such stress, that the maxProcessClumps limit was reduced to 250 or 300.

...

This task will be run twice: Pass 1 will perform event classification for source and transient events and allow analysis to produce diffuse class IRFs; Pass 2 will be identical to Pass 1 but will include diffuse classification. The latest word from C&A is that diffuse response will only be calculated for 'source' class events.

...

  • 8/31/2010 - Pass 1 of this task is complete (through 31 July 2010)
  • 8/30/2010 - Problem with makeFT1 stressing /u38 (very large temporary file needed when using xml representation of event classes was being written to $PWD). Jim makes update to fitsGenApps => ST 09-18-03, put into production at stream 1400.
  • 8/29/2010 - Begin Pass 1 of task...

Configuration (version 2)

Task Location

/nfs/farm/g/glast/u38/Reprocess-tasks/P120-FT1

Task Status

http://glast-ground.slac.stanford.edu/Pipeline-II/exp/Fermi/indextask.jspImage Removed?task=65047878

Input Data Selection

MERIT (from P120-MERIT) P120-MERIT)

spacecraft data

FT2 from P105 (runs 239557414 - 271844560), then from current Level 1 production

Input Run List

ftp://ftp-glast.slac.stanford.edu/glast.u38/Reprocess-tasks/P120-FT1/config/runFile.txtImage Removed

Reprocessing Mode

reFT1 evtClassDefs

00-17-00

meritFilter

FT1EventClass!=0

evtClassDefs

00-19-01

eventClassifier

Pass7_Classifier.py (OBSOLETE)

eventClassMap

EvtClassDefs_P7V3P7V6.xml (in evtUtils)

s/c data

FT2 from P105 (runs 239557414 - 271844560), then from current Level 1 production

ScienceTools

09-18-01 through stream 1399, then 09-18-03 (SCons build) 24-00

Code Variants

redhat4-i686-32bit-gcc34 and redhat5-i686-32bit-gcc41 (Optimized)

Diffuse Model

based on contents of /afs/slac.stanford.edu/g/glast/ground/releasesGLAST_EXT/analysisFiles/diffuse/v2/source_model_v02.xml diffuseModels/v2r0
(see

https://confluence.slac.stanford.edu/display/SCIGRPS/DiffuseQuick+ModelStart+for+Analysis+of+LAT+DataImage Removedwith+Pass+7

)

Diffuse Response IRFs

P7trans_v3mc, P7source_v3mc (TEMPORARY)

'source' using P7SOURCE_V6 IRF
'clean' using P7CLEAN_V6 IRF

IRFs

P6V7, contained within ScienceTools release

IRFs

implemented as 'custom irf', files in /afs/slac.stanford.edu/g/glast/ground/PipelineConfig/IRFS/Pass7.3

Output Data Products

FT1, LS1, EXTENDEDFT1, EXTENDEDLS1, ELECTRONFT1

Processing chain for FITS data products

Data Product

makeFT1

gtdiffrsp

gtmktime

gtltcube

Product

selection

makeFT1

gtdiffrsp

gtmktime

gtltcube

FT1

'source' and above
EVENT_CLASS bits 2,3,4

true

true

true

false

LS1

'transient' and above
EVENT_CLASS bits 0,2,3,4

true

true

true

false

FT1EXTENDED

FT1EventClass!=0FT1

true

true forevclsmin==1

true

falseLS1

LS1EXTENDED

FT1EventClass!=0

true

falsetrue

true

false

ELECTRONFT1

CTBParticleType==1

true

false

true

false

Note that diffuse response is calculated only for 'source' class events (and not transient); diffuse class events are not yet classified.

Note on 'Code Variant': The SLAC batch farm contains a mixture of architectures , both hardware (Intel/AMD 64-bit) and software (RedHat Enterprise Linux 4, and 5, gcc 3.4, 4.1, etc.).

Note on diffuse response calculation: (OBSOLETE)gtdiffrsp is called three times in succession. The first time with IRF P7_v2_diff and evclsmin==8, followed by IRF P7_v2_extrad and evclsmin==9, and finally IRF P7_v2_datac and evclsmin==10. The resulting FT1 file has six columns of diffuse response, two columns (galactic and extragalactic response) for each of the three IRFs. This creates a non-standard FT1 file by FSSC standards as they expect only five diffuse response columns.(OBSOLETE)

Timing

false

Note that diffuse response is calculated for 'source' and 'clean' event classes only.

Note on 'Code Variant': The SLAC batch farm contains a mixture of architectures , both hardware (Intel/AMD 64-bit) and software (RHEL5-64, gcc v4.1, etc.).

Timing

  • 1/28/2011 - Without diffuse response, the mergeClumps jobs are taking about 10 hequ-minutes of CPU to complete.
  • 8/31/2010 - The primary batch job, mergeClumps, took a (mean) time of 42 cpu minutes (primarily a mixture of hequ and fell class machines). 8/31/2010 - With P120-MERIT files nicely distributed across xroot servers, there were no xroot limitations to the processing. After the update to makeFT1, there was no longer an issue with overloading /u38 ($PWD). The next bottleneck was the pipeline processing itself. This task consists of three batch jobs and four scriptlets; it was observed that the pipeline allowed hundreds of jobs to dwell in the READY state for extended periods of time, thus making it impossible to keep LSF saturated. Nevertheless, the maximum number of simultaneous jobs approached 2000. The task essentially completed in 8 hours, although some lingerers kept 'running' for another nine hours (mostly in SSUSP). A profile of job processing rate appears in this plot:

...

P120-LEO-MERIT
Anchor
P120-LEO-MERIT
P120-LEO-MERIT

Status chronology

  • 8/16/2010 - Task complete (199 runs)
  • 8/13/2010 - Create task

Configuration

Identical to the P120-MERIT task, except use FT2 files from P110 reprocessing.