Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
  1. Determine *when* the runs were taken.  You'll need the day number of the year.  One method is: 
    No Format
    python
    >>> from datetime import *
    >>> datetime(<yyyy>, <mm>, <dd>).toordinal() - datetime(<yyyy>, 1, 1).toordinal() + 1
    
    Copy the data from /gnfs/data/DVD/RawPackets/2006<day of year>2006<DDD>/ to a local directory, say ~/sci (where DDD is the day of year from above).  (It is probably best to use the glast account for this to ensure that you are working with the correct environment.)  The date and time found in the eLog runs section is the same as used in the .../RawArchive/* directories.  Only the *_SCI.pkt files that span the run of interest are needed.
  2. mkdir ~/sci/lsf
  3. Using the glast account on a lat-licos* or equivalent (i.e., a machine that has LICOS_Scripts installed):
    No Format
    python $ONLINE_ROOT/LICOS_Scripts/analysis/LsfWriter.py --lsd_dir=~/sci --lsf_dir=~/sci/lsf
    
    Determine which of the output files is the run you want.  File format is:
  4. No Format
    <ApId>-<GroundId>-<Time>-<nFirst>-<nEnd>.lsf
    
    GroundId is the run number, without the 077000000, in hex.  Verify that nFirst is 0000. 
  5. Login to lat-dmz01 and enter the following to connect to the MySQL database:
  1. No Format
    mysql --user elogbook elogbook --password
    
    (enter the password at the prompt)
  2. Enter the following command to reset the analysis state:
    No Format
    update LICOS_activities set JobCompleteList = '[None]', AnalysisState_FK = 7 where RunID_FK = '999999999077RRRRRR';
    
    (replace 999999999 077RRRRRR with the acquire run id)
  3. Copy the lsf file(s) to "/nfs/data/lsf/.".  If the corresponding DAQ run had multiple steps, there will be multiple lsf files.
  4. AnalysisEngine will see them and should be able to reprocess the run.