Versions Compared


  • This line was added.
  • This line was removed.
  • Formatting was changed.
Comment: Migration of unmigrated content due to installation of a new plugin
  1. Determine *when* the runs were taken.  You'll need the day number of the year.  One method is:
    No Format
    >>> from datetime import *
    >>> datetime(<yyyy>, <mm>, <dd>).toordinal() - datetime(<yyyy>, 1, 1).toordinal() + 1
    Copy the data from /gnfs/data/DVD/RawPackets/2006<DDD>/ to a local directory, say ~/sci (where DDD is the day of year from above).  The date and time found in the eLog runs section is the same as used in the .../RawArchive/* directories.  Only the *_SCI.pkt files that span the run of interest are needed.
  2. mkdir ~/sci/lsf
  3. Using the glast account on a lat-licos* or equivalent (i.e., a machine that has LICOS_Scripts installed):
    No Format
    python $ONLINE_ROOT/LICOS_Scripts/analysis/ --lsd_dir=~/sci --lsf_dir=~/sci/lsf
    Determine which of the output files is the run you want.  File format is:
  4. No Format
    GroundId is the run number, without the 077000000, in hex.  Verify that nFirst is 0000. 
  5. Login to lat-dmz01 and enter the following to connect to the MySQL database:
    No Format
    mysql --user elogbook elogbook --password
    (enter the password at the prompt)
  6. Enter the following command to reset the analysis state:
    No Format
    update LICOS_activities set JobCompleteList = '[None]', AnalysisState_FK = 7 where RunID_FK = '077RRRRRR';
    (replace 077RRRRRR with the acquire run id)
  7. Copy the lsf file(s) to "/nfs/data/lsf/.".  If the corresponding DAQ run had multiple steps, there will be multiple lsf files.
  8. AnalysisEngine will see them and should be able to reprocess the run.