Search/Navigation:
Related:
SLAC/EPP/HPS Public
Jefferson Lab/Hall B/HPS Run Wiki
S30XL-LESA/LDMX
Bottom COB has dtm1, data dpm going from dpm0 to dpm6 and control dpm is dpm7. It is marked as "0" with a marker, but can also be seen from the ID tag ending in 0 on the front of the COB.
Top COB had dtm2, data dpm going from dpm8 to dpm14 and control dpm is dpm15. It is marked as "1" with a marker, but can also be identified from the ID barcod tag ending in 1 on the front of the COB.
The RTM C03-01 and C03-02 were placed in the bottom and top location, respectively.
Fibers were tested in fixed locations on the RTM.
To ensure that things are working as expected from this table, they should be placed in the same locations on the RTM according to the 2021 mapping.
Fiber / RTM combination | Status |
---|---|
A | OK |
B | OK |
C | OK |
D | OK |
E | OK |
F | OK |
G | OK |
H | OK |
Image of connection on the RTM Side.
Use connectors with yellow strip (they seem to work best). Slow control channels have been tested to be working if placed in this configuration
Image of connections on the RTM data side. The non-marked fiber is the type-B fiber which comes out from the breakout box.
We decided to avoid connections on the DPM6 as this particular RCE showed few errors during testing.
To do so we used a breakout box to re-map the data connections.
For the HALL-B set-up:
1) Take breakout box labelled as "DATA"
2) Connect fiber type B in MTP1
3) Connect the octopus cable to fiber H
4) Connect the single fibers in the following order (from 0 to 11)
black, red, yellow, grey, teal, pink, white, purple, brown, green, orange, blue
5) In this configuration the mapping should be (X means absent):
dpm pgp
3 3
5 3
X X
1 3
X X
X X
5 1
X X
4 2
5 0
4 0
4 1
This table shows the connections on the flange board
QA of Flange Board ↔ FE0Bs
FEB | FEB SN | Flange Board (HW) | Flange Channel | I At Boot (AnaN, AnaP, Digi) | I Mod ON | I Mod ON + Config | Ctrl Ch (dpm / ch) (1 is top cob) | Fiber (ctrl / Data) | jTag OK | Feb Address | Data Taking | Comment | DAQ MAP |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
L0-1t | 12 | 2 | 2 | 0.242 - 1.212 - 1.250 | 0.292 - 2.317 - 1.392 | 0.292 - 2.833 - 1.322 | 0 / 3 | C / G | Y | 0 | OK | Lost clock after few hrs up | |
L2-3t | 17 | 3 | 0 | 0.256 - 1.248 - 1.239 | 0.317 - 2.319 - 1.314 | 0.313 - 3.112 - 1.313 | 0 / 8 | D / H | Y | 2 | OK | ||
L5t | 05 | 0 | 0 | 0.236 - 1.216 - 1.287 | 0.272 - 2.101 - 1.334 | 0.271 - 2.715 - 1.334 | 1 / 5 | A / E | Y | 6 | OK | Lost clock after few hrs up. Also lost clock at the transition between download and prestart | |
L6t | 16 | 0 | 1 | 0.249 - 1.222 - 1.362 | 0.313 - 2.743 - 1.434 | 0.308 - 3.539 - 1.434 | 1 / 4 | A / E | Y | 8 | OK | Lost clock during download. | |
L4b ( In EPICS is L4t) | 08 | 0 (??) | 2 (??) | 0.248 - 1.241 - 1.248 | 0 / 4 | C / G | Y | 5 | OK | Only 2 HV pins on one hybrid (instead of 4) | |||
L4t ( In EPICS is L4b) | 11 | 1 | 0 | 0.252 1.153 1.374 | 0.307 - 2.8755 - 1.448 | 0.304 - 3.564 - 1.449 | 1 / 8 | B / F | Y | 4 | OK | ||
L5b | 20 | 1 | 1 | 0.246 - 1.169 - 1.243 | 0.306 - 2.242 - 1.346 | 0.302 - 3.063 - 1.321 | 1 / 7 | B / F | Y | 7 | OK | 13 / 1, 13 / 0, 12/ 2, 12/1 | |
L6b | 04 | 1 | 2 | 0.248-1.199-1.278 | 0.307 - 2.705 -1.345 | 0.304 - 3.525 - 1.348 | 1 / 6 | B / F | Y | 9 | OK | ||
L0-1b | 07 | 2 | 0 | 0.241 - 1.284 - 1.313 | 0.293 - 2.235 - 1.394 | 0.293 - 2.931 - 1.394 | 0 / 5 | C / G | Y | 1 | OK | ||
L2-3b | 01 | 3 | 1 | 0.220 - 1.330 - 1.388 | 0.311 - 2.452 - 1.456 | 0.305 - 3.286 - 1.456 | 0 / 7 | D / H | Y | 3 | OK |
This table shows a visual of the connections from the cooling plate to the flange, looking from the detector toward the flange. X indicates no connection.
Flange Board Number | ||||
---|---|---|---|---|
3 | 2 | 1 | 0 | |
FEB | L23T | L1B | L4T | L5T |
L23B | X | L5B | L6T | |
X | L1T | L6B | L4B | |
List of issues we observed during local data taking relative to the Back - End.
NOTICE!!! The dpm numbers in the daq map after 7 are off by one. Example: DataDpm[13] in the DAQ Map is hostname dpm14.
Server machines: clonfarm1, clon10new
TiCrate: hdice1
COBS: dpm0 : dpm6, dpm7, dtm1
dpm8 : dpm14, dpm15, dtm2
Connect to clonfarm1 as clasrun and change to bash shell
SDK software installation to talk to the atca crate
source /usr/clas12/release/1.4.0/slac_svt_new/V3.4.0/i86-linux-64/tools/envs-sdk.sh examples: cob_dump --all atca1 => dumps the RCE status (booted is 0xea) cob_rce_reset atca1 ==> resets all the RCEs cob_rce_reset atca1/1/0/2 ==> resets a particular dpm (in this case dpm1) cob_cold_data_reset atca1 ==> "power cycles" the RCEs (sometimes they do not come back up nicely so rce_reset might be needed after)
SVT software is installed in
/data/hps/slac_svt/ => diskless is exported to the cobs via nfs (server hosted on clonfarm1, NFS has to be v2 !! important ) => daq is exported to the cobs via nfs. Has been compiled on DTM1 => server contains the current software installation
On clonfarm1 start the rogue server
source /usr/clas12/release/1.4.0/slac_svt_new/anaconda3/etc/profile.d/conda.sh conda activate rogue_5.9.3 cd /data/hps/slac_svt/server/heavy-photon-daq/software/scripts/ python SvtCodaRun.py --local --env JLAB
On clon10new start the dummy server
python SvtCodaDummy.py --local --env JLAB
Here are the instructions to how to start Coda with clasrun user
python SvtCodaDummy.py --local --env JLAB 1) Open a terminal on clonfarm1 > bash > sconda #setup conda > crogue #activates rogue environment > startCodaRun #starts Rogue Coda Run Gui 2) Open a terminal on clon10new > bash > sconda; crogue; startCodaDummy #starts the dummy rogue server 3) Open a terminal on clonfarm1 > runcontrol -rocs Then: connectconfigure (hdice1_clonfarm1_clon10new to run with both cobs, hdice1_clonfarm1 to run with top cob, hdice1_clon10new to run with bottom cob)download (for both cobs use hps_v2_noThr.trg)PrestartGo
We are using conda release 1.4.0, installed in
/usr/clas12/release/1.4.0 HPS trigger configurations are stored in /usr/clas12/release/1.4.0/parms/trigger/HPS/
If event_builder or rogue_lite change, coda libraries need to be updates. Just run
sh /data/hps/slac_svt/copy_libraries.sh
This is in the event builder and indicates that we are dropping packets.
This is raised in RssiContributor.cc in the event_builder by acceptFrame thread
Effect:
rol BUSY
Causes: Dropped frames.
Fix: Reset COBs
This usually indicates that the FEBs lost clock
Effect: not possible to run
Causes: FEBs lost clock
Fix: Recycle FEBs and run control