Search/Navigation:
Related:
SLAC/EPP/HPS Public
Jefferson Lab/Hall B/HPS Run Wiki
S30XL-LESA/LDMX
Fibers were tested in fixed locations on the RTM.
To ensure that things are working as expected from this table, they should be placed in the same locations on the RTM according to the 2021 mapping.
Fiber / RTM combination | Status |
---|---|
A | OK |
B | OK |
C | OK |
D | OK |
E | OK |
F | OK |
G | OK |
H | OK |
QA of Flange Board ↔ FE0Bs
FEB | FEB SN | Flange Board (HW) | Flange Channel | I At Boot (AnaN, AnaP, Digi) | I Mod ON | I Mod ON + Config | Ctrl Ch (dpm / ch) (1 is top cob) | Fiber (ctrl / Data) | jTag OK | Feb Address | Data Taking | Comment |
---|---|---|---|---|---|---|---|---|---|---|---|---|
L0-1t | 12 | 2 | 2 | 0.242 - 1.212 - 1.250 | 0.292 - 2.317 - 1.392 | 0.292 - 2.833 - 1.322 | 0 / 3 | C / G | Y | 0 | OK | Lost clock after few hrs up |
L2-3t | 17 | 3 | 0 | 0.256 - 1.248 - 1.239 | 0.317 - 2.319 - 1.314 | 0.313 - 3.112 - 1.313 | 0 / 8 | D / H | Y | 2 | OK | |
L5t | 05 | 0 | 0 | 0.236 - 1.216 - 1.287 | 0.272 - 2.101 - 1.334 | 0.271 - 2.715 - 1.334 | 1 / 5 | A / E | Y | 6 | OK | Lost clock after few hrs up. Also lost clock at the transition between download and prestart |
L6t | 16 | 0 | 1 | 0.249 - 1.222 - 1.362 | 0.313 - 2.743 - 1.434 | 0.308 - 3.539 - 1.434 | 1 / 4 | A / E | Y | 8 | OK | Lost clock during download. |
L4b ( In EPICS is L4t) | 08 | 2 | 1 | 0.248 - 1.241 - 1.248 | 0 / 4 | C / G | Y | 5 | OK | Only 2 HV pins on one hybrid (instead of 4) | ||
L4t ( In EPICS is L4b) | 11 | 1 | 0 | 0.252 1.153 1.374 | 0.307 - 2.8755 - 1.448 | 0.304 - 3.564 - 1.449 | 1 / 8 | B / F | Y | 4 | OK | |
L5b | 20 | 1 | 1 | 0.246 - 1.169 - 1.243 | 0.306 - 2.242 - 1.346 | 0.302 - 3.063 - 1.321 | 1 / 7 | B / F | Y | 7 | OK | |
L6b | 04 | 1 | 2 | 0.248-1.199-1.278 | 0.307 - 2.705 -1.345 | 0.304 - 3.525 - 1.348 | 1 / 6 | B / F | Y | 9 | OK | |
L0-1b | 07 | 2 | 0 | 0.241 - 1.284 - 1.313 | 0.293 - 2.235 - 1.394 | 0.293 - 2.931 - 1.394 | 0 / 5 | C / G | Y | 1 | OK | |
L2-3b | 01 | 3 | 1 | 0.220 - 1.330 - 1.388 | 0.311 - 2.452 - 1.456 | 0.305 - 3.286 - 1.456 | 0 / 7 | D / H | Y | 3 | OK |
List of issues we observed during local data taking relative to the Back - End.
NOTICE!!! The dpm numbers in the daq map after 7 are off by one. Example: DataDpm[13] in the DAQ Map is hostname dpm14.
MTP1 => typeB fiber
0 1 2 3 4 5 6 7 8 9 10 11
L L N L N N L N L L L L
33 53 62 13 51 42 50 40 41
For the HALL-B:
1) Take breakout box labelled as "DATA"
2) Connect fiber type B in MTP1
3) Connect the octopus cable to fiber H
4) Connect the single fibers in the following order (from 0 to 11)
black, red, yellow, grey, teal, pink, white, purple, brown, green, orange, blue
5) In this configuration the mapping should be:
dpm pgp
3 3
5 3
X X
1 3
X X
X X
5 1
X X
4 2
5 0
4 0
4 1
Server machines: clonfarm1, clon10new
TiCrate: hdice1
COBS: dpm0 : dpm6, dpm7, dtm1
dpm8 : dpm14, dpm15, dtm2
Connect to clonfarm1 as clasrun and change to bash shell
SDK software installation to talk to the atca crate
source /usr/clas12/release/1.4.0/slac_svt_new/V3.4.0/i86-linux-64/tools/envs-sdk.sh examples: cob_dump --all atca1 => dumps the RCE status (booted is 0xea) cob_rce_reset atca1 ==> resets all the RCEs cob_rce_reset atca1/1/0/2 ==> resets a particular dpm (in this case dpm1) cob_cold_data_reset atca1 ==> "power cycles" the RCEs (sometimes they do not come back up nicely so rce_reset might be needed after)
SVT software is installed in
/data/hps/slac_svt/ => diskless is exported to the cobs via nfs (server hosted on clonfarm1, NFS has to be v2 !! important ) => daq is exported to the cobs via nfs. Has been compiled on DTM1 => server contains the current software installation
On clonfarm1 start the rogue server
source /usr/clas12/release/1.4.0/slac_svt_new/anaconda3/etc/profile.d/conda.sh conda activate rogue_5.9.3 cd /data/hps/slac_svt/server/heavy-photon-daq/software/scripts/ python SvtCodaRun.py --local --env JLAB
On clon10new start the dummy server
python SvtCodaDummy.py --local --env JLAB
We are using conda release 1.4.0, installed in
/usr/clas12/release/1.4.0 HPS trigger configurations are stored in /usr/clas12/release/1.4.0/parms/trigger/HPS/
If event_builder or rogue_lite change, coda libraries need to be updates. Just run
sh /data/hps/slac_svt/copy_libraries.sh
Error:
setHeartError: 0 >sys 0, mask 21<
WARN: HeartBeat[0]: heartbeat=1628785839(1628785839) heartmask=21
UDP_cancel: cancel >inf:clonfarm1 sys 0, mask 21<
UDP_cancel: cancel >inf:clonfarm1 sys 0, mask 21<
TcpServer(external): wait: request in progress
Effect:
rol BUSY
Causes: multiple. Suggest to add printouts around rogueCodaTrigger and rogueCodaEvent in the hpssvt1.c ROL
This is in the event builder and indicates that we are dropping packets.
This is raised in RssiContributor.cc in the event_builder by acceptFrame thread