You are viewing an old version of this page. View the current version.

Compare with Current View Page History

Version 1 Next »

The XPP large detector is 4Mpx and 25kHz.  From Vincent: Two multimode MPO 24 fiber bundles per megapixel (8 bundles total) to the MM to SM conversion box.
This means 8x24=192 fibers (96 pairs) which is 2 full SM-MM conversion box. The SparkPix scales similarly with the number of pixels, so it only needs 24 fibers (1x MPO 24).

For TXI (2Mpx 5kHz) based on the epixhremu study from ric/stefano/valerio/mikhail/mona it feels like there is a good chance we can do it on 20 nodes with CPU (each with ~50 cores).  The XPP detector has 10x that data volume, and 200 nodes feels too difficult.  My best guess is that we should try to target GPUs for this to reduce the node count (ideally to 24 nodes, each of which would take 1 LR4 multi-color fiber).

If using 1 LR4 fiber per node isn't doable (i.e. we need more nodes) we should still use LR4 to reduce fiber using in the BOS and use a SM-MM box to split out the LR4 into more PLR4 fibers.

To-do items:

  • test common-mode speed on gpu (seshu has demonstrated Mikhail’s calibration formula without common-mode will be fast on a GPU)
  • we should meet with XPP scientists to understand what data-reduction algorithms are needed for their hutch
  • if libsz is one of the algorithms, we should understand its performance on gpu
  • benchmark other data-reduction algorithms on gpu
  • consider multiple options for algorithm implementation: cupy, cunumeric, cuda kernels
  • talk to TID engineers about dma'ing the kcu1500 data directly to gpu
  • No labels