Confluence will be unusable 23-July-2024 at 06:00 due to a Crowd upgrade.
In 2021 LCLS switch to the SLURM batch system.
Information on submitting jobs to the SLURM system at LCLS can be found on this page: Submitting SLURM Batch Jobs
Information on the Automatic Run Processing system (ARP) can be found on this page: Automatic Run Processing (ARP). This is also usable at sites like NERSC and SDF.
A "cheat sheet" showing similar commands on LSF and SLURM can be found here: https://slurm.schedmd.com/rosetta.pdf
Refer to the table below for the batch resources available in psana.
Table of Contents |
---|
Depending on your data access you may need to submit jobs to a specific farm. This is accomplished by submitting to the appropriate LSF batch queue. Refer to the table below. Jobs for the current experiment should be submitted to the high priority queues psnehhiprioq and psfehhiprioq running against the Fast Feedback storage layer (FFB) located at /reg/d/ffb/<hutch>/<experiment> as shown HERE. Jobs for the off-shift experiment should be submitted to psnehprioq and psfehprioq. Only psneh(hi)prioq/psfeh(hi)prioq should access the FFB. When in doubt, use psanaq.
Submit your job from an interactive node (where you land after doing ssh psana). LSF will run the submitted job on the specified queue using nodes listed in the table below. All nodes in the queues listed below run rhel7RHEL7. By submitting from an interactive node (, also running rhel7)RHEL7, you will ensure that your job inherits a rhel7 environment.
RHEL7 environment.
Note 1: Jobs for the current experiment can be submitted to fast feedback (FFB) queues, which allocate resources for the most recent experiments. The FFB queues in the tables below are for LCLS-II experiments (TMO, RIX and UED). The FEH experiments (LCLS-I, including XPP) can submit FFB jobs to the new Fast Feedback System.
Warning |
---|
As of February 2023, the offline compute resources have been consolidated into the psanaq. The priority queues have been removed. |
sQueue name | Node names on SLURM queues | Number of Nodes | Comments | Throughput | Cores/ Node | RAM [GB/node] | Time limit |
---|---|---|---|---|---|---|---|
psanaq | psana15xx | 34 | Primary psana queue | 40 | 16 | 128 | 48hrs |
psanagpuq | psanagpu113-psanagpu118 | 6 | GPU nodes | 10 |
Location
Queue
Nodes
Data
Comments
Throughput
(Gbit/s)
RAM
(GB/node)
Default
Time
Limit
NEH
psnehhiprioq
psana15xx
FFB for AMO, SXR, XPP
Current NEH experiment on FFB ONLY
psnehprioq
psana15xx
FFB for AMO, SXR, XPP
Off-shift NEH experiment on FFB ONLY
psnehq
psana15xx
Jobs preemptable by psneh(hi)prioq
FEH
psfehhiprioq
psana16xx
FFB for XCS, CXI, MEC
Current FEH experiment on FFB ONLY
psfehprioq
psana16xx
FFB for XCS, CXI, MEC
Off-shift FEH experiment on FFB ONLY
psfehq
psana16xx
Jobs preemptable by psfeh(hi)prioq
16 | 128 | 48hrs |