LCLS users are responsible for complying with the data management and curation policies of their home institutions and funding agents and authorities. To enhance the scientific productivity of the LCLS user community, LCLS supplies on-site disk, tape and compute resources for prompt analysis of LCLS data, and software to access those resources consistent with the published data retention policy. Compute resources are preferentially allocated to recent and running experiments.
You will need a valid SLAC UNIX account in order to use the LCLS computing system. The instructions for getting a SLAC UNIX account are here.
You can get into the LCLS photon computing system by ssh'ing to:
pslogin pool you can then reach the analysis nodes (see below).
You can also consider connecting to the LCLS system using NX technology: this approach allows to preserve the status of a connection across multiple sessions and it provides better performance for displaying remote graphics.
Each control room has a number of nodes for local login. These nodes have access to the Internet and are named
The controls and DAQ nodes used for operating an instrument work in kiosk mode so you don't need a personal account to run an experiment from the control room. Remote access to these nodes is not allowed for normal users.
LCLS provides space for all your experiment's data at no cost for you. This includes the raw data from the detectors as well as the data derived from your analysis. Your raw data are available as XTC files or, on demand, as HDF5 files. The path to the experimental data is:
You can use the
psexport pool for copying your data:
The recommended tools for exporting the data offsite are
bbcp and Globus. The former,
bbcp, is slightly simpler to setup. On the other hand Globus is more efficient when transferring large amount of data because it babysits the overall process by, for example, automatically restarting a failed or stalled transfer. The performance of the two tools is very similar.
All control rooms and the overflow room in FEH have one or more taps on the Visitor Data Network. These taps can be used to transfer data to a laptop or a storage device. These devices will be automatically assigned an IP address through DHCP as soon as they are connected to the network tap.
The analysis framework is documented in the Data Analysis page. This section describes the resources available for running the analysis.
The following figure shows a logic diagram of the LCLS data flow and indicates the different stages where data analysis can be performed in LCLS:
In order to get access to the interactive nodes, connect to
psana pool is currently made of 16 servers with the following specifications:
Each psana interactive nodes has a GPU accelerator.
There are a number of batch farms (i.e. collections of compute nodes) located in the central computing building and in the NEH and FEH experimental halls. Instructions describing how to submit jobs can be found on the Submitting Batch Job page.
The main batch farm currently consist of 80 nodes with the following general specifications:
In addition, each experimental hall provides 20 nodes with access to the fast feedback storage layer (FFB). These nodes provide the resources for the high priority queues for the running experiments and have the following specifications:
When performing analysis on the psana interactive nodes, it is useful to display plots on your host machine. For host machines near SLAC, using ssh with X-windows forwarding (the -X or -Y options) suffices. X windows forwarding can get slow for host machines in Europe. Some users have found better performance with technology called nomachine, this is documented on the Remote Visualization page.
There are various ways to get real time information about the data acquired by an LCLS experiment. Users should be aware of the different possibilities and choose the approach that works best for their experiment. The methods for doing prompt analysis are described here.