SLAC can provide a hosted instance of cryosparc that can make use of the slurm cluster that we have onsite.
You will need to make one small change due to our AFS file system before you will be able to submit jobs tho:
ssh into a slac machine (eg slacgpu.slac.stanford.edu) and run
cd ~${USER}/.cache
rm -rf pycuda
mkdir /gpfs/slac/cryo/fs1/u/${USER}/.cache/pycuda
ln -sf /gpfs/slac/cryo/fs1/u/${USER}/.cache/pycuda ~/.cache/pycuda
otherwise, you'll get a permissions error about the pycuda cache