High Energy Physics CMS Tier-2 Facilities#
Overview of Resources#
Access to the computing cluster is available through SSH:#
Batch Processing using Condor#
HTCondor may be used directly for running CMS jobs on our cluster. Once submitted, the jobs flock all over the UW campus grid. In particular, the jobs run on Grid Laboratory Of Wisconsin (GLOW) or Wisconsin’s Center for High Throughput Computing (CHTC) cluster. #
Abridged instructions for CMS Users#
The batch jobs are started by submitting a script to condor: condor_submit
#
You can watch the progress of your jobs using: condor_q -nobatch
#
You can check the status of the compute pools using:#
condor_status -pool condor.hep.wisc.edu
condor_status -pool cm.chtc.wisc.edu
You can ssh to a job to debug it interactively using:#
condor_ssh_to_job *jobid-here*
Accessing files from your HTCondor job#
A job can read files in this way: Reading files.#
A CMSSW program cannot write to HDFS directly. Instead, it must write to the local working directory. The script that executes the program must then copy the output file HDFS as described here.#
“farmout” shell script for submitting a bunch of CMSSW jobs#
You may use my sample shell script called “farmout” that submits CMSSW jobs to fully process a dataset. This automatically submits the jobs from an appropriate /scratch directory where logs will accumulate. The data files are copied to HDFS.#
Example simulation submission:#
farmoutRandomSeedJobs \
dataset-name \
total-events \
events-per-job \
/path/to/CMSSW \
/path/to/configTemplate.py
Example analysis submission:#
farmoutAnalysisJobs \
jobName \
/path/to/CMSSW \
/path/to/configTemplate.py
Use the –help option to see the options used by these scripts.#
Your data files will be stored in HDFS under /hdfs/store/user/username. See the FAQ for more information on how to manage your files, how to use farmout, and how to use Condor.#