DEPRECATION NOTICE: The on-ramp codelabs are currently in the process of being deprecated in favor of content in the user manual. If anything seems out of date in this content, please refer to the user manual content on the CRC's wrapper scripts.
CRC provides a few helper scripts that are intended to make the user experience simpler. These include
Checking disk quota with crc-quota
[username@login1 ~]$ crc-quota User: 'username' -> ihome: 70.11 GB / 75.0 GB Group: 'groupname' -> bgfs: 35.91 GB / 5.0 TB
View your group's SLURM allocation details with crc-usage
. Using this command alone will show the details for your user account's default allocation.
[nlc60@login1 ~] : crc-usage Please enter your CRC login password: +-----------------------------------------------------------------------------+ | Resource Allocation Request Information for 'sam' | +---------+--------------------------------------------+----------------------+ | ID | TITLE | EXPIRATION DATE | +---------+--------------------------------------------+----------------------+ | 35544 | Resource Allocation Request for sam | 2025-05-01 | +---------+--------------------------------------------+----------------------+ | | CLUSTER | SERVICE UNITS | | | ---- | ---- | | | smp | 100000 | | | mpi | 100000 | | | htc | 100000 | | | gpu | 100000 | | | | | +---------+--------------------------------------------+----------------------+ +-----------------------------------------------------------------------------+ | Summary of Usage Across All Clusters | +-------+-------------------------+-----------------------+-------------------+ | smp | TOTAL USED: 1279 | AWARDED: 100000 | % USED: 1 | +-------+-------------------------+-----------------------+-------------------+ | | USER | USED | % USED | | | ---- | ---- | ---- | | | chx33 | 1192 | 1 | | | leb140 | 42 | <1 | | | yak73 | 28 | <1 | | | kimwong | 16 | <1 | | | nlc60 | 0 | <1 | | | | | | +-------+-------------------------+-----------------------+-------------------+ | mpi | TOTAL USED: 45313 | AWARDED: 100000 | % USED: 45 | +-------+-------------------------+-----------------------+-------------------+ | | USER | USED | % USED | | | ---- | ---- | ---- | | | leb140 | 44799 | 44 | | | chx33 | 353 | <1 | | | yak73 | 160 | <1 | | | kimwong | 2 | <1 | | | | | | +-------+-------------------------+-----------------------+-------------------+ | htc | TOTAL USED: 931 | AWARDED: 100000 | % USED: 0 | +-------+-------------------------+-----------------------+-------------------+ | | USER | USED | % USED | | | ---- | ---- | ---- | | | kimwong | 621 | <1 | | | yak73 | 218 | <1 | | | fangping | 49 | <1 | | | chx33 | 41 | <1 | | | djp81 | 1 | <1 | | | leb140 | 1 | <1 | | | clcgenomics | 0 | <1 | | | | | | +-------+-------------------------+-----------------------+-------------------+ | gpu | TOTAL USED: 33257 | AWARDED: 100000 | % USED: 33 | +-------+-------------------------+-----------------------+-------------------+ | | USER | USED | % USED | | | ---- | ---- | ---- | | | yak73 | 32464 | 32 | | | leb140 | 712 | <1 | | | chx33 | 74 | <1 | | | kimwong | 4 | <1 | | | fangping | 3 | <1 | | | | | | +-------+-------------------------+-----------------------+-------------------+
Look for available resources across the clusters with crc-idle
.
[username@login1 ~]$ crc-idle Cluster: smp, Partition: smp ============================ 2 nodes w/ 1 idle cores 5 nodes w/ 2 idle cores 1 nodes w/ 3 idle cores 9 nodes w/ 4 idle cores 2 nodes w/ 5 idle cores 11 nodes w/ 6 idle cores 30 nodes w/ 7 idle cores 35 nodes w/ 8 idle cores 1 nodes w/ 9 idle cores 11 nodes w/ 12 idle cores 4 nodes w/ 15 idle cores 1 nodes w/ 16 idle cores 1 nodes w/ 18 idle cores 1 nodes w/ 21 idle cores 1 nodes w/ 22 idle cores 20 nodes w/ 23 idle cores Cluster: smp, Partition: high-mem ================================= 6 nodes w/ 8 idle cores 2 nodes w/ 12 idle cores Cluster: smp, Partition: legacy =============================== 1 nodes w/ 1 idle cores 1 nodes w/ 8 idle cores Cluster: gpu, Partition: gtx1080 ================================ 3 nodes w/ 1 idle GPUs 1 nodes w/ 2 idle GPUs 4 nodes w/ 3 idle GPUs 4 nodes w/ 4 idle GPUs Cluster: gpu, Partition: titanx =============================== 1 nodes w/ 1 idle GPUs 1 nodes w/ 2 idle GPUs 1 nodes w/ 3 idle GPUs 3 nodes w/ 4 idle GPUs Cluster: gpu, Partition: k40 ============================ 1 nodes w/ 2 idle GPUs Cluster: gpu, Partition: v100 ============================= No idle GPUs Cluster: mpi, Partition: opa ============================ No idle cores Cluster: mpi, Partition: opa-high-mem ===================================== No idle cores Cluster: mpi, Partition: ib =========================== 14 nodes w/ 20 idle cores Cluster: htc, Partition: htc ============================ 2 nodes w/ 2 idle cores 1 nodes w/ 5 idle cores 1 nodes w/ 6 idle cores 1 nodes w/ 10 idle cores 3 nodes w/ 11 idle cores 1 nodes w/ 12 idle cores 20 nodes w/ 16 idle cores 4 nodes w/ 24 idle cores 1 nodes w/ 25 idle cores 1 nodes w/ 37 idle cores 5 nodes w/ 48 idle cores
You can request an interactive session on a compute-node with crc-interactive
[username@login1 ~]$ crc-interactive --help crc-interactive -- An interactive Slurm helper Usage: crc-interactive (-s | -g | -m | -i | -d) [-hvzo] [-t <time>] [-n <num-nodes>] [-p <partition>] [-c <num-cores>] [-u <num-gpus>] [-r <res-name>] [-b <memory>] [-a <account>] [-l <license>] [-f <feature>] Positional Arguments: -s --smp Interactive job on smp cluster -g --gpu Interactive job on gpu cluster -m --mpi Interactive job on mpi cluster -i --invest Interactive job on invest cluster -d --htc Interactive job on htc cluster Options: -h --help Print this screen and exit -v --version Print the version of crc-interactive -t --time <time> Run time in hours, 1 <= time <= 12 [default: 1] -n --num-nodes <num-nodes> Number of nodes [default: 1] -p --partition <partition> Specify non-default partition -c --num-cores <num-cores> Number of cores per node [default: 1] -u --num-gpus <num-gpus> Used with -g only, number of GPUs [default: 0] -r --reservation <res-name> Specify a reservation name -b --mem <memory> Memory in GB -a --account <account> Specify a non-default account -l --license <license> Specify a license -f --feature <feature> Specify a feature, e.g. `ti` for GPUs -z --print-command Simply print the command to be run -o --openmp Run using OpenMP style submission [username@login1 ~]$ crc-interactive -g -p titanx -n 1 -c 1 -u 1 -t 12 srun: job 260065 queued and waiting for resources srun: job 260065 has been allocated resources [username@gpu-stage06 ~]$ nvidia-smi Wed Jan 26 08:42:04 2022 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 460.32.03 Driver Version: 460.32.03 CUDA Version: 11.2 | |-------------------------------+----------------------+----------------------+ | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 GeForce GTX TIT... On | 00000000:02:00.0 Off | N/A | | 48% 82C P2 236W / 250W | 794MiB / 12212MiB | 99% Default | | | | N/A | +-------------------------------+----------------------+----------------------+ | 1 GeForce GTX TIT... On | 00000000:03:00.0 Off | N/A | | 22% 28C P8 16W / 250W | 1MiB / 12212MiB | 0% Default | | | | N/A | +-------------------------------+----------------------+----------------------+ | 2 GeForce GTX TIT... On | 00000000:81:00.0 Off | N/A | | 22% 28C P8 15W / 250W | 1MiB / 12212MiB | 0% Default | | | | N/A | +-------------------------------+----------------------+----------------------+ | 3 GeForce GTX TIT... On | 00000000:82:00.0 Off | N/A | | 22% 27C P8 14W / 250W | 1MiB / 12212MiB | 0% Default | | | | N/A | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | 0 N/A N/A 23796 C pmemd.cuda 790MiB | +-----------------------------------------------------------------------------+ [username@gpu-stage06 ~]$ exit exit [username@login1 ~]$
There are a few other helper scripts that you can see by typing crc-
followed two strokes of the tab
key
[username@login1 ~]$ crc- crc-idle crc-job-stats crc-quota crc-squeue crc-usage crc-interactive crc-proposal-end crc-scancel crc-sinfo crc-sus [username@login1 ~]$ crc-
crc-job-stats
- Include at the bottom of a job script, outputs the details of the job after completion
crc-squeue
- Show details for active jobs submitted by your user
crc-scancel
- Cancel a job by job ID from crc-squeue
crc-proposal-end
- Outputs the end date of your user's default allocation
crc-sus
- Show a concise view of the service units awarded on each cluster
crc-show-config
- Show configuration information about the partitions for each cluster
crc-sinfo
- Show the status of the partitions of each cluster and their nodes
The best way to get help on a specific issue is to submit a help ticket. You should log in to the CRC website using your Pitt credentials first.