Skip to content

Grace

Grace

The Grace cluster is is named for the computer scientist and United States Navy Rear Admiral Grace Murray Hopper, who received her Ph.D. in Mathematics from Yale in 1934.

Grace is a shared-use resource for the Faculty of Arts and Sciences (FAS). It consists of a variety of compute nodes networked over low-latency InfiniBand and mounts several shared filesystems.


Partitions and Hardware

Grace is made up of several kinds of compute nodes. We group them into (sometimes overlapping) Slurm partitions meant to serve different purposes. By combining the --partition and --constraint Slurm options you can more finely control what nodes your jobs can run on.

Public Partitions

See each tab below for more information about the available common use partitions.

Use the day partition for most batch jobs. This is the default if you don't specify one with --partition.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the day partition are subject to the following limits:

Limit Value
Max job time limit 1-00:00:00
Maximum CPUs per group 900
Maximum CPUs per user 640

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
135 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard
60 E5-2660_v2 20 121 ivybridge, E5-2660_v2, nogpu, standard, oldest
44 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard
72 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard

Use the interactive partition to jobs with which you need ongoing interaction. For example, exploratory analyses or debugging compilation.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the interactive partition are subject to the following limits:

Limit Value
Max job time limit 06:00:00
Maximum CPUs per user 4
Maximum memory per user 32G
Maximum running jobs per user 1
Maximum submitted jobs per user 1

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
2 E5-2660_v2 20 121 ivybridge, E5-2660_v2, nogpu, standard, oldest
1 6126 24 176 skylake, avx2, avx512, 6126, nogpu, standard

Use the week partition for jobs that need a longer runtime than day allows.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the week partition are subject to the following limits:

Limit Value
Max job time limit 7-00:00:00
Maximum CPUs per group 250
Maximum CPUs per user 100

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
36 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard
7 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard

Use the transfer partition to stage data for your jobs to and from cluster storage.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the transfer partition are subject to the following limits:

Limit Value
Max job time limit 1-00:00:00
Maximum running jobs per user 2
Maximum CPUs per job 1

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
2 E5-2660_v2 20 121 ivybridge, E5-2660_v2, nogpu, standard, oldest

Use the gpu partition for jobs that make use of GPUs. You must request GPUs explicitly with the --gres option in order to use them. For example, --gres=gpu:gtx1080ti:2 would request 2 GeForce GTX 1080Ti GPUs per node.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the gpu partition are subject to the following limits:

Limit Value
Max job time limit 2-00:00:00
Maximum GPUs per user 24

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
2 6136 24 90 v100 2 16 skylake, avx2, avx512, 6136, doubleprecision
5 E5-2660_v3 20 121 k80 4 12 haswell, avx2, E5-2660_v3, doubleprecision
6 E5-2660_v4 28 247 p100 1 16 broadwell, avx2, E5-2660_v4, doubleprecision
4 6240 36 372 v100 4 16 cascadelake, avx2, avx512, 6240, doubleprecision
5 6240 36 183 rtx2080ti 4 11 cascadelake, avx2, avx512, 6240, singleprecision

Use the gpu_devel partition to debug jobs that make use of GPUs, or to develop GPU-enabled code.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the gpu_devel partition are subject to the following limits:

Limit Value
Max job time limit 04:00:00
Maximum CPUs per user 10
Maximum submitted jobs per user 1

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
1 6240 36 372 v100 4 16 cascadelake, avx2, avx512, 6240, doubleprecision
1 E5-2660_v3 20 121 k80 4 12 haswell, avx2, E5-2660_v3, doubleprecision

Use the bigmem partition for jobs that have memory requirements other partitions can't handle.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the bigmem partition are subject to the following limits:

Limit Value
Max job time limit 1-00:00:00
Maximum CPUs per user 40
Maximum memory per user 1500G

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
2 E7-4820_v4 40 1507 broadwell, avx2, E7-4820_v4, nogpu
1 6240 36 1506 cascadelake, avx2, avx512, 6240, nogpu, standard
2 6240 36 1506 cascadelake, avx2, avx512, 6240, nogpu

Use the mpi partition for tightly-coupled parallel programs that make efficient use of multiple nodes. See our MPI documentation if your workload fits this description.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --exclusive --mem=92160

Job Limits

Jobs submitted to the mpi partition are subject to the following limits:

Limit Value
Max job time limit 1-00:00:00
Maximum nodes per group 48
Maximum nodes per user 32

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
120 6136 24 90 hdr, skylake, avx2, avx512, 6136, nogpu, standard

Use the scavenge partition to run preemptable jobs on more resources than normally allowed. For more information about scavenge, see the Scavenge documentation.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the scavenge partition are subject to the following limits:

Limit Value
Max job time limit 1-00:00:00
Maximum CPUs per user 10000

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
4 6240 36 372 v100 4 16 cascadelake, avx2, avx512, 6240, doubleprecision
2 6136 24 90 v100 2 16 skylake, avx2, avx512, 6136, doubleprecision
136 6136 24 90 hdr, skylake, avx2, avx512, 6136, nogpu, standard
16 6136 24 90 edr, skylake, avx2, avx512, 6136, nogpu, standard
3 6142 32 183 skylake, avx2, avx512, 6142, nogpu, standard
207 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard
8 6240 36 372 cascadelake, avx2, avx512, 6240, nogpu, standard
1 E7-4820_v2 32 1003 ivybridge, E7-4820_v2, nogpu
1 E7-4809_v3 32 2011 haswell, avx2, E7-4809_v3, nogpu
4 E7-4820_v4 40 1507 broadwell, avx2, E7-4820_v4, nogpu
78 E5-2660_v2 20 121 ivybridge, E5-2660_v2, nogpu, standard, oldest
136 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard
20 E5-2660_v3 20 247 haswell, avx2, E5-2660_v3, nogpu, standard
8 E5-2660_v3 20 247 k80 2 12 haswell, avx2, E5-2660_v3, doubleprecision
6 E5-2660_v3 20 121 k80 4 12 haswell, avx2, E5-2660_v3, doubleprecision
161 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard
7 E5-2660_v4 28 247 p100 1 16 broadwell, avx2, E5-2660_v4, doubleprecision
1 6136 24 751 skylake, avx2, avx512, 6136, nogpu
1 E5-2637_v4 8 121 gtx1080ti 4 11 broadwell, avx2, E5-2637_v4, singleprecision
9 6136 24 183 p100 4 16 skylake, avx2, avx512, 6136, doubleprecision
2 5122 8 183 rtx2080 4 8 skylake, avx2, avx512, 5122, singleprecision
1 6254 36 1506 rtx4000,rtx8000,v100 4,2,2 8,48,16 cascadelake, avx2, avx512, 6254
20 8260 96 183 cascadelake, avx2, avx512, 8260, nogpu
1 6240 36 1506 cascadelake, avx2, avx512, 6240, nogpu, standard
2 6240 36 1506 cascadelake, avx2, avx512, 6240, nogpu
7 6240 36 183 rtx2080ti 4 11 cascadelake, avx2, avx512, 6240, singleprecision

Use the scavenge_gpu partition to run preemptable jobs on more GPU resources than normally allowed. For more information about scavenge, see the Scavenge documentation.

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the scavenge_gpu partition are subject to the following limits:

Limit Value
Max job time limit 1-00:00:00
Maximum GPUs per user 30

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
2 6136 24 90 v100 2 16 skylake, avx2, avx512, 6136, doubleprecision
8 E5-2660_v3 20 247 k80 2 12 haswell, avx2, E5-2660_v3, doubleprecision
6 E5-2660_v3 20 121 k80 4 12 haswell, avx2, E5-2660_v3, doubleprecision
7 E5-2660_v4 28 247 p100 1 16 broadwell, avx2, E5-2660_v4, doubleprecision
1 E5-2637_v4 8 121 gtx1080ti 4 11 broadwell, avx2, E5-2637_v4, singleprecision
9 6136 24 183 p100 4 16 skylake, avx2, avx512, 6136, doubleprecision
2 5122 8 183 rtx2080 4 8 skylake, avx2, avx512, 5122, singleprecision
1 6254 36 1506 rtx4000,rtx8000,v100 4,2,2 8,48,16 cascadelake, avx2, avx512, 6254
4 6240 36 372 v100 4 16 cascadelake, avx2, avx512, 6240, doubleprecision
5 6240 36 183 rtx2080ti 4 11 cascadelake, avx2, avx512, 6240, singleprecision

Private Partitions

With few exceptions, jobs submitted to pi_ partitions are not considered when calculating your group's Fairshare. Your group can purchase additional hardware for private use, which we will make available as a pi_groupname partition. These nodes are purchased by you, but supported and administered by us. After vendor support expires, we retire compute nodes. Compute nodes can range from $10K to upwards of $50K depending on your requirements. If you are interested in purchasing nodes for your group, please contact us.

PI Partitions (click to expand)

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_altonji partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
2 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_anticevic partition are subject to the following limits:

Limit Value
Max job time limit 100-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
20 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard
16 E5-2660_v3 20 247 haswell, avx2, E5-2660_v3, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_anticevic_bigmem partition are subject to the following limits:

Limit Value
Max job time limit 100-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
1 E7-4809_v3 32 2011 haswell, avx2, E7-4809_v3, nogpu

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the pi_anticevic_gpu partition are subject to the following limits:

Limit Value
Max job time limit 100-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
8 E5-2660_v3 20 247 k80 2 12 haswell, avx2, E5-2660_v3, doubleprecision

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_anticevic_z partition are subject to the following limits:

Limit Value
Max job time limit 100-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
3 E5-2660_v3 20 247 haswell, avx2, E5-2660_v3, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_balou partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
9 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard
30 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_berry partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
1 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard
1 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_cowles partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00
Maximum CPUs per user 120
Maximum nodes per user 5

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
14 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_cowles_nopreempt partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00
Maximum CPUs per user 120
Maximum nodes per user 5

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
10 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_econ_io partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
6 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_econ_lp partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
5 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_esi partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00
Maximum CPUs per user 648

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
36 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=3840

Job Limits

Jobs submitted to the pi_fedorov partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
12 6136 24 90 hdr, skylake, avx2, avx512, 6136, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_gelernter partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
1 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard
1 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_gerstein partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
1 E7-4820_v2 32 1003 ivybridge, E7-4820_v2, nogpu
32 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_glahn partition are subject to the following limits:

Limit Value
Max job time limit 100-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
1 E5-2660_v3 20 247 haswell, avx2, E5-2660_v3, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=3840

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the pi_hammes_schiffer partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
16 6136 24 90 edr, skylake, avx2, avx512, 6136, nogpu, standard
1 6136 24 751 skylake, avx2, avx512, 6136, nogpu
1 E5-2637_v4 8 121 gtx1080ti 4 11 broadwell, avx2, E5-2637_v4, singleprecision
2 5122 8 183 rtx2080 4 8 skylake, avx2, avx512, 5122, singleprecision

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_hodgson partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
1 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_holland partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
2 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard
8 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_howard partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
1 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_jetz partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
2 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_kaminski partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
8 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the pi_lederman partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
1 6254 36 1506 rtx4000,rtx8000,v100 4,2,2 8,48,16 cascadelake, avx2, avx512, 6254

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=1952

Job Limits

Jobs submitted to the pi_levine partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
20 8260 96 183 cascadelake, avx2, avx512, 8260, nogpu

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=3840

Job Limits

Jobs submitted to the pi_lora partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
4 6136 24 90 hdr, skylake, avx2, avx512, 6136, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_mak partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
3 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the pi_manohar partition are subject to the following limits:

Limit Value
Max job time limit 180-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
2 E7-4820_v4 40 1507 broadwell, avx2, E7-4820_v4, nogpu
8 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard
1 E5-2660_v4 28 247 p100 1 16 broadwell, avx2, E5-2660_v4, doubleprecision
4 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the pi_ohern partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
3 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard
2 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard
16 E5-2660_v2 20 121 ivybridge, E5-2660_v2, nogpu, standard, oldest
9 6136 24 183 p100 4 16 skylake, avx2, avx512, 6136, doubleprecision

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_owen_miller partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
5 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

GPU jobs need GPUs!

Jobs submitted to this partition do not request a GPU by default. You must request one with the --gres option.

Job Limits

Jobs submitted to the pi_panda partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) GPU Type GPUs/Node vRAM/GPU (GB) Node Features
2 6240 36 183 rtx2080ti 4 11 cascadelake, avx2, avx512, 6240, singleprecision

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_poland partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
8 6240 36 372 cascadelake, avx2, avx512, 6240, nogpu, standard
10 E5-2660_v4 28 247 broadwell, avx2, E5-2660_v4, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_polimanti partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
1 6240 36 183 cascadelake, avx2, avx512, 6240, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_seto partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
3 6142 32 183 skylake, avx2, avx512, 6142, nogpu, standard

Request Defaults

Unless specified, your jobs will run with the following options to srun and sbatch options for this partition.

--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120

Job Limits

Jobs submitted to the pi_tsmith partition are subject to the following limits:

Limit Value
Max job time limit 28-00:00:00

Available Compute Nodes

Requests for --cpus-per-task and --mem can't exceed what is available on a single compute node.

Nodes CPU Type CPUs/Node Memory/Node (GiB) Node Features
1 E5-2660_v3 20 121 haswell, avx2, E5-2660_v3, nogpu, standard

Storage

Grace has access to a number of filesystems. /gpfs/loomis is Grace's primary filesystem where home, project, and scratch60 directories are located. For more details on the different storage spaces, see our Cluster Storage documentation.

You can check your current storage usage & limits by running the getquota command. Your ~/project and ~/scratch60 directories are shortcuts. Get a list of the absolute paths to your directories with the mydirectories command. If you want to share data in your Project or Scratch directory, see the permissions page.

Warning

Files stored in scratch60 are purged if they are older than 60 days. You will receive an email alert one week before they are deleted.

Partition Root Directory Storage File Count Backups
home /gpfs/loomis/home.grace 125GiB/user 500,000 Yes
project /gpfs/loomis/project 1TiB/group, increase to 4TiB on request 5,000,000 No
scratch60 /gpfs/loomis/scratch60 20TiB/group 15,000,000 No

Last update: August 5, 2020