Misha
Misha is a cluster intended for use on projects associated with the Wu Tsai Institute, an interdisciplinary research endeavor at Yale University connecting neuroscience and data science to accelerate breakthroughs in understanding cognition.
Misha is named for Dr. Misha Mahowald, an American computational neuroscientist in the neuromorphic engineering, known for her work on the silicon retina.
Beta
Misha is currently in closed beta. For access, please contact Ping Luo (ping.luo@yale.edu).
Access the Cluster
Once you have an account, the cluster can be accessed via ssh or through the Open OnDemand web portal.
System Status and Monitoring
For system status messages and the schedule for upcoming maintenance, please see the system status page.
Partitions and Hardware
Misha is made up of several kinds of compute nodes. We group them into (sometimes overlapping) Slurm partitions meant to serve different purposes. By combining the --partition
and --constraint
Slurm options you can more finely control what nodes your jobs can run on.
Job Submission Limits
-
You are limited to 4 interactive app instances (of any type) at one time. Additional instances will be rejected until you delete older open instances. For OnDemand jobs, closing the window does not terminate the interactive app job. To terminate the job, click the "Delete" button in your "My Interactive Apps" page in the web portal.
-
Job submissions are limited to 200 jobs per hour. See the Rate Limits section in the Common Job Failures page for more info.
Public Partitions
See each tab below for more information about the available common use partitions.
Use the day partition for most batch jobs. This is the default if you don't specify one with --partition
.
Request Defaults
Unless specified, your jobs will run with the following options to salloc
and sbatch
options for this partition.
--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120
Job Limits
Jobs submitted to the day partition are subject to the following limits:
Limit | Value |
---|---|
Maximum job time limit | 1-00:00:00 |
Maximum CPUs per group | 384 |
Maximum memory per group | 3840G |
Maximum CPUs per user | 128 |
Maximum memory per user | 1280G |
Available Compute Nodes
Requests for --cpus-per-task
and --mem
can't exceed what is available on a single compute node.
Count | CPU Type | CPUs/Node | Memory/Node (GiB) | Node Features |
---|---|---|---|---|
18 | 6458 | 64 | 480 | sapphirerapids, avx512, 6458q, common |
Use the devel partition to jobs with which you need ongoing interaction. For example, exploratory analyses or debugging compilation.
Request Defaults
Unless specified, your jobs will run with the following options to salloc
and sbatch
options for this partition.
--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120
Job Limits
Jobs submitted to the devel partition are subject to the following limits:
Limit | Value |
---|---|
Maximum job time limit | 06:00:00 |
Maximum CPUs per user | 4 |
Maximum memory per user | 32G |
Available Compute Nodes
Requests for --cpus-per-task
and --mem
can't exceed what is available on a single compute node.
Count | CPU Type | CPUs/Node | Memory/Node (GiB) | Node Features |
---|---|---|---|---|
2 | 6458 | 64 | 480 | sapphirerapids, avx512, 6458q, common |
Use the week partition for jobs that need a longer runtime than day allows.
Request Defaults
Unless specified, your jobs will run with the following options to salloc
and sbatch
options for this partition.
--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120
Job Limits
Jobs submitted to the week partition are subject to the following limits:
Limit | Value |
---|---|
Maximum job time limit | 7-00:00:00 |
Maximum CPUs per group | 192 |
Maximum memory per group | 1920G |
Maximum CPUs per user | 32 |
Maximum memory per user | 320G |
Available Compute Nodes
Requests for --cpus-per-task
and --mem
can't exceed what is available on a single compute node.
Count | CPU Type | CPUs/Node | Memory/Node (GiB) | Node Features |
---|---|---|---|---|
6 | 6458 | 64 | 480 | sapphirerapids, avx512, 6458q, common |
Use the gpu partition for jobs that make use of GPUs. You must request GPUs explicitly with the --gpus
option in order to use them. For example, --gpus=gtx1080ti:2
would request 2 GeForce GTX 1080Ti GPUs per node.
Request Defaults
Unless specified, your jobs will run with the following options to salloc
and sbatch
options for this partition.
--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120
GPU jobs need GPUs!
Jobs submitted to this partition do not request a GPU by default. You must request one with the --gpus
option.
Job Limits
Jobs submitted to the gpu partition are subject to the following limits:
Limit | Value |
---|---|
Maximum job time limit | 2-00:00:00 |
Maximum CPUs per group | 192 |
Maximum GPUs per group | 24 |
Maximum CPUs per user | 96 |
Maximum GPUs per user | 12 |
Available Compute Nodes
Requests for --cpus-per-task
and --mem
can't exceed what is available on a single compute node.
Count | CPU Type | CPUs/Node | Memory/Node (GiB) | GPU Type | GPUs/Node | vRAM/GPU (GB) | Node Features |
---|---|---|---|---|---|---|---|
6 | 6326 | 32 | 975 | a40 | 4 | 48 | icelake, avx512, 6326, doubleprecision, a40, common |
5 | 6326 | 32 | 1001 | a100 | 4 | 80 | icelake, avx512, 6326, doubleprecision, a100-80g, common |
Use the gpu_devel partition to debug jobs that make use of GPUs, or to develop GPU-enabled code.
Request Defaults
Unless specified, your jobs will run with the following options to salloc
and sbatch
options for this partition.
--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120
GPU jobs need GPUs!
Jobs submitted to this partition do not request a GPU by default. You must request one with the --gpus
option.
Job Limits
Jobs submitted to the gpu_devel partition are subject to the following limits:
Limit | Value |
---|---|
Maximum job time limit | 06:00:00 |
Maximum CPUs per user | 10 |
Maximum GPUs per user | 2 |
Maximum memory per user | 70G |
Available Compute Nodes
Requests for --cpus-per-task
and --mem
can't exceed what is available on a single compute node.
Count | CPU Type | CPUs/Node | Memory/Node (GiB) | GPU Type | GPUs/Node | vRAM/GPU (GB) | Node Features |
---|---|---|---|---|---|---|---|
1 | 6326 | 32 | 1001 | icelake, avx512, 6326, doubleprecision, a100-80g-MIG, common |
Use the bigmem partition for jobs that have memory requirements other partitions can't handle.
Request Defaults
Unless specified, your jobs will run with the following options to salloc
and sbatch
options for this partition.
--time=01:00:00 --nodes=1 --ntasks=1 --cpus-per-task=1 --mem-per-cpu=5120
Job Limits
Jobs submitted to the bigmem partition are subject to the following limits:
Limit | Value |
---|---|
Maximum job time limit | 1-00:00:00 |
Maximum CPUs per user | 64 |
Maximum memory per user | 2T |
Available Compute Nodes
Requests for --cpus-per-task
and --mem
can't exceed what is available on a single compute node.
Count | CPU Type | CPUs/Node | Memory/Node (GiB) | Node Features |
---|---|---|---|---|
2 | 6458 | 64 | 1992 | sapphirerapids, avx512, 6458q, common |
Storage
/gpfs/radev
is Misha's filesystem where home, project, and scratch directories are located. For more details on the different storage spaces, see our Cluster Storage documentation.
You can check your current storage usage & limits by running the getquota
command. Note that the per-user usage breakdown only update once daily.
For information on data recovery, see the Backups and Snapshots documentation.
Warning
Files stored in scratch
are purged if they are older than 60 days. You will receive an email alert one week before they are deleted. Artificial extension of scratch file expiration is forbidden without explicit approval from the YCRC. Please purchase storage if you need additional longer term storage.
Partition | Root Directory | Storage | File Count | Backups | Snapshots |
---|---|---|---|---|---|
home | /gpfs/radev/home |
125GiB/user | 500,000 | Not yet | >=2 days |
project | /gpfs/radev/project |
1TiB/group, increase to 4TiB on request | 5,000,000 | No | >=2 days |
scratch | /gpfs/radev/scratch |
10TiB/group | 15,000,000 | No | No |