Overview#
Hardware Resources#
Octopus
is a mixed architecture Intel/AMD Beowulf cluster with the
following specifications:
- 880 cores
376 AMD EPYC 7551p vCPUs
96 Intel Xeon E5-2695 v4 vCPUs
288 Intel Xeon E5-2665 physical cores
48 Intel Xeon E5-2643 v2 vCPUs
72 misellaneous CPUs and vCPUs for management / storage and other tasks
3.0 TB main memory
11 x Nvidia V100 PCI-E GPUs
8 x Nvidia GK110GL Tesla K20m GPUs
10 Gbit/s CISCO interconnect used for storage and computing
40 Gbit/s Infiniband interconnect QLogic 12200 InfiniBand QDR switch switch
100 TB shared storage and scratch space

Operating system#
All the nodes of Octopus
run Linux (CentOS 7).
The following types of jobs can be run on the cluster:
batch jobs (no user interaction)
GPU jobs (e.g scientific computing using GPGPUs or deep learning)
memory intensive jobs (up to 256GB RAM on a single machine available as a SMP host)
IO intensive jobs using the scratch partition (e.g several TB processing per job)
Interactive Jupyer jobs running on the compute hosts
Fully interactive desktop environment running on a compute node
Scheduler#
The scheduler used in Octopus
is open source SLURM
For more information on using the scheduler please consult the SLURM cheatsheet
Partitions#
There list below summarizes the main partitions:
normal
: 12 hosts with 16 vCPUs each with 64GB RAM.
gpu
: 12 hosts ( 8 wiht Nvidia K20m cards and 4 with Nvidia V100 cards).
large
: 4 hosts with 64 cores each and 256 GB RAM.
arza
: 18 hosts with 16 cores each and 64 GB RAM connected with an Infiniband network.
medium
: 5 hosts with 12 cores each and 24 GB RAM.
These partitions are broken down into smaller partitions with different time limits and resource limits and hardware accelerators.
Partition name |
Timelimit |
Nodes |
cores |
Memory |
Accelerators |
---|---|---|---|---|---|
normal |
1-00:00:00 |
12 |
16 |
64000 |
|
medium |
1-00:00:00 |
4 |
12 |
20000 |
|
gpu |
6:00:00 |
8 |
16 |
64000 |
Nvidia K20m |
gpu |
6:00:00 |
4 |
8 |
32000-128000 |
Nvidia V100 |
msfea-ai |
3-00:00:00 |
2 |
8 |
32000-128000 |
Nvidia V100 |
cmps-ai |
3-00:00:00 |
2 |
8 |
32000-128000 |
Nvidia V100 |
physics |
1-00:00:00 |
4 |
64 |
256000 |
|
large |
1-00:00:00 |
4 |
64 |
256000 |
|
arza |
1-00:00:00 |
8 |
16 |
64000 |
Nvidia K20m |
arza |
1-00:00:00 |
8 |
16 |
64000 |
|
interactive |
2:00:00 |
8 |
16 |
64000 |
Nvidia K20m |
interactive |
2:00:00 |
23 |
12+ |
20000+ |
|
interactive-gpu |
2:00:00 |
8 |
16 |
64000 |
Nvidia K20m |
interactive-gpu |
2:00:00 |
4 |
8 |
32000-128000 |
Nvidia V100 |
cudadev |
3:00:00 |
8 |
16 |
64000 |
Nvidia K20m |
cudadev |
3:00:00 |
4 |
8 |
32000-128000 |
Nvidia V100 |
For more information on using the paritions with the information on the resources and time limits please consult the hosts and partitions section.
Storage#
All the hosts’ mount the /home
directory and the /apps
directory. The quota
of the home directory is set to 25 GB. The /home
directory is backed up regularly.
For larger storage space the /scratch
partition can be used that has a quota 1 TB
per user. The maximum number of files that can be owned by a user is 1,000,000.