Galileo100

Galileo100 is a new infrastructure co-funded by the European ICEI (Interactive Computing e-Infrastructure) project and engineered by DELL. It is the national Tier-1 system for scientific research and is available to the Italian public and industrial researchers since September 2021. It also features 77 cloud computing servers and was expanded in November 2022 with 82 additional nodes. Galileo100 is used for high-end technical and industrial HPC projects, as well as meteorology and environmental studies.

The specific guide for the Galileo100 cluster contains unique information that deviates from the general behavior described in the HPC Clusters sections.

Access to the System

The machine is reachable via ssh (secure Shell) protocol at hostname point: login.g100.cineca.it.

The connection is established, automatically, to one of the available login nodes. It is possible to connect to Galileo100 using one the specific login hostname points:

  • login01-ext.g100.cineca.it

  • login02-ext.g100.cineca.it

  • login03-ext.g100.cineca.it

Warning

The mandatory access to Galileo100 is the two-factor authetication (2FA). Get more information at section Access to the Systems.

System Architecture

Aggiungere qualcosa?

Hardware Details

Type

Specific

Models

Dual-soket Dell PowerEdge

Nodes

630

Processors/node

2xCPU x86 Intel Xeon Platinum 8276/L 2.4GHz

CPU/node

48

Accelerators/node

2xGPU Nvidia V100 PCIe3 with 32 GB Ram on 36 Viz Nodes

RAM/node

384 GiB (+ 3.0 TiB Optane on 180 fat nodes)

Peak Performance

2 PFlop/s (3.53 TFlop/s in single node)

Internal Network

Mellanox Infiniband 100GbE

Disks and Filesystems

The storage organization conforms to CINECA infrastructure. General information are reported in File Systems and Data Management section. In the following, only differences with respect to general behavior are listed and explained.

Job Managing and SLURM Partitions

Partition

QOS

#Cores per job

Walltime

Max jobs/resources per user

Max memory per node (MB)

Priority

Notes

g100_all_serial

(default)

noQOS

4 cores

04:00:00

4 cores

120 submitted jobs

31,200

(30 GB)

40

on two login nodes

budget free

g100_all_serial

(default)

qos_install

16 cores

04:00:00

16 cores

1 running job

100 GB

40

request to superc@cineca.it

g100_usr_dbg

noQOS

2 nodes

01:00:00

375,300

(366 GB)

40

g100_usr_dbg

qos_ind

Depending on the specific agreement

Depending on the specific agreement

375,300

(366 GB)

90

Partition dedicated to specific kinds of users.

g100_usr_prod

g100_usr_smem

g100_usr_pmem

noQOS

min = 1

max = 32 nodes

24:00:00

100 running jobs

120 submitted jobs

375,300

(366 GB)

40

runs on thin and persistent memory nodes

runs only on thin nodes

runs only on persistent memory nodes

g100_usr_prod

g100_usr_smem

g100_usr_pmem

g100_qos_bprod

min = 1537 (33 nodes)

max = 3072 (64 nodes)

24:00:00

100 running jobs

120 submitted jobs

375,300

(366 GB)

60

runs on thin and persistent memory nodes

runs only on thin nodes

runs only on persistent memory nodes

g100_usr_prod

g100_usr_smem

g100_usr_pmem

g100_qos_lprod

min = 1

max = 2 nodes

4-00:00:00

2 nodes

100 running jobs

120 submitted jobs

375,300

(366 GB)

40

runs on thin and persistent memory nodes

runs only on thin nodes

runs only on persistent memory nodes

g100_usr_prod

g100_usr_smem

g100_usr_pmem

qos_special

> 32 nodes

> 24:00:00

375,300

(366 GB)

40

request to superc@cineca.it

g100_usr_bmem

noQOS

25 nodes

24:00:00

100 running jobs

120 submitted jobs

3,036,000

(3 TB)

40

runs on fat nodes

g100_usr_interactive

noQOS

max = 0.5 node

8:00:00

100 running jobs

120 submitted jobs

375,300

(366 GB)

40

on nodes with GPUs

–gres=gpu:N (N=1)

g100_meteo_prod

qos_meteo

24:00:00

375,300

(366 GB)

40

Partition reserved to meteo services, NOT opened to production.

Runs on thin nodes

Dedicated Services

Interactive Computing

Galileo 100 resources are also accessible via web browser on a Jupyter-based interface at the following link: https://jupyter.g100.cineca.it/

Further details are reported at the following Interactive Computing. Please note that the service is considered in pre-production, thus the resources are not accounted from the budget and the service is provided with no warranty.