Apply to this round for a project on the Berzelius SuperPOD at NSC. Berzelius is not a NAISS system. It is a direct donation to NSC from the Knut and Alice Wallenberg foundation (KAW). Projects with financial support from KAW should have priority according to the donation letter.
To be granted resources on Berzelius, both the Principal Investigator (PI) — defined here as the main applicant in SUPR — and the project must meet the conditions stated
here.
Deadlines and Decisions
Handling of projects requesting a default allocation:
Proposals requesting a default allocation will normally be handled within a week.
Handling of proposals requesting more than a default allocation:
Monthly evaluation of proposals during the year.
Proposals submitted at the latest on the 15th will undergo review during the same month.
July and December have different schedules.
Proposals submitted after June 15 will be processed in August. i.e. with a first possible allocation starting September 1st for larger than default requests!
This round is open for proposals until 2026-01-01 00:00.
|
Resource |
Centre |
Upper Limit |
Default Storage |
Available |
Unit |
Note |
|
|
Berzelius Ampere |
NSC |
28 800 |
|
345 600 |
GPU-h/month |
Formerly known as "Berzelius Compute". Applications are normally evaluated during the last week each month
|
|
Submit your proposal at least one week before the end of a month to be considered for an allocation from the first of the following month. Received proposals will be evaluated against each other and time that become available as project ends at the end of a month will be allocated to the proposed projects accordingly.
Berzelius Ampere is an NVIDIA® SuperPOD consisting of 94 NVIDIA® DGX-A100 compute nodes supplied by
Atos/Eviden and 8 CPU nodes also supplied by Eviden. The original 60 "thin" DGX-A100 nodes are each equipped with 8 NVIDIA® A100 Tensor Core GPUs, 2 AMD Epyc™ 7742 CPUs, 1 TB RAM and 15 TB of local NVMe SSD storage. The A100 GPUs have 40 GB on-board HBM2 VRAM. The 34 newer DGX-A100 nodes "fat" are each equipped with 8 NVIDIA® A100 Tensor
Core GPUs, 2 AMD Epyc™ 7742 CPUs, 2 TB RAM and 30 TB of local NVMe SSD storage. The A100 GPUs have 80 GB on-board HBM2 VRAM. The CPU nodes are each equipped with 2 AMD Epyc™ 9534 CPUs, 1.1 TB RAM and 6.4 TB of local NVMe SSD storage.
Fast compute interconnect is provided via 8x NVIDIA® Mellanox® HDR per DGX
connected in a non-blocking fat-tree topology. In addition, every node is
equipped with NVIDIA® Mellanox® HDR dedicated storage interconnect.
All nodes have a local disk where applications can store temporary files. The
size of this disk (available to jobs as `/scratch/local`) is 15 TB on "thin"
nodes, 30 TB on "fat" nodes, and 6.4 TB on CPU nodes, and is shared between
all jobs using the node.
|
|
|
Berzelius Hopper |
NSC |
11 520 |
|
92 160 |
GPU-h/month |
Applications are normally evaluated during the last week each month
|
|
Submit your proposal at least one week before the end of a month to be considered for an allocation from the first of the following month. Received proposals will be evaluated against each other and time that become available as project ends at the end of a month will be allocated to the proposed projects accordingly.
The latest phase of the Berzelius service is Berzelius Hopper. Berzelius Hopper consist of 16 NVIDIA® DGX-H200 compute nodes supplied by Eviden and 8 CPU nodes also supplied by Eviden.
The DGX H200 are equiped with 8 NVIDIA® H200 141GB GPUs, 2 Intel® 8480C CPUs, and 2.1 TB RAM. The CPU nodes are each equipped with 2 AMD Epyc™ 9534 CPUs, 1.1 TB RAM and 6.4 TB of local NVMe SSD storage. The DGX H200 nodes are connected to a fast interconnect with 8x NVIDIA® Mellanox® NDR per DGX in a non-blocking fat-tree topology. This is a separate interconnect from that which connects the DGX A100 nodes in Berzelius Ampere.
All nodes have a local disk where applications can store temporary files. The
size of this disk (available to jobs as `/scratch/local`) is 30 TB on H200 nodes, and 6.4 TB
on CPU nodes, and is shared between all jobs using the node.
Berzelius Hopper is accessed through a new set of login nodes separate from those in the original Berzelius installation and also has new servers for other supporting tasks.
Berzelius Hopper is currently in a testpilot phase before being released for new projects in SUPR.
|
|
|
Berzelius Storage |
NSC |
— |
2 000 |
931 322 |
GiB |
Applications are normally evaluated during the last week each month.
|
|
Submit your proposal at least one week before the end of a month to be considered for an allocation from the first of the following month. Received proposals will be evaluated against each other and time that become available as project ends at the end of a month will be allocated to the proposed projects accordingly.
Shared, central storage accessible from all Berzelius Ampere and Berzelius Hopper compute and login nodes is provided by a storage cluster from VAST Data consisting of 8 CBoxes and 3 DBoxes using an NVMe-oF architecture. The storage servers are connected end-to-end to the GPUs using a high bandwidth interconnect separate from the East-West compute interconnect. The installed physical storage capacity is 3 PB, but due to compression and deduplication this will be higher in practice.
NSC centre storage (as available on Tetralith) is not accessible on Berzelius.
|