This round has been closed as all proposals have been handled.
To apply, you must be a scientist in Swedish academia, at least at the level of assistant professor.
Deadlines and Decisions
Monthly evaluation of proposals during the year.
Proposals submitted at the latest on the 15th will undergo review during the same month.
July and December have different schedules.
Proposals submitted after June 15 will be processed in August. i.e. with a first possible allocation starting September 1st!
This round is open for proposals until 2023-01-01 00:00.
|
Resource |
Centre |
Total Requested |
Upper Limit |
Available |
Unit |
Note |
|
Alvis |
C3SE |
394 070 |
20 000 |
390 000 |
GPU-h/month |
The Alvis resource is dedicated for AI/ML research.
|
|
The Alvis cluster is a national NAISS resource dedicated to Artificial Intelligence and
Machine Learning research.
Note: Significant generation of training data is expected to be done elsewhere.
The system is built around Graphical Processing Units
(GPUs) accelerator cards. The first phase of the resource has 160 NVIDIA T4, 44
V100, and 4 A100 GPUs. The second phase is based on 340 NVIDIA A40 and 336
A100 GPUs.
|
|
Kebnekaise |
HPC2N |
6 706 |
200 |
2 000 |
x 1000 core-h/month |
Kebnekaise is, due to age, no longer available in SNIC rounds. HPC2N partner sites can apply for time in ‘HPC2N Local’ rounds.
|
|
Kebnekaise will, due to its age, not be available for SNIC/NAIS Medium/Small rounds after 2022-12-31. We have therefore deactivated these resources in the corresponding rounds. For HPC2N partner sites there is a possibility to apply for resources in the ‘HPC2N Local Compute/Storage’ rounds with the caveat listed on the resources in the Local rounds.
This resource is for access to the CPU nodes in Kebnekaise.
For GPU nodes see resource 'Kebnekaise GPU'.
For large memory nodes see resource 'Kebnekaise Large Memory'.
Kebnekaise is a heterogeneous computing resource currently consisting of:
- Compute nodes:
- GPU nodes (separate resource):
- 10 Intel® Xeon Gold 6132 Processor (Skylake), 2x14 cores, 192 GB/node
- 2 AMD® EPYC 7413 (Zen3), 2x24 cores, 512 GB/node
- 1 AMD® EPYC 7413 (Zen3), 2x24 cores, 512 GB/node
- 1 AMD® EPYC 9254 (Zen4), 2x24 cores, 384 GB/node
- 10 AMD® EPYC 9254 (Zen4), 2x24 cores, 384 GB/node
- 2 AMD® EPYC 9454 (Zen4), 2x48 cores, 768 GB/node
- 1 AMD® EPYC 9334 (Zen4), 2x32 cores, 768 GB/node
- 2 AMD® EPYC 9334 (Zen4), 2x32 cores, 768 GB/node
- Large Memory nodes (separate resource):
Notes:
- Access to the GPU nodes are handled through the 'Kebnekaise GPU' resource.
- Access to the Large Memory nodes are handled through the 'Kebnekaise Large Memory' resource.
- New nodes will be procured on a semi-regular basis.
|
|
Kebnekaise Large Memory |
HPC2N |
111 |
10 |
50 |
x 1000 core-h/month |
Kebnekaise is, due to age, no longer available in SNIC rounds. HPC2N partner sites can apply for time in ‘HPC2N Local’ rounds.
|
|
Kebnekaise will, due to its age, not be available for SNIC/NAIS Medium/Small rounds after 2022-12-31. We have therefore deactivated these resources in the corresponding rounds. For HPC2N partner sites there is a possibility to apply for resources in the ‘HPC2N Local Compute/Storage’ rounds with the caveat listed on the resources in the Local rounds.
This resource is for access to the Large Memory nodes in Kebnekaise.
For CPU nodes see resource 'Kebnekaise'.
For GPU nodes see resource 'Kebnekaise GPU'.
Kebnekaise is a heterogeneous computing resource currently consisting of:
- Compute nodes (separate resource):
- GPU nodes (separate resource):
- 10 Intel® Xeon Gold 6132 Processor (Skylake), 2x14 cores, 192 GB/node
- 2 AMD® EPYC 7413 (Zen3), 2x24 cores, 512 GB/node
- 1 AMD® EPYC 7413 (Zen3), 2x24 cores, 512 GB/node
- 1 AMD® EPYC 9254 (Zen4), 2x24 cores, 384 GB/node
- 10 AMD® EPYC 9254 (Zen4), 2x24 cores, 384 GB/node
- 2 AMD® EPYC 9454 (Zen4), 2x48 cores, 768 GB/node
- 1 AMD® EPYC 9334 (Zen4), 2x32 cores, 768 GB/node
- 2 AMD® EPYC 9334 (Zen4), 2x32 cores, 768 GB/node
- Large Memory nodes:
Notes:
- Access to the CPU nodes are handled through the 'Kebnekaise' resource.
- Access to the GPU nodes are handled through the 'Kebnekaise GPU' resource.
- New nodes will be procured on a semi-regular basis.
|
|
Tetralith |
NSC |
21 572 |
200 |
11 500 |
x 1000 core-h/month |
|
|
Tetralith is a general computational resource hosted by NSC at Linköping University.
Tetralith servers have two Intel Xeon Gold 6130 processors, providing 32 cores per server. 1844 of the servers are equipped with 96 GiB of primary memory and 64 servers with 384 GiB. All servers are interconnected with a 100 Gbit/s Intel Omni- Path network which is also used to connect the existing storage. Each server has a local SSD disk for ephemeral storage (approx. 200GiB per thin node, 900GiB per fat node). An IBM Spectrum Scale system comprises the centre storage. 170 of the Tetralith nodes are equipped with one NVIDIA Tesla T4 GPU each as well as a high- performance NVMe SSD scratch disk of 2TB.
|
|
Dardel |
PDC |
13 450 |
200 |
11 900 |
x 1000 core-h/month |
|
|
Dardel is a Cray EX system from Hewlett Packard Enterprise, based on AMD EPYC processors with an accompanying Lustre storage system.
The nodes are interconnected using Slingshot HPC Ethernet.
|
|
Rackham |
UPPMAX |
8 015 |
200 |
3 500 |
x 1000 core-h/month |
SNIC Life Science system at UPPMAX, mounts the Crex filesystem
|
|
SNIC Life Science system at UPPMAX, mounts the Crex filesystem
Rackham provides 9720 cores in the form of 486 nodes with two 10-core Intel Xeon V4 CPUs each. 4 fat nodes have 1 TB of memory, 32 fat nodes have 256 GB, and the rest have 128 GB.
The interconnect is Infiniband.
|