NAISS
SUPR
SUPR
NAISS Medium Compute 2023

Decided

This round has been closed as all proposals have been handled.

To apply, you must be a scientist in Swedish academia, at least at the level of assistant professor.

See further information.

Deadlines and Decisions

Monthly evaluation of proposals during the year. Proposals submitted at the latest on the 15th will undergo review during the same month. July and December have different schedules. Proposals submitted at the latest on the 13th December will undergo review before Christmas. Proposals submitted later in December will be processed in January. Proposals submitted after June 15 will be processed in August. i.e. with a first possible allocation starting September 1st!

This round is open for proposals until 2024-01-01 00:00.

Resources

Resource Centre Total
Requested
Upper
Limit
Default
Storage
Available Unit Note
Alvis C3SE 565 998 20 000 390 000 GPU-h/month The Alvis resource is dedicated for AI/ML research.
The Alvis resource is dedicated for research in and research using AI/ML techniques. For general use of GPU:s instead use Dardel.

The Alvis cluster is a national NAISS resource dedicated to Artificial Intelligence and Machine Learning research.

Note: Significant generation of training data is expected to be done elsewhere.

The system is built around Graphical Processing Units (GPUs) accelerator cards. The first phase of the resource has 160 NVIDIA T4, 44 V100, and 4 A100 GPUs. The second phase is based on 340 NVIDIA A40 and 336 A100 GPUs.

Mimer C3SE 14 000 500 100 000 GiB

Project storage attached to Alvis and Vera, dedicated for AI/ML

Mimer is an all-flashed based storage system based on as solution from WEKA IO. It consists of an 0.6 PB all-flash tier and a 7 PB Ceph based bulk storage tier (with spinning disk).

Tetralith NSC 31 107 400 14 500 x 1000 core-h/month
Projects will receive a default 500 GiB storage allocation on Centre Storage at NSC. If you need more storage, please apply for a Storage project and decline default storage from this compute proposal.

Tetralith is a general computational resource hosted by NSC at Linköping University.

Tetralith servers have two Intel Xeon Gold 6130 processors, providing 32 cores per server. 1844 of the servers are equipped with 96 GiB of primary memory and 64 servers with 384 GiB. All servers are interconnected with a 100 Gbit/s Intel Omni- Path network which is also used to connect the existing storage. Each server has a local SSD disk for ephemeral storage (approx. 200GiB per thin node, 900GiB per fat node). An IBM Spectrum Scale system comprises the centre storage. 170 of the Tetralith nodes are equipped with one NVIDIA Tesla T4 GPU each as well as a high- performance NVMe SSD scratch disk of 2TB.

Centre Storage NSC 63 000 500 60 000 GiB

Project storage for NAISS as well as LiU Local projects with compute allocations on resources hosted by NSC.

Centre Storage @ NSC is designed for fast access from compute resources at NSC. It consists of one IBM ESS GL6S building block and one IBM ESS 5000 SC4 building block.

In total there are 946 spinning hard disks and a small number of NVRAM devices and SSDs which act as a cache to speed up small writes. The total disk space that is usable for storing files is approximately 6.9 PiB.

Dardel PDC 24 022 400 28 500 x 1000 core-h/month
Dardel is a Cray EX system from Hewlett Packard Enterprise, based on AMD EPYC processors with an accompanying Lustre storage system. The nodes are interconnected using Slingshot HPC Ethernet.
Dardel-GPU PDC 28 760 4 000 106 000 GPU-h/month
GPU nodes on Dardel will probably be generally available 2023-01-01, but there is a risk for delays due to server maintenance to accomodate the GPUs. Also, these GPUs are not nVIDIA GPUs but rather AMD GPUs, so if your software runs using CUDA, a certain amount of conversion of the code is needed. You can read information about this at https://www.lumi-supercomputer.eu/preparing-codes-for-lumi-converting-cuda-applications-to-hip/

Dardel-GPU is the accelerated partition based on AMD’s Instinct MI250X GPU of the Cray EX system from Hewlett Packard Enterprise. It has an accompanying Lustre storage system. The nodes are interconnected using Slingshot HPC Ethernet.
Klemming PDC 55 000 500 300 000 GiB
More information about project directories in Klemming can be found at https://www.pdc.kth.se/support/documents/data_management/lustre.html.

Project storage for NAISS as well as PDC projects with compute allocations on resources hosted by PDC.

Klemming is designed for fast access from compute resources at PCD. It uses the Lustre parallel file system, which is optimized for handling data from many clients at the same time. The total size of Klemming is 12 PB.

Rackham UPPMAX 8 257 200 3 500 x 1000 core-h/month
Rackham provides 9720 cores in the form of 486 nodes with two 10-core Intel Xeon V4 CPUs each. 4 fat nodes have 1 TB of memory, 32 fat nodes have 256 GB, and the rest have 128 GB. The interconnect is Infiniband.
Crex 1 UPPMAX 9 472 128 20 000 GiB Default storage for projects with Rackham allocations.
Default storage for projects with Rackham allocations. Limited to 128 GB. If more storage is needed, please apply for a Storage project.

Active data storage for Rackham projects. Primarily for life science projects.

Click above to show more information about the resource.