NAISS
SUPR
SUPR
NAISS Large Fall 2025

Open for Proposals

To apply, you must be at least an assistant professor at an eligible Swedish institution. NAISS is financed by Swedish tax payers with the objective to strengthen science and research in Sweden. To apply for these resources, you must therefore hold an appointment of at least 60 percent at a Swedish university or research institution recognized by the Swedish Research Council and have been the Principal Investigator of a NAISS Medium project or similar.

The deadline for submitting proposals is 2025-10-16 15:00.

See further information.

Resources

Resource Centre Upper
Limit
Default
Storage
Available Unit Note
Alvis C3SE 124 971 GPU-h/month This resource is only intended for AI/ML research.
This resource is only intended for research on AI/ML or research using AI/ML methods. Allocations on Alvis will be scaled and transferred to the new NAISS system Arrhenius once the latter becomes available. Our current best estimate is that this will occur in the spring of 2026. Alvis is guaranteed to operate until 2026-07-01. More information will be announced as the procurement and installation of Arrhenius progresses.

The Alvis cluster is a national NAISS resource dedicated to Artificial Intelligence and Machine Learning research.

Note: Significant generation of training data is expected to be done elsewhere.

The system is built around Graphical Processing Units (GPUs) accelerator cards. The first phase of the resource has 160 NVIDIA T4, 44 V100, and 4 A100 GPUs. The second phase is based on 340 NVIDIA A40 and 336 A100 GPUs.

Mimer C3SE 1 000 2 000 000 GiB
Mimer provide storage for Alvis. Allocations on Mimer will be scaled and transferred to storage on the new NAISS system Arrhenius once the latter becomes available. Our current best estimate is that this will occur in the spring of 2026. Mimer is guaranteed to operate until 2026-07-01. More information will be announced as the procurement and installation of Arrhenius progresses.

Project storage attached to Alvis and Vera, dedicated for AI/ML

Mimer is an all-flashed based storage system based on as solution from WEKA IO. It consists of an 0.6 PB all-flash tier and a 7 PB Ceph based bulk storage tier (with spinning disk).

Tetralith NSC 14 500 x 1000 core-h/month
Allocations on Tetralith will be scaled and transferred to the new NAISS system Arrhenius once the latter becomes available. Our current best estimate is that this will occur in the spring of 2026. More information will be announced as the procurement and installation of Arrhenius progresses.

Tetralith is a general computational resource hosted by NSC at Linköping University.

Tetralith servers have two Intel Xeon Gold 6130 processors, providing 32 cores per server. 1844 of the servers are equipped with 96 GiB of primary memory and 64 servers with 384 GiB. All servers are interconnected with a 100 Gbit/s Intel Omni- Path network which is also used to connect the existing storage. Each server has a local SSD disk for ephemeral storage (approx. 200GiB per thin node, 900GiB per fat node). An IBM Spectrum Scale system comprises the centre storage. 170 of the Tetralith nodes are equipped with one NVIDIA Tesla T4 GPU each as well as a high- performance NVMe SSD scratch disk of 2TB.

Centre Storage NSC 500 2 560 000 GiB Storage resource connected to compute resources Tetralith and Sigma at NSC.
Storage resource connected to compute resources Tetralith and Sigma at NSC. The end date of granted storage allocations will be adjusted to match the end date of compute projects that will utilizing the storage allocation. Allocations on Centre Storage will be scaled and transferred to storage on the new NAISS system Arrhenius once the latter becomes available. Our current best estimate is that this will occur in the spring of 2026. More information will be announced as the procurement and installation of Arrhenius progresses.

Project storage for NAISS as well as LiU Local projects with compute allocations on resources hosted by NSC.

Centre Storage @ NSC is designed for fast access from compute resources at NSC. It consists of one IBM ESS GL6S building block and one IBM ESS 5000 SC4 building block.

In total there are 946 spinning hard disks and a small number of NVRAM devices and SSDs which act as a cache to speed up small writes. The total disk space that is usable for storing files is approximately 6.9 PiB.

Dardel PDC 28 000 x 1000 core-h/month
Dardel is a Cray EX system from Hewlett Packard Enterprise, based on AMD EPYC processors with an accompanying Lustre storage system. The nodes are interconnected using Slingshot HPC Ethernet.
Dardel-GPU PDC 105 000 GPU-h/month
Dardel-GPU is the accelerated partition based on AMD’s Instinct MI250X GPU of the Cray EX system from Hewlett Packard Enterprise. It has an accompanying Lustre storage system. The nodes are interconnected using Slingshot HPC Ethernet.
Klemming PDC 500 2 925 000 GiB Storage resource connected to compute resources Dardel at PDC and Dardel-GPU at PDC.
Storage resource connected to compute resources Dardel at PDC and Dardel-GPU at PDC. The end date of granted storage allocations will be adjusted to match the end date of compute projects that will utilizing the storage allocation.

Project storage for NAISS as well as PDC projects with compute allocations on resources hosted by PDC.

Klemming is designed for fast access from compute resources at PCD. It uses the Lustre parallel file system, which is optimized for handling data from many clients at the same time. The total size of Klemming is 12 PB.

Bianca UPPMAX 1 100 x 1000 core-h/month NAISS SENS
System for projects involved sensitive data. Do not apply for this resource if you do not intend to work with sensitive data. Note that the Bianca system will be replaced by the new NAISS SENS system, Maja. Your compute allocation will at that point be transferred. Your data will remain on the same physical storage system (Cygnus), and will not need to be moved. It is NAISS's best estimate now that this will occur during winter 2025/2026.

Bianca is a research system dedicated to analysing sensitive personal data, or other types of sensitive data.

Bianca provides 4480 cores in the form of 204 dual CPU (Intel Xeon E5-2630 v3) Huawei XH620 V3 nodes with 128GB memory, 75 fat nodes with 256 GB of memory, 15 nodes with 512 GB of memory, ten nodes with two NVIDIA A100 40GB GPUs each.

Cygnus /proj UPPMAX 1 000 300 000 GiB Backed-up secure storage connected to Bianca.
Storage for projects involved sensitive data. Do not apply for this resource if you do not intend to work with sensitive data. This resource features 30-day incremental backup. Regardless, this is not a suitable system for primary data storage, and good data management and information security practices demand that important data must be replicated elsewhere.

Cygnus is the new storage resource attached to Bianca, the NAISS SENS research cluster. The /proj area is backed up.
Cygnus /proj/nobackup UPPMAX 1 000 300 000 GiB Secure storage connected to Bianca.
Storage for projects involving sensitive data. Do not apply for this resource if you do not intend to work with sensitive data. This storage resource contains the Wharf, which allows data transfers in and out of Bianca. You must request a non-zero amount of this resource if you apply for Bianca.

Cygnus is the new storage resource attached to Bianca, the NAISS SENS research cluster. The /proj/nobackup area is not backed up.

Click above to show more information about the resource.