NAISS
SUPR
SUPR
NAISS Medium Compute 2024

Open for Proposals

To apply, you must be a scientist in Swedish academia, at least at the level of assistant professor.

Deadlines and Decisions

Proposals are processed monthly during the year. Note that staff will be on vacation during the summer and Christmas and New Year holidays. Thus, July and December have different schedules than other months.
If Submitted Latest On Expect Decisions Before Note
2024-04-15 2024-05-01
2024-05-15 2024-06-01
2024-06-15 2024-07-01
2024-08-15 2024-09-01 Vacation Period
2024-09-15 2024-10-01
2024-10-15 2024-11-01
2024-11-15 2024-12-01
2024-12-12 2025-01-01 Christmas and New Year
2024-12-31 2025-02-01 Round close at New Year

Resources

Resource Centre Upper
Limit
Available Unit Note
Alvis C3SE 20 000 390 000 GPU-h/month The Alvis resource is dedicated for AI/ML research.
The Alvis resource is dedicated for research in and research using AI/ML techniques. For general use of GPU:s instead use Dardel GPU.

The Alvis cluster is a national NAISS resource dedicated to Artificial Intelligence and Machine Learning research.

The system is built around Graphical Processing Units (GPUs) accelerator cards. The first phase of the resource has 160 NVIDIA T4, 44 V100, and 4 A100 GPUs. The second phase is based on 340 NVIDIA A40 and 336 A100 GPUs.

Tetralith NSC 400 14 500 x 1000 core-h/month
Projects will receive a default 500 GiB storage allocation on Centre Storage at NSC. If you need more storage, please apply for a Storage project and decline default storage from this compute proposal.

Tetralith is a general computational resource hosted by NSC at Linköping University.

Tetralith servers have two Intel Xeon Gold 6130 processors, providing 32 cores per server. 1844 of the servers are equipped with 96 GiB of primary memory and 64 servers with 384 GiB. All servers are interconnected with a 100 Gbit/s Intel Omni- Path network which is also used to connect the existing storage. Each server has a local SSD disk for ephemeral storage (approx. 200GiB per thin node, 900GiB per fat node). An IBM Spectrum Scale system comprises the centre storage. 170 of the Tetralith nodes are equipped with one NVIDIA Tesla T4 GPU each as well as a high- performance NVMe SSD scratch disk of 2TB.

Dardel PDC 400 28 500 x 1000 core-h/month
Dardel is a Cray EX system from Hewlett Packard Enterprise, based on AMD EPYC processors with an accompanying Lustre storage system. The nodes are interconnected using Slingshot HPC Ethernet.
Dardel-GPU PDC 4 000 106 000 GPU-h/month
These GPUs are not nVIDIA GPUs but rather AMD GPUs, so if your software runs using CUDA, a certain amount of conversion of the code is needed. You can read information about this at https://www.lumi-supercomputer.eu/preparing-codes-for-lumi-converting-cuda-applications-to-hip/ Reporting on GPU consumption on Dardel is not working yet.

Dardel-GPU is the accelerated partition based on AMD’s Instinct MI250X GPU of the Cray EX system from Hewlett Packard Enterprise. It has an accompanying Lustre storage system. The nodes are interconnected using Slingshot HPC Ethernet.
Rackham UPPMAX 200 3 500 x 1000 core-h/month Restrictive policy for NEW projects on Rackham.
Rackham will be decommissioned on 2024-12-31. No allocations will be made beyond this date. See https://www.uppmax.uu.se/uppmax-news/?tarContentId=1069594 Few new Medium-scale projects will be accepted. Continuation proposals are welcome. New projects must carefully describe a plan for the project from now until 2024-12-31. UU-affiliated projects can be moved automatically to the new local system. Other projects must have a clear and concrete exit plan.

Rackham provides 9720 cores in the form of 486 nodes with two 10-core Intel Xeon V4 CPUs each. 4 fat nodes have 1 TB of memory, 32 fat nodes have 256 GB, and the rest have 128 GB. The interconnect is Infiniband.

Click above to show more information about the resource.