NAISS
SUPR
SUPR
SNAC Medium, 2016

Decided

This round has been closed as all proposals have been handled.

See further information.

Resources

Resource Centre Total
Requested
Upper
Limit
Available Unit Note
Glenn C3SE 140 120 800 x 1000 core-h/month Note that Glenn only is available the first months of 2016!
Glenn will be taken out of the SNAC round during the spring. Please only apply for time on Glenn to finish off existing projects. Hebbe is the replacement resource for Glenn.

The Glenn cluster is built on AMD Opteron 6220 (code-named "Interlagos") CPU's. The system consists of: In total 379 compute nodes (total of 6080 cores) and 18.1 TB of RAM. More specific:
  • 224 nodes with 16 cores and 32 GB of RAM
  • 135 nodes with 16 cores and 64 GB of RAM
  • 13 nodes with 16 cores and 128 GB of RAM
  • 1 node with 32 cores and 512 GB of RAM
  • 4 nodes with 16 cores, 32 GB of RAM and 1 NVidia Fermi M2050 GPU
  • 2 nodes with 16 cores, 32 GB of RAM and 1 PCoIP adapter for remote graphics
There are also 3 system servers used for accessing and managing the cluster. There's a Gigabit Ethernet network used for logins and file system access, a dedicated management network and an Infiniband high-speed/low-latency network for parallell computations. The nodes are equipped with Mellanox ConnectX-2 QDR Infiniband 40Gbps HCA's. The server and compute node hardware is built by Supermicro and delivered by South Pole.
Hebbe C3SE 2 612 100 1 100 x 1000 core-h/month
The Hebbe cluster is built on Intel Xeon E5-2650v3 (code-named "haswell") CPU's. The system has a total of 323 compute nodes (total of 6480 cores) with 27 TiB of RAM and 6 GPUs. More specific:
  • 260 x 64 GB of RAM (249 of these available for SNIC users)
  • 46 x 128 GB of RAM (31 of these available for SNIC users)
  • 7 x 256 GB of RAM (not available for SNIC users)
  • 3 x 512 GB of RAM (1 of these available for SNIC users)
  • 1 x 1024 GB of RAM
  • 4 x 64 GB of RAM and NVIDIA Tesla K40 GPU (2 of these available for SNIC users)
  • 2 x 256 GB of RAM and NVIDIA k4200 for remote graphics
Each node have 2 CPUs with 10 cores each. There's a 10Gigabit Ethernet network used for logins, and a dedicated management network and an Infiniband high-speed/low-latency network for parallel computations and filesystem access. The nodes are equipped with Mellanox ConnectX-3 FDR Infiniband 56Gbps HCA's.
Abisko HPC2N 7 959 160 2 000 x 1000 core-h/month Proposals will be evaluated at most once per month
The cluster has 15744 cores with a peak performance of over 150 Tflops/s. For high parallel performance, the system is equipped with a high bandwidth, low latency QDR InfiniBand interconnect, with full bisectional bandwidth. All nodes have at least 2 GB/core and some nodes have over 8 GB/core. For more information about the system and available software see the HPC2N web-pages.
Akka HPC2N 100 80 2 000 x 1000 core-h/month Proposals will be evaluated at most once per month
The cluster contains 672 nodes each in a dual quad-core configuration, for a total of 5376 cores. The CPU:s are the Intel L5420 low-power CPU:s running at 2.5 GHz, giving each node a total peak performance of 80 Gflops/s. Each node is equipped with 16 GB RAM for a total of more than 10 Terabytes of RAM. For high parallel performance, the system is equipped with a high bandwidth, low latency InfiniBand interconnect. Scratch storage is provided by a 100 TB disk-subsystem running GPFS. For more information about the system and available software see the HPC2N web-pages. Akka will be closed down at the end of 2012. In this round we will only award 6 months allocations.
Kebnekaise HPC2N 1 466 200 1 000 x 1000 core-h/month Proposals will be evaluated at most once per month

This resource is for access to the CPU nodes in Kebnekaise.

For GPU nodes see resource 'Kebnekaise GPU'.

For large memory nodes see resource 'Kebnekaise Large Memory'.

 

Kebnekaise is a heterogeneous computing resource currently consisting of: Notes:
  1. Access to the GPU nodes are handled through the 'Kebnekaise GPU' resource.
  2. Access to the Large Memory nodes are handled through the 'Kebnekaise Large Memory' resource.
  3. New nodes will be procured on a semi-regular basis.
Kebnekaise Large Memory HPC2N 53 20 100 x 1000 core-h/month Proposals will be evaluated at most once per month

This resource is for access to the Large Memory nodes in Kebnekaise.

For CPU nodes see resource 'Kebnekaise'.

For GPU nodes see resource 'Kebnekaise GPU'.

 

Kebnekaise is a heterogeneous computing resource currently consisting of: Notes:
  1. Access to the CPU nodes are handled through the 'Kebnekaise' resource.
  2. Access to the GPU nodes are handled through the 'Kebnekaise GPU' resource.
  3. New nodes will be procured on a semi-regular basis.
Aurora LUNARC 3 160 100 1 100 x 1000 core-h/month
Aurora is the Lund university compute resource and is operated by LUNARC
Erik LUNARC 106 8 66 x 1000 core-h/month
Erik is the SNIC GPU Pilot system. Special review is done before accepting applications. Allocation is done in ch/m based on gpu:s in the system. Erik will retire as a SNIC resource on 2017-01-01 NOTE: SNIC GPU Pilot allocation info, http://www.lunarc.lu.se/Support/snic-gpu-pilot-allocation-information
Triolith NSC 13 635 200 6 000 x 1000 core-h/month Applications will be evaluated once a month.
Allocations on Triolith will be reduced to 60 % from some time in the spring 2017 due to ending funding from SNIC. Currently we estimate that the reduction will be from 2017-04-01.

Triolith (triolith.nsc.liu.se) was a capability cluster with a total of 24320 cores and a peak performance of 428 Tflops/s. However, Triolith was shrunk by 576 nodes on April 3rd, 2017 as a result of a delay in funding a replacement system and now has a peak performance of 260 Teraflop/sec and 16,368 compute cores. It is equipped with a fast interconnect for high performance for parallel applications. The operating system is CentOS 6.x x86_64. Each of the 1520 (now 944) HP SL230s compute servers is equipped with two Intel E5-2660 (2.2 GHz Sandybridge) processors with 8 cores each (i.e. 16 cores per compute server). 56 of the compute servers have 128 GiB memory each and the remaining 888 have 32 GiB each. The fast interconnect is Infiniband from Mellanox (FDR IB, 56 Gb/s) in a 2:1 blocking configuration. Triolith have been replaced with a new system, Tetralith, that was made available to users on August 23, 2018. NSC currently plan to keep Triolith in operation and available to users until September 21st, 2018. After that, Triolith will be permanently shut down and decommissioned.
Beskow PDC 16 500 200 7 500 x 1000 core-h/month
After a proposal has been approved it can take a couple of days before it is activated
Tintin UPPMAX 1 425 80 260 x 1000 core-h/month Tintin is scheduled to be decommissioned in early 2017
Tintin is scheduled to be decommissioned in early 2017

160 compute servers, each with two AMD Opteron 6220 (3.0 GHz 8 core Bulldozer) CPUs. 144 of the nodes have 64 GB memory and 16 have 128 GB memory. All nodes have 1 TB local disc. All nodes are interconnected with a 2:1 oversubscribed QDR Infiniband fabric. There is also 4 nodes with Tesla GPU graphics card. SNICs share of the Tintin system is 85.9%, i.e. about 137 compute servers.

Click above to show more information about the resource.