NAISS
SUPR
SUPR
SNAC Large, Spring 2016

Decided

This round has already been decided by the committee and is not open for proposals.

See further information.

Resources

Resource Centre Total
Requested
Upper
Limit
Available Unit Note
Hebbe C3SE 1 070 600 x 1000 core-h/month
The Hebbe cluster is built on Intel Xeon E5-2650v3 (code-named "haswell") CPU's. The system has a total of 323 compute nodes (total of 6480 cores) with 27 TiB of RAM and 6 GPUs. More specific:
  • 260 x 64 GB of RAM (249 of these available for SNIC users)
  • 46 x 128 GB of RAM (31 of these available for SNIC users)
  • 7 x 256 GB of RAM (not available for SNIC users)
  • 3 x 512 GB of RAM (1 of these available for SNIC users)
  • 1 x 1024 GB of RAM
  • 4 x 64 GB of RAM and NVIDIA Tesla K40 GPU (2 of these available for SNIC users)
  • 2 x 256 GB of RAM and NVIDIA k4200 for remote graphics
Each node have 2 CPUs with 10 cores each. There's a 10Gigabit Ethernet network used for logins, and a dedicated management network and an Infiniband high-speed/low-latency network for parallel computations and filesystem access. The nodes are equipped with Mellanox ConnectX-3 FDR Infiniband 56Gbps HCA's.
Abisko HPC2N 8 630 3 000 x 1000 core-h/month
During the summer of 2016 a new HPC resources to replace Abisko will be taken into operation. The specs of the new system will be known during the evaluation of the spring 2016 Large round. SNAC will make an assessment of which successful applications for large allocations of computing time that are suitable for allocations on the new system.

The cluster has 15744 cores with a peak performance of over 150 Tflops/s. For high parallel performance, the system is equipped with a high bandwidth, low latency QDR InfiniBand interconnect, with full bisectional bandwidth. All nodes have at least 2 GB/core and some nodes have over 8 GB/core. For more information about the system and available software see the HPC2N web-pages.
Aurora LUNARC 1 150 600 x 1000 core-h/month
Aurora is the Lund university compute resource and is operated by LUNARC
Triolith NSC 16 720 6 000 x 1000 core-h/month
Triolith (triolith.nsc.liu.se) was a capability cluster with a total of 24320 cores and a peak performance of 428 Tflops/s. However, Triolith was shrunk by 576 nodes on April 3rd, 2017 as a result of a delay in funding a replacement system and now has a peak performance of 260 Teraflop/sec and 16,368 compute cores. It is equipped with a fast interconnect for high performance for parallel applications. The operating system is CentOS 6.x x86_64. Each of the 1520 (now 944) HP SL230s compute servers is equipped with two Intel E5-2660 (2.2 GHz Sandybridge) processors with 8 cores each (i.e. 16 cores per compute server). 56 of the compute servers have 128 GiB memory each and the remaining 888 have 32 GiB each. The fast interconnect is Infiniband from Mellanox (FDR IB, 56 Gb/s) in a 2:1 blocking configuration. Triolith have been replaced with a new system, Tetralith, that was made available to users on August 23, 2018. NSC currently plan to keep Triolith in operation and available to users until September 21st, 2018. After that, Triolith will be permanently shut down and decommissioned.
Beskow PDC 21 625 11 200 x 1000 core-h/month
Tegner PDC 0 210 x 1000 core-h/month
Pre/Post system for medium/large SNIC allocations on Beskow

Tegner is the pre/post processing cluster for Beskow
Tintin UPPMAX 680 260 x 1000 core-h/month
Tintin may be decommissioned during the allocation period. At that time, projects will be transferred to the replacement system, Rackham.

160 compute servers, each with two AMD Opteron 6220 (3.0 GHz 8 core Bulldozer) CPUs. 144 of the nodes have 64 GB memory and 16 have 128 GB memory. All nodes have 1 TB local disc. All nodes are interconnected with a 2:1 oversubscribed QDR Infiniband fabric. There is also 4 nodes with Tesla GPU graphics card. SNICs share of the Tintin system is 85.9%, i.e. about 137 compute servers.

Click above to show more information about the resource.