The Glenn cluster is built on AMD Opteron 6220 (code-named "Interlagos") CPU's. The system consists of:
In total 379 compute nodes (total of 6080 cores) and 18.1 TB of RAM. More specific:
224 nodes with 16 cores and 32 GB of RAM
135 nodes with 16 cores and 64 GB of RAM
13 nodes with 16 cores and 128 GB of RAM
1 node with 32 cores and 512 GB of RAM
4 nodes with 16 cores, 32 GB of RAM and 1 NVidia Fermi M2050 GPU
2 nodes with 16 cores, 32 GB of RAM and 1 PCoIP adapter for remote graphics
There are also 3 system servers used for accessing and managing the cluster.
There's a Gigabit Ethernet network used for logins and file system access, a dedicated management network and an Infiniband high-speed/low-latency network for parallell computations. The nodes are equipped with Mellanox ConnectX-2 QDR Infiniband 40Gbps HCA's.
The server and compute node hardware is built by Supermicro and delivered by South Pole.
Abisko
HPC2N
7 430
—
4 500
x 1000 core-h/month
The cluster has 15744 cores with a peak performance of over 150 Tflops/s. For high parallel performance, the system is equipped with a high bandwidth, low latency QDR InfiniBand interconnect, with full bisectional bandwidth. All nodes have at least 2 GB/core and some nodes have over 8 GB/core. For more information about the system and available software see the HPC2N web-pages.
Alarik
LUNARC
500
—
450
x 1000 core-h/month
Alarik was cluster consisting of 208 nodes. Each nodes had 2 eightcore CPU:s with 32-64 GB memory and 250 GB local scratch. The interconnect was Infiniband with 40 Gb/s. The system was connected to a gigabit Ethernet.
Alarik has been decomissioned.
Triolith
NSC
17 470
—
6 000
x 1000 core-h/month
Triolith (triolith.nsc.liu.se) was a capability cluster with a total of 24320 cores and a peak performance of 428 Tflops/s. However, Triolith was shrunk by 576 nodes on April 3rd, 2017 as a result of a delay in funding a replacement system and now has a peak performance of 260 Teraflop/sec and 16,368 compute cores. It is equipped with a fast interconnect for high performance for parallel applications. The operating system is CentOS 6.x x86_64. Each of the 1520 (now 944) HP SL230s compute servers is equipped with two Intel E5-2660 (2.2 GHz Sandybridge) processors with 8 cores each (i.e. 16 cores per compute server). 56 of the compute servers have 128 GiB memory each and the remaining 888 have 32 GiB each. The fast interconnect is Infiniband from Mellanox (FDR IB, 56 Gb/s) in a 2:1 blocking configuration.
Triolith have been replaced with a new system, Tetralith, that was made available to users on August 23, 2018. NSC currently plan to keep Triolith in operation and available to users until September 21st, 2018. After that, Triolith will be permanently shut down and decommissioned.
Lindgren
PDC
13 350
—
7 900
x 1000 core-h/month
Only 6 months allocations!
Lindgren will only be available for 6 months alloctaions in this round
Tintin
UPPMAX
1 370
—
390
x 1000 core-h/month
160 compute servers, each with two AMD Opteron 6220 (3.0 GHz 8 core Bulldozer) CPUs. 144 of the nodes have 64 GB memory and 16 have 128 GB memory. All nodes have 1 TB local disc. All nodes are interconnected with a 2:1 oversubscribed QDR Infiniband fabric. There is also 4 nodes with Tesla GPU graphics card. SNICs share of the Tintin system is 85.9%, i.e. about 137 compute servers.
Click above to show more information about the resource.