SNAC Large, Spring 2018

Closed for New Proposals

This round is closed for new proposals. The outcome of the round is pending committee decision.

More information about this round is available at


Resource Centre Total
Available Capacity Unit Note
Hebbe C3SE 1 400 600 x 1000 core-h/month
The Hebbe cluster is built on Intel Xeon E5-2650v3 (code-named "haswell") CPU's. The system has a total of 323 compute nodes (total of 6480 cores) with 27 TiB of RAM and 6 GPUs. More specific:
  • 260 x 64 GB of RAM (249 of these available for SNIC users)
  • 46 x 128 GB of RAM (31 of these available for SNIC users)
  • 7 x 256 GB of RAM (not available for SNIC users)
  • 3 x 512 GB of RAM (1 of these available for SNIC users)
  • 1 x 1024 GB of RAM
  • 4 x 64 GB of RAM and NVIDIA Tesla K40 GPU (2 of these available for SNIC users)
  • 2 x 256 GB of RAM and NVIDIA k4200 for remote graphics
Each node have 2 CPUs with 10 cores each. There's a 10Gigabit Ethernet network used for logins, and a dedicated management network and an Infiniband high-speed/low-latency network for parallel computations and filesystem access. The nodes are equipped with Mellanox ConnectX-3 FDR Infiniband 56Gbps HCA's.
Abisko HPC2N 5 600 2 500 x 1000 core-h/month
Abisko has reached its intended End-of-Life. As Abisko is quite a popular resource, HPC2N have decided to make it available for the Large Spring 2018 allocation round. Allocations may be reduced and/or moved to other resources, depending on availability of free time on other resources, due to parts failing and other causes.

The cluster has 15744 cores with a peak performance of over 150 Tflops/s. For high parallel performance, the system is equipped with a high bandwidth, low latency QDR InfiniBand interconnect, with full bisectional bandwidth. All nodes have at least 2 GB/core and some nodes have over 8 GB/core. For more information about the system and available software see the HPC2N web-pages.
Kebnekaise HPC2N 12 800 4 000 x 1000 core-h/month
Kebnekaise is a heterogeneous computing resource consisting of:


  1. Access to the Large Memory nodes are handled through the 'Kebnekaise Large Memory' resource.
  2. Note: It is important that requests for GPU nodes and KNL nodes are explicitly specified in the user's proposal. Also to note that the GPU nodes and the KNL nodes will be charged differently than ordinary computing nodes.
Kebnekaise Large Memory HPC2N 1 520 500 x 1000 core-h/month
This resource is for access to the 'Large Memory nodes' in Kebnekaise. For standard, GPU and KNL nodes see resource 'Kebnekaise'.

Kebnekaise is a heterogeneous computing resource consisting of:

Aurora Lunarc 1 600 500 x 1000 core-h/month
Was opened for test usage at the end of January 2016.
Triolith NSC 13 688 3 600 x 1000 core-h/month Triolith will be replaced with a new system named Tetralith. Awarded allocations will be transferred and users migrated.
Triolith will be replaced with a new system named Tetralith. Awarded Triolith allocations will be transferred to Tetralith and users migrated. Details will be announced later. The Tetralith installation will take place in two stages. The first stage will have a capacity that exceeds the current computing capacity of Triolith. NSC plan to have the first stage available from July 1, 2018. The second part is installed after Triolith is decommissioned and dismounted. NSC plan to have the entire Tetralith in operation by November. Existing centre storage will remain and be reconnected to Tetralith. Tetralith will be running a CentOS 7 version of the NSC Cluster Software Environment. Tetralith will consist of 1892 servers. Each server has two Intel Xeon Gold 6130 processors providing 32 cores per server. 1832 of the servers is equipped with 96 GiB of primary memory and 60 servers with 384 GiB. All servers are interconnected with a 100 Gbit Intel Omni-Path network which is also used to connect the existing storage. More information is available at:

Triolith ( was a capability cluster with a total of 24320 cores and a peak performance of 428 Tflops/s. However, Triolith was shrunk by 576 nodes on April 3rd, 2017 as a result of a delay in funding a replacement system and now has a peak performance of 260 Teraflop/sec and 16,368 compute cores. It is equipped with a fast interconnect for high performance for parallel applications. The operating system is CentOS 6.x x86_64. Each of the 1520 (now 944) HP SL230s compute servers is equipped with two Intel E5-2660 (2.2 GHz Sandybridge) processors with 8 cores each (i.e. 16 cores per compute server). 56 of the compute servers have 128 GiB memory each and the remaining 888 have 32 GiB each. The fast interconnect is Infiniband from Mellanox (FDR IB, 56 Gb/s) in a 2:1 blocking configuration.
Beskow PDC 21 010 11 200 x 1000 core-h/month
Tegner PDC 233 210 x 1000 core-h/month
Tegner is the pre/post processing cluster for Beskow
Crex 1 UPPMAX 20 000 1 000 000 GiB Storage resource attached to Rackham
Crex is the centre storage at UPPMAX, attached to the Rackham compute cluster. Proposals requesting Crex storage in SNAC LARGE must also include requests for compute resources totalling more the limits of SNAC MEDIUM (100 kch/month), at least part of which are on Rackham.

Active data storage for SNIC UPPMAX projects.
Rackham UPPMAX 1 320 1 000 x 1000 core-h/month
Rackham provides 6860 cores in the form of 334 nodes with two 10-core Intel Xeon V4 CPUs each. 32 fat nodes will have 256 GB of memory, with the rest having 128 GB. The interconnect is Infiniband.

Click the ▶ to show more information about the resource.