Logo

HPC @ Uni.lu

High Performance Computing in Luxembourg

This website is deprecated, the old pages are kept online but you should refer in priority to the new web site hpc.uni.lu and the new technical documentation site hpc-docs.uni.lu

This is a recurring question, especially in light of the heterogeneity of clusters and their overlapping capabilities.

Here are the relevant guidelines, with cases and answers:

Nature of HPC activity System(s) Comments
For running either serial or parallel jobs chaos/gaia reserve exactly CPU resources you need
For developing or trying out applications chaos/gaia reserve exactly CPU resources you need
For small serial jobs chaos/gaia  
If you need linear scaling across nodes chaos/gaia Prefer whole-node reservations for #N>12
If you are running job-arrays chaos/gaia  
Use checkpointing to overcome X-days limit per job chaos/gaia Either CRIU or BLCR or user-space technologies should work
For lengthy jobs, >3 days chaos/gaia ask a project description and explain operational aspects
To run MATLAB chaos/gaia module load MATLAB
To run serial Matlab jobs chaos/gaia remember to scale it with compiled code
To run parallel Matlab jobs chaos/gaia remember to scale it with compiled code
To run statistical computing software like R, Rmpi, Snow, Snowfall chaos/gaia The non-bare version of R in modules has strong features
For post-processing and visualization chaos/gaia Normal system nodes should be used for this, ask for 1 core
Run precompiled applications chaos/gaia long list is visible at [1], [2]
Use EasyBuild to develop software chaos/gaia get familiar with the content at [3]
If you need a large, single memory system chaos/gaia class bigmem offers 1TB per 32 cores (gaia-74/r-cluster-1)
If you need a large, single memory system with AVX commandset gaia class bigmem offers 1TB per 32 cores (gaia-74)
If you need a large, single memory system with multiple cores gaia node #73 is a bigsmp (ccNUMA) architecture with 160 cores
If you need to experiment with accelerators and GPU computing gaia nodes #61-#72 have nVidia GPU cards; see. http://hpc.uni.lu
For tightly coupled multinode parallel jobs w. low latency network gaia all nodes are on infiniband, although #74 is out of lustre
To run Stata gaia module load Stata
If you prefer Intel CPUs chaos/gaia Intel Cluster Studio XE 2013 has been purchased, use it
If you prefer AMD CPUs N/A  

The users are kindly asked to check what are the actual requirements for each application and use the resources appropriately and in a rational way.

[1] http://hpcbios.readthedocs.org/en/latest/HPCBIOS_2012-92.html

[2] https://github.com/hpcugent/easybuild/wiki/List-of-supported-software-packages

[3] https://easybuild.readthedocs.org/