HPE Cray Programming Environment¶
In this session we discuss some of the basics of the operating system and programming environment on LUMI. Whether you like it or not, every user of a supercomputer like LUMI gets confronted with these elements at some point.
Why do I need to know this?¶
The typical reaction of someone who only wants to run software on an HPC system when confronted with a talk about development tools is "I only want to run some programs, why do I need to know about programming environments?"
The answer is that development environments are an intrinsic part of an HPC system. No HPC system is as polished as a personal computer and the software users want to use is typically very unpolished. And some of the essential middleware that turns the hardware with some variant of Linux into a parallel supercomputers is part of the programming environment. The binary interfaces to those libraries are also not as standardised as for the more common Linux system libraries.
Programs on an HPC cluster are preferably installed from sources to generate binaries optimised for the system. CPUs have gotten new instructions over time that can sometimes speed-up execution of a program a lot, and compiler optimisations that take specific strengths and weaknesses of particular CPUs into account can also gain some performance. Even just a 10% performance gain on an investment of 160 million EURO such as LUMI means a lot of money. When running, the build environment on most systems needs to be at least partially recreated. This is somewhat less relevant on Cray systems as we will see at the end of this part of the course, but if you want reproducibility it becomes important again.
Compiling on the system is also the easiest way to guarantee compatibility of the binaries with the system.
Even when installing software from prebuilt binaries some modules might still be needed. Prebuilt binaries will typically include the essential runtime libraries for the parallel technologies they use, but these may not be compatible with LUMI. In some cases this can be solved by injecting a library from LUMI, e.g., you may want to inject an optimised MPI library as we shall see in the container section of this course. But sometimes a binary is simply incompatible with LUMI and there is no other solution than to build the software from sources.
Clusters in Belgium
There are differences in the setup of the operating system on LUMI compared to the VSC, CÉCI and Cenaero clusters in Belgium.
Lucia, the Cenaero cluster and Walloon tier-1 system, has a programming environment which is similar to LUMI. But all other Belgian systems have a programming environment that is more inspired on the GNU and other Open Source software approach. (I don't want to say "a more traditional" programming environment as in fact the HPE Cray PE goes back in its approach to the environment on UNIX workstations and supercomputers in the '90s, so is in fact the more traditional one. It's just that we have forgotten those traditions...)
The operating system on LUMI¶
The login nodes of LUMI run a regular SUSE Linux Enterprise Server 15 SP5 distribution. That alone already causes subtle differences with the Red Hat clones and Debian Linux derivatives (like Ubuntu) that are popular on many clusters. The compute nodes however run Cray OS, a restricted version of the SUSE Linux that runs on the login nodes. Some daemons are inactive or configured differently and Cray also does not support all regular file systems. The goal of this is to minimize OS jitter, interrupts that the OS handles and slow down random cores at random moments, that can limit scalability of programs. Yet on the GPU nodes there was still the need to reserve one core for the OS and driver processes. This in turn led to an asymmetry in the setup so now 8 cores are reserved, one per CCD, so that all CCDs are equal again.
This also implies that some software that works perfectly fine on the login nodes may not
work on the compute nodes. E.g., you will see that there is no /run/user/$UID
directory.
Large HPC clusters also have a small system image, so don't expect all the bells-and-whistles from a Linux workstation to be present on a large supercomputer (and certainly not in the same way as they would be on a workstation). Since LUMI compute nodes are diskless, the system image actually occupies RAM which is another reason to keep it small.
Some missing pieces
Compute nodes don't run a per-user dbus daemon, so some if not all DBUS functionality is missing. And D-Bus may sometimes show up in places where you don't expect it... It may come from freedesktop.org but is is not only used for desktop software.
Compute nodes on a Cray system have Lustre as the main file system. They do not import any networked file system like NFS, GPFS or CernVM-FS (the latter used by, e.g., Cern for distributing software for the Large Haedron Collider and the EESSI project). Instead these file systems are mounted on external servers in the admin section of the cluster and the Cray Data Virtualisation Service (DVS) is then used to access those file systems from the compute nodes over the high-speed interconnect.
Low-noise mode¶
Low-noise mode has meant different things throughout the history of Cray systems. Sometimes the mode described above, using only a selection of the regular Linux daemons on the compute nodes, was already called low-noise mode while some Cray systems provided another mode in which those daemons were activated. Depending on the cluster this was then called "emulation mode" or "Cluster Compatibility Mode". The latter is not implemented on LUMI, and even if it would, compatibility would still be limited by the special requirements to use the Slingshot interconnect and to have GPU-aware communication over Slingshot. The mode we use on the LUMI CPU-nodes used to be called "Extreme Scalability Mode" (ESM) in the Cray documentation.
However, it turned out that even the noise reduction described above was not yet sufficient to pass some large-scale scalability tests, and therefore another form of "low-noise" mode is implemented on the GPU nodes of LUMI where OS processes are restricted to a reserved core, actually core 0. Cray calls this "core specialization" and this idea already goes back to around 2010 when we got the first compute nodes with more than 20 cores (with the AMD "Magny Cours" Opteron). This leaves us with an asymmetric structure of the node, where the first CCD has 7 available cores while the other ones have 8, but as that created a headache for users to get a proper distribution of tasks and threads over the CPU (see the "Process and thread distribution and binding" chapter), the choice was made to also disable the first core on each of the other CCDs so that users now effectively see a 56-core node with 8 CCDs with 7 cores each.
Programming models¶
On LUMI we have several C/C++ and Fortran compilers. These will be discussed more in this session.
There is also support for MPI and SHMEM for distributed applications. And we also support RCCL, the ROCm-equivalent of the CUDA NCCL library that is popular in machine learning packages.
All compilers have some level of OpenMP support, and two compilers support OpenMP offload to the AMD GPUs, but again more about that later.
OpenACC, the other directive-based model for GPU offloading, is only supported in the Cray Fortran compiler. There is no commitment of neither HPE Cray or AMD to extend that support to C/C++ or other compilers, even though there is work going on in the LLVM community and several compilers on the system are based on LLVM.
The other important programming model for AMD GPUs is HIP (Heterogeneous-Compute Interface for Portability), which is their alternative for the proprietary CUDA model. It does not support all CUDA features though (basically it is more CUDA 7 or 8 level) and there is also no equivalent to CUDA Fortran.
The commitment to OpenCL is very unclear, and this actually holds for other GPU vendors also.
We also try to provide SYCL as it is a programming language/model that works on all three GPU families currently used in HPC.
Python is of course pre-installed on the system but we do ask to use big Python installations in a special way as Python puts a tremendous load on the file system. More about that later in this course.
Some users also report some success in running Julia. We don't have full support though and have to depend on binaries as provided by julialang.org.
It is important to realise that there is no CUDA on AMD GPUs and there will never be as this is a proprietary technology that other vendors cannot implement. The visualisation nodes in LUMI have NVIDIA rendering GPUs but these nodes are meant for visualisation and not for compute.
The development environment on LUMI¶
Long ago, Cray designed its own processors and hence had to develop their own compilers. They kept doing so, also when they moved to using more standard components, and had a lot of expertise in that field, especially when it comes to the needs of scientific codes, programming models that are almost only used in scientific computing or stem from such projects. As they develop their own interconnects, it does make sense to also develop an MPI implementation that can use the interconnect in an optimal way. They also have a long tradition in developing performance measurement and analysis tools and debugging tools that work in the context of HPC.
The first important component of the HPE Cray Programming Environment is the compilers. Cray still builds its own compilers for C/C++ and Fortran, called the Cray Compiling Environment (CCE). Furthermore, the GNU compilers are also supported on every Cray system, though at the moment AMD GPU support is not enabled. Depending on the hardware of the system other compilers will also be provided and integrated in the environment. On LUMI two other compilers are available: the AMD AOCC compiler for CPU-only code and the AMD ROCm compilers for GPU programming. Both contain a C/C++ compiler based on Clang and LLVM and a Fortran compiler which is currently based on the former PGI frontend with LLVM backend. The ROCm compilers also contain the support for HIP, AMD's CUDA clone.
The second component is the Cray Scientific and Math libraries, containing the usual suspects as BLAS, LAPACK and ScaLAPACK, and FFTW, but also some data libraries and Cray-only libraries.
The third component is the Cray Message Passing Toolkit. It provides an MPI implementation optimized for Cray systems, but also the Cray SHMEM libraries, an implementation of OpenSHMEM 1.5.
The fourth component is some Cray-unique sauce to integrate all these components, and support for hugepages to make memory access more efficient for some programs that allocate huge chunks of memory at once.
Other components include the Cray Performance Measurement and Analysis Tools and the Cray Debugging Support Tools that will not be discussed in this one-day course, and Python and R modules that both also provide some packages compiled with support for the Cray Scientific Libraries.
Besides the tools provided by HPE Cray, several of the development tools from the ROCm stack are also available on the system while some others can be user-installed (and one of those, the Grafana-based GUI for Omniperf, is not available due to security concerns). Furthermore there are some third party tools available on LUMI, including Linaro Forge (previously ARM Forge) and Vampir and some open source profiling tools.
Specifically not on LUMI is the Intel programming environment, nor is the
regular Intel oneAPI HPC Toolkit. The classic Intel compilers pose problems on AMD CPUs
as -xHost
cannot be relied on, but it appears that the new compilers that are based on
Clang and an LLVM backend behave better. Various MKL versions are also troublesome, with
different workarounds for different versions, though here also it seems that Intel now has
code that works well on AMD for many MKL routines. We have experienced problems with Intel
MPI when testing it on LUMI though in principle it should be possible to use Cray MPICH as
they are derived from the same version of MPICH.
The NVIDIA programming environment doesn't make sense on an AMD GPU system,
but it could be useful for some visualisation software on the
visualisation nodes so it is currently installed on those nodes.
We will now discuss some of these components in a little bit more detail, but refer to the 4-day trainings that we organise several times a year with HPE for more material.
The Cray Compiling Environment¶
The Cray Compiling Environment are the default compilers on many Cray systems and on LUMI. These compilers are designed specifically for scientific software in an HPC environment. The current versions are LLVM-based with extensions by HPE Cray for automatic vectorization and shared memory parallelization, technology that they have experience with since the late '70s or '80s.
The compiler offers extensive standards support. The C and C++ compiler is essentially their own build of Clang with LLVM with some of their optimisation plugins and OpenMP run-time. The version numbering of the CCE currently follows the major versions of the Clang compiler used. The support for C and C++ language standards corresponds to that of Clang. The Fortran compiler uses a frontend and optimiser developed by HPE Cray, but an LLVM-based code generator. The compiler supports most of Fortran 2018 (ISO/IEC 1539:2018). The CCE Fortran compiler is known to be very strict with language standards. Programs that use GNU or Intel extensions will usually fail to compile, and unfortunately since many developers only test with these compilers, much Fortran code is not fully standards compliant and will fail.
All CCE compilers support OpenMP, with offload for AMD and NVIDIA GPUs. In their most recent versions, they claim full OpenMP 5.0 support with partial (and growing) support for OpenMP 5.1 and 5.2. More information about the OpenMP support is found by checking a manual page:
man intro_openmp
cce
module is loaded,
or the web version of that page
which may be for a more recent version of the programming environment than available on LUMI.
The Fortran compiler also supports OpenACC for AMD and NVIDIA GPUs. That implementation
claims to be fully OpenACC 2.0 compliant, and offers partial support for OpenACC 2.x/3.x.
Information is available via
man intro_openacc
The CCE compilers also offer support for some PGAS (Partitioned Global Address Space) languages. UPC 1.2 is supported, as is Fortran 2008 coarray support. These implementations do not require a preprocessor that first translates the code to regular C or Fortran. There is also support for debugging with Linaro Forge.
Lastly, there are also bindings for MPI.
Scientific and math libraries¶
Cray Scientific and Math Libraries overview web page
Some mathematical libraries have become so popular that they basically define an API for which several implementations exist, and CPU manufacturers and some open source groups spend a significant amount of resources to make optimal implementations for each CPU architecture.
The most notorious library of that type is BLAS, a set of basic linear algebra subroutines for vector-vector, matrix-vector and matrix-matrix implementations. It is the basis for many other libraries that need those linear algebra operations, including Lapack, a library with solvers for linear systems and eigenvalue problems.
The HPE Cray LibSci library contains BLAS and its C-interface CBLAS, and LAPACK and its C interface LAPACKE. It also adds ScaLAPACK, a distributed memory version of LAPACK, and BLACS, the Basic Linear Algebra Communication Subprograms, which is the communication layer used by ScaLAPACK. The BLAS library combines implementations from different sources, to try to offer the most optimal one for several architectures and a range of matrix and vector sizes.
LibSci also contains one component which is HPE Cray-only: IRT, the Iterative Refinement Toolkit, which allows to do mixed precision computations for LAPACK operations that can speed up the generation of a double precision result with nearly a factor of two for those problems that are suited for iterative refinement. If you are familiar with numerical analysis, you probably know that the matrix should not be too ill-conditioned for that.
There is also a GPU-optimized version of LibSci, called LibSci_ACC, which contains a subset of the routines of LibSci. We or the LUMI User Support Team don't have much experience with this library though. It can be compared with what Intel is doing with oneAPI MKL which also offers GPU versions of some of the traditional MKL routines.
Another separate component of the scientific and mathematical libraries is FFTW3, Fastest Fourier Transforms in the West, which comes with optimized versions for all CPU architectures supported by recent HPE Cray machines.
Finally, the scientific and math libraries also contain HDF5 and netCDF libraries in sequential and parallel versions. These are included because it is essential that they interface properly with MPI parallel I/O and the Lustre file system to offer the best bandwidth to and from storage.
Cray used to offer more pre-installed third party libraries for which the only added value was that they compiled the binaries. Instead they now offer build scripts in a GitHub repository.
Cray MPI¶
HPE Cray build their own MPI library with optimisations for their own interconnects. The Cray MPI library is derived from the ANL MPICH 3.4 code base and fully supports the ABI (Application Binary Interface) of that implementation which implies that in principle it should be possible to swap the MPI library of applications build with that ABI with the Cray MPICH library. Or in other words, if you can only get a binary distribution of an application and that application was build against an MPI library compatible with the MPICH 3.4 ABI (which includes Intel MPI), it should be possible to exchange that library for the Cray one to have optimised communication on the Cray Slingshot interconnect.
Cray MPI contains many tweaks specifically for Cray systems. HPE Cray claim improved algorithms for many collectives, an asynchronous progress engine to improve overlap of communications and computations, customizable collective buffering when using MPI-IO, and optimized remote memory access (MPI one-sided communication) which also supports passive remote memory access.
When used in the correct way (some attention is needed when linking applications) it is also fully GPU aware with currently support for AMD and NVIDIA GPUs.
The MPI library also supports bindings for Fortran 2008.
MPI 3.1 is almost completely supported, with two exceptions. Dynamic process management is not
supported (and a problem anyway on systems with batch schedulers), and when using CCE
MPI_LONG_DOUBLE
and MPI_C_LONG_DOUBLE_COMPLEX
are also not supported.
The Cray MPI library does not support the mpirun
or mpiexec
commands, which is in fact
allowed by the standard which only requires a process starter and suggest mpirun
or mpiexec
depending on the version of the standard. Instead the Slurm srun
command is used as the
process starter. This actually makes a lot of sense as the MPI application should be mapped
correctly on the allocated resources, and the resource manager is better suited to do so.
Cray MPI on LUMI is layered on top of libfabric, which in turn uses the so-called Cassini provider to interface with the hardware. UCX is not supported on LUMI (but Cray MPI can support it when used on InfiniBand clusters). It also uses a GPU Transfer Library (GTL) for GPU-aware MPI.
GPU-aware MPI¶
Cray MPICH does support GPU-aware MPI, so it is possible to directly use GPU-attached communication buffers using device pointers. The implementation supports (a) GPU-NIC RDMA for efficient inter-node MPI transfers and (b) GPU Peer2Peer IPC for efficient intra-node transfers. The latter mechanism comes with some restrictions though that we will discuss in the chapter "Process and thread distribution and binding".
GPU-aware MPI needs to be enabled explicitly, which you can do by setting an environment variable:
export MPICH_GPU_SUPPORT_ENABLED=1
In addition to this, if the GPU code does use MPI operations that access GPU-attached memory regions it is best to also set
export MPICH_OFI_NIC_POLICY=GPU
to tell MPICH to always use the NIC closest to the GPU.
If only CPU communication buffers are used, then it may be better to set
export MPICH_OFI_NIC_POLICY=NUMA
which tells MPICH to use the NIC closest to the CPU NUMA domain.
Depending on how Slurm is used, Peer2Peer IPC may not work and in those cases you may want to turn it off using
export MPICH_GPU_IPC_ENABLED=0
export MPICH_SMP_SINGLE_COPY_MODE=NONE
More information about Cray MPICH and the many environment variables to fine-tune performance can be found in the manual page
man intro_mpi
or its web-based version which may be for a newer version than available on LUMI.
Lmod¶
Virtually all clusters use modules to enable the users to configure the environment and select the versions of software they want. There are three different module systems around. One is an old implementation that is hardly evolving anymore but that can still be found on a number of clusters. HPE Cray still offers it as an option. Modulefiles are written in TCL, but the tool itself is in C. The more popular tool at the moment is probably Lmod. It is largely compatible with modulefiles for the old tool, but prefers modulefiles written in LUA. It is also supported by the HPE Cray PE and is our choice on LUMI. The final implementation is a full TCL implementation developed in France and also in use on some large systems in Europe.
Fortunately the basic commands are largely similar in those implementations, but what differs is the way to search for modules. We will now only discuss the basic commands, the more advanced ones will be discussed in the next session of this tutorial course.
Modules also play an important role in configuring the HPE Cray PE, but before touching that topic we present the basic commands:
module avail
: Lists all modules that can currently be loaded.module list
: Lists all modules that are currently loadedmodule load
: Command used to load a module. Add the name and version of the module.module unload
: Unload a module. Using the name is enough as there can only one version be loaded of a module.module swap
: Unload the first module given and then load the second one. In Lmod this is really equivalent to amodule unload
followed by amodule load
.
Lmod supports a hierarchical module system. Such a module setup distinguishes between installed modules and available modules. The installed modules are all modules that can be loaded in one way or another by the module systems, but loading some of those may require loading other modules first. The available modules are the modules that can be loaded directly without loading any other module. The list of available modules changes all the time based on modules that are already loaded, and if you unload a module that makes other loaded modules unavailable, those will also be deactivated by Lmod. The advantage of a hierarchical module system is that one can support multiple configurations of a module while all configurations can have the same name and version. This is not fully exploited on LUMI, but it is used a lot in the HPE Cray PE. E.g., the MPI libraries for the various compilers on the system all have the same name and version yet make different binaries available depending on the compiler that is being used.
Different configurations on some Belgian clusters
Depending on the configuration Lmod can behave rather differently on different systems. Some clusters in Belgium have Lmod configured to mimic the original Tcl module system better rather than exposing the full power of Lmod.
E.g., on LUMI, module swap
isn't really needed as the auto-swap feature of Lmod is
enabled. Automatically unloading a module if another module with the same name is
loaded, is also enabled. Both features make with a hierarchical scheme much more powerful
and using the HPE Cray PE with these features disabled would be very difficult.
Compiler wrappers¶
The HPE Cray PE compilers are usually used through compiler wrappers.
The wrapper for C is cc
, the one for C++ is CC
and the one for Fortran is ftn
.
The wrapper then calls the selected compiler. Which compiler will be called is determined
by which compiler module is loaded. As shown on the slide
"Development environment on LUMI", on LUMI
the Cray Compiling Environment (module cce
), GNU Compiler Collection (module gcc-native
or gcc
depending on the version of the Cray Programming Environment),
the AMD Optimizing Compiler for CPUs (module aocc
) and the ROCm LLVM-based compilers
(module amd
) are available. On the visualisation nodes, the NVIDIA HPC compiler is currently
also installed (module nvhpc
). On other HPE Cray systems, you may also find the Intel
compilers.
The target architectures for CPU and GPU are also selected through modules, so it is better
to not use compiler options such as -march=native
. This makes cross compiling also easier.
The wrappers will also automatically link in certain libraries, and make the include files available, depending on which other modules are loaded. In some cases it tries to do so cleverly, like selecting an MPI, OpenMP, hybrid or sequential option depending on whether the MPI module is loaded and/or OpenMP compiler flag is used. This is the case for:
- The MPI libraries. There is no
mpicc
,mpiCC
,mpif90
, etc. on LUMI (well, there is nowadays, but their use is discouraged). The regular compiler wrappers do the job as soon as thecray-mpich
module is loaded. - LibSci and FFTW are linked automatically if the corresponding modules are loaded. So no need
to look, e.g., for the BLAS or LAPACK libraries: They will be offered to the linker if the
cray-libsci
module is loaded (and it is an example of where the wrappers try to take the right version based not only on compiler, but also on whether MPI is loaded or not and the OpenMP compiler flag). - netCDF and HDF5
It is possible to see which compiler and linker flags the wrappers add through the -craype-verbose
flag.
The wrappers do have some flags of their own, but also accept all flags of the selected compiler and simply pass those to those compilers.
The compiler wrappers are provided by the craype
module (but you don't have to load that module by hand).
Selecting the version of the CPE¶
The version numbers of the HPE Cray PE are of the form yy.dd
, e.g., 24.03
for the version
released in March 2024. There are several releases each year, but not all
of them are offered on LUMI as installing them may require downtime and as changes in
the system every few months are not appreciated by all users.
There is always a default version assigned by the sysadmins when installing the programming
environment. It is possible to change the default version for loading further modules
by loading one of the versions of the cpe
module. E.g., assuming the 24.03 version would be
present on the system, it can be loaded through
module load cpe/24.03
module load
twice will fix this:
module load cpe/24.03
module load cpe/24.03
module load
of cpe
when one is already loaded, as it then first unloads the already
loaded cpe
module). The warning can be ignored, but keep in mind that what it says is true,
it cannot restore the environment you found on LUMI at login.
The cpe
module is also not needed when using the LUMI software stacks, but more about that
later.
The target modules¶
The target modules are used to select the CPU and GPU optimization targets and to select the network communication layer.
On LUMI there are three CPU target modules that are relevant:
craype-x86-rome
selects the Zen2 CPU family code named Rome. These CPUs are used on the login nodes and the nodes of the data analytics and visualisation partition of LUMI. However, as Zen3 is a superset of Zen2, software compiled to this target should run everywhere, but may not exploit the full potential of the LUMI-C and LUMI-G nodes (though the performance loss is likely minor).craype-x86-milan
is the target module for the Zen3 CPUs code named Milan that are used on the CPU-only compute nodes of LUMI (the LUMI-C partition).craype-x86-trento
is the target module for the Zen3 CPUs code named Trento that are used on the GPU compute nodes of LUMI (the LUMI-G partition).
Two GPU target modules are relevant for LUMI:
craype-accel-gfx90a
: Compile offload code for the MI200 series GPUs that are used on LUMI-G.craype-accel-host
: Will tell some compilers to compile offload code for the host instead.
Two network target modules are relevant for LUMI:
craype-network-ofi
selects the libfabric communication layer which is needed for Slingshot 11.craype-network-none
omits all network specific libraries.
The compiler wrappers also have corresponding compiler flags that can be used to overwrite these
settings: -target-cpu
, -target-accel
and -target-network
.
PrgEnv and compiler modules¶
In the HPE Cray PE, the PrgEnv-*
modules are usually used to load a specific variant of the
programming environment. These modules will load the compiler wrapper (craype
), compiler,
MPI and LibSci module and may load some other modules also.
The following table gives an overview of the available PrgEnv-*
modules and the compilers they
activate:
PrgEnv | Description | Compiler module | Compilers |
---|---|---|---|
PrgEnv-cray | Cray Compiling Environment | cce |
craycc , crayCC , crayftn |
PrgEnv-gnu | GNU Compiler Collection | gcc-native gcc (*) |
gcc-13 , g++-13 , gfortran-13 gcc , g++ , gfortran |
PrgEnv-aocc | AMD Optimizing Compilers (CPU only) |
aocc |
clang , clang++ , flang |
PrgEnv-amd | AMD ROCm LLVM compilers (GPU support) |
amd |
amdclang , amdclang++ , amdflang |
(*) In the 23.12 programming environment, HPE Cray switched from own-built gcc binaries to
those from SUSE, and the module name was changed to gcc-native
to reflect this change.
There is also a second module that offers the AMD ROCm environment, rocm
. That module
has to be used with PrgEnv-cray
, PrgEnv-gnu
and recently also with PrgEnv-amd
to
enable MPI-aware GPU, hipcc with the GNU compilers or GPU support with the Cray compilers.
Changes to the GNU compilers since 23.12
The HPE Cray PE has changed the way it offers the GNU compilers in releases starting from 23.12. Rather than packaging the GNU compilers, HPE Cray now uses the development compiler packages of SUSE Linux (not to be confused with the system default which is still 7.5, the compiler that was offered with the initial release of SUSE Enterprise Linux 15).
In releases up to the 23.09, the GNU compilers are offered through the
gcc
compiler module. When loaded, it adds newer versions of the gcc
, g++
and gfortran
compilers
to the path, calling the version indicated by the version of the gcc
module.
In releases from 23.12 on, that compiler module is now called gcc-native
, and the compilers
now have the major version attached to the name of binary, e.g.,
gcc-13
, g++-13
and gfortran-13
, while
gcc
, g++
and gfortran
will compile with version 7.5, the default version for SUSE 15.
Getting help¶
Help on the HPE Cray Programming Environment is offered mostly through manual pages and compiler flags. Online help is limited and difficult to locate.
For the compilers, the following man pages are relevant:
PrgEnv | C | C++ | Fortran |
---|---|---|---|
PrgEnv-cray | man craycc |
man crayCC |
man crayftn |
PrgEnv-gnu | man gcc-13 |
man g++-13 |
man gfortran-13 |
PrgEnv-aocc/PrgEnv-amd | - | - | - |
Wrappers | man 1 cc |
man 1 CC |
man 1 ftn |
Recently, HPE Cray have also created a web version of some of the CPE documentation.
Some compilers also support the --help
flag, e.g., amdclang --help
. For the wrappers,
the switch -help
should be used instead as the double dash version is passed to the
compiler.
The wrappers have a number of options specific to them. Information about them can be obtained
by using the --craype-help
flag with the wrappers.
The wrappers also support the -dumpversion
flag to show the version of the underlying compiler.
Many other commands, including the actual compilers, use --version
to show the version.
For Cray Fortran compiler error messages, the explain
command is also helpful. E.g.,
$ ftn
ftn-2107 ftn: ERROR in command line
No valid filenames are specified on the command line.
$ explain ftn-2107
Error : No valid filenames are specified on the command line.
At least one file name must appear on the command line, with any command-line
options. Verify that a file name was specified, and also check for an
erroneous command-line option which may cause the file name to appear to be
an argument to that option.
On older Cray systems this used to be a very useful command with more compilers but as
HPE Cray is using more and more open source components instead there are fewer commands
that give additional documentation via the explain
command.
Lastly, there is also a lot of information in the "Developing" section of the LUMI documentation.
Google, ChatGPT and LUMI¶
When looking for information on the HPE Cray Programming Environment using search engines such as Google, you'll be disappointed how few results show up. HPE doesn't put much information on the internet, and the environment so far was mostly used on Cray systems of which there are not that many.
The same holds for ChatGPT. In fact, much of the training of the current version of ChatGPT was done with data of two or so years ago and there is not that much suitable training data available on the internet either.
The HPE Cray environment has a command line alternative to search engines though: the man -K
command
that searches for a term in the manual pages. It is often useful to better understand some error messages.
E.g., sometimes Cray MPICH will suggest you to set some environment variable to work around some problem.
You may remember that man intro_mpi
gives a lot of information about Cray MPICH, but if you don't and,
e.g., the error message suggests you to set FI_CXI_RX_MATCH_MODE
to either software
or hybrid
, one way
to find out where you can get more information about this environment variable is
man -K FI_CXI_RX_MATCH_MODE
The online documentation is now also complete enough that it makes sense trying the search box on that page instead.
Other modules¶
Other modules that are relevant even to users who do not do development:
- MPI:
cray-mpich
. - LibSci:
cray-libsci
- Cray FFTW3 library:
cray-fftw
- HDF5:
cray-hdf5
: Serial HDF5 I/O librarycray-hdf5-parallel
: Parallel HDF5 I/O library
- NetCDF:
cray-netcdf
cray-netcdf-hdf5parallel
cray-parallel-netcdf
- Python:
cray-python
, already contains a selection of packages that interface with other libraries of the HPE Cray PE, including mpi4py, NumPy, SciPy and pandas. - R:
cray-R
The HPE Cray PE also offers other modules for debugging, profiling, performance analysis, etc. that are not covered in this short version of the LUMI course. Many more are covered in the 4-day courses for developers that we organise several times per year with the help of HPE and AMD.
Warning 1: You do not always get what you expect...¶
The HPE Cray PE packs a surprise in terms of the libraries it uses, certainly for users who come from an environment where the software is managed through EasyBuild, but also for most other users.
The PE does not use the versions of many libraries determined by the loaded modules at runtime
but instead uses default versions of libraries (which are actually in /opt/cray/pe/lib64
on the system)
which correspond to the version of the programming environment that is set as the system default when installed.
This is very much the behaviour of Linux applications also that pick standard libraries in a few standard
directories and it enables many programs built with the HPE Cray PE to run without reconstructing the
environment and in some cases to mix programs compiled with different compilers with ease (with the
emphasis on some as there may still be library conflicts between other libraries when not using the
so-called rpath linking). This does have an annoying side effect though: If the default PE on the system
changes, all applications will use different libraries and hence the behaviour of your application may
change.
Luckily there are some solutions to this problem.
By default the Cray PE uses dynamic linking, and does not use rpath linking, which is a form of dynamic
linking where the search path for the libraries is stored in each executable separately. On Linux, the
search path for libraries is set through the environment variable LD_LIBRARY_PATH
. Those Cray PE modules
that have their libraries also in the default location, add the directories that contain the actual version
of the libraries corresponding to the version of the module to the PATH-style environment variable
CRAY_LD_LIBRARY_PATH
. Hence all one needs to do is to ensure that those directories are put in
LD_LIBRARY_PATH
which is searched before the default location:
export LD_LIBRARY_PATH=$CRAY_LD_LIBRARY_PATH:$LD_LIBRARY_PATH
Small demo of adapting LD_LIBRARY_PATH
:
An example that can only be fully understood after the section on the LUMI software stacks:
$ module load LUMI/23.09
$ module load lumi-CPEtools/1.1-cpeGNU-239.09
$ ldd $EBROOTLUMIMINCPETOOLS/bin/mpi_check
linux-vdso.so.1 (0x00007fff399fe000)
libdl.so.2 => /lib64/libdl.so.2 (0x00007feccb848000)
libmpi_gnu_91.so.12 => /opt/cray/pe/lib64/libmpi_gnu_91.so.12 (0x00007fecc8a00000)
...
$ export LD_LIBRARY_PATH=$CRAY_LD_LIBRARY_PATH:$LD_LIBRARY_PATH
$ ldd $EBROOTLUMIMINCPETOOLS/bin/mpi_check
linux-vdso.so.1 (0x00007ffea02a5000)
libdl.so.2 => /lib64/libdl.so.2 (0x00007f9945496000)
libmpi_gnu_91.so.12 => /opt/cray/pe/mpich/8.1.27/ofi/gnu/9.1/lib/libmpi_gnu_91.so.12 (0x00007f9942600000)
...
ldd
command shows which libraries are used by an executable. Only a part of the very long
output is shown in the above example. But we can already see that in the first case, the library
libmpi_gnu_91.so.12
is taken from opt/cray/pe/lib64
which is the directory with the default
versions, while in the second case it is taken from /opt/cray/pe/mpich/8.1.27/ofi/gnu/9.1/lib/
which clearly is for a specific version of cray-mpich
.
We do provide the module lumi-CrayPath
that tries to fix LD_LIBRARY_PATH
in a way that unloading
the module fixes LD_LIBRARY_PATH
again to the state before adding CRAY_LD_LIBRARY_PATH
and that
reloading the module adapts LD_LIBRARY_PATH
to the current value of CRAY_LD_LIBRARY_PATH
. Loading that
module after loading all other modules should fix this issue for most if not all software.
The second solution would be to use rpath-linking for the Cray PE libraries, which can be done by setting
the CRAY_ADD_RPATH
environment variable:
export CRAY_ADD_RPATH=yes
However, there is also a good side to the standard Cray PE behaviour. Updates of the underlying operating system or network software stack may break older versions of the MPI library. By letting the applications use the default libraries and updating the defaults to a newer version, most applications will still run while they would fail if any of the two tricks to force the use of the intended library version are used. This has actually happened after a big LUMI update in March 2023, when all software that used rpath-linking had to be rebuilt as the MPICH library that was present before the update did not longer work.
Warning 2: Order matters¶
Lmod is a hierarchical module scheme and this is exploited by the HPE Cray PE. Not all modules are available right away and some only become available after loading other modules. E.g.,
cray-fftw
only becomes available when a processor target module is loadedcray-mpich
requires both the network target modulecraype-network-ofi
and a compiler module to be loadedcray-hdf5
requires a compiler module to be loaded andcray-netcdf
in turn requirescray-hdf5
but there are many more examples in the programming environment.
In the next section of the course we will see how unavailable modules can still be found
with module spider
. That command can also tell which other modules should be loaded
before a module can be loaded, but unfortunately due to the sometimes non-standard way
the HPE Cray PE uses Lmod that information is not always complete for the PE, which is also
why we didn't demonstrate it here.
Note: Compiling without the HPE Cray PE wrappers¶
It is now possible to work without the HPE Cray PE compiler wrappers and to use the compilers in a way you may be more familiar with from other HPC systems.
In that case, you would likely want to load a compiler module without loading the
PrgEnv-*
module and craype
module (which would be loaded automatically by the
PrgEnv-*
module). The compiler module and compiler driver names are then given
by the following table:
Description | Compiler module | Compilers |
---|---|---|
Cray Compiling Environment | cce |
craycc , crayCC , crayftn |
PrgEnv-gnu | gcc-native gcc (*) |
gcc-13 , g++-13 , gfortran-13 gcc , g++ , gfortran |
AMD Optimizing Compilers (CPU only) |
aocc |
clang , clang++ , flang |
AMD ROCm LLVM compilers (GPU support) |
amd |
amdclang , amdclang++ , amdflang |
Recent versions of the cray-mpich
module now also provide the traditional MPI compiler wrappers
such as mpicc
, mpicxx
or mpifort
. Note that you will still need to ensure that the network
target module craype-network-ofi
is loaded to be able to load the cray-mpich
module!
The cray-mpich
module also defines the environment variable MPICH_DIR
that points to the
MPI installation for the selected compiler.
To manually use the BLAS and LAPACK libraries, you'll still have to load the cray-libsci
module.
This module defines the CRAY_PE_LIBSCI_PREFIX_DIR
environment variable that points to the directory
with the library and include file subdirectories for the selected compiler.
(Or CRAY_LIBSCI_PREFIX_DIR
before release 23.12 of the programming environment.)
See the intro_libsci
manual page
for information about the different libraries.
To be able to use the cray-fftw
FFTW libraries, you still need to load the right CPU target module,
even though you need to specify the target architecture yourself now when calling the compilers.
This is because the HPE Cray PE does not come with a multi-cpu version of the FFTW libraries, but
specific versions for each CPU (or sometimes group of similar CPUs). Here again some environment
variables may be useful to point the compiler and linker to the installation: FFTW_ROOT
for the
root of the installation for the specific CPU (the library is otherwise compiler-independent),
FFTW_INC
for the subdirectory with the include files and FFTW_DIR
for the directory with the
libraries.
Other modules that you may want to use also typically define some useful environment variables.