Building Software
The phrase "building software" is a common way to describe the process of producing a machine-readable executable file from source files written in C, Fortran, or some other programming language. In its simplest form, building software involves a simple, one-line call or short shell script that invokes a compiler. More typically, the process leverages the power of makefiles, so you can change a line or two in the source code, then rebuild in a systematic way only the components affected by the change. Increasingly, however, the build process is a sophisticated multi-step automated workflow managed by a special framework like autotools or cmake
, intended to achieve a repeatable, maintainable, portable mechanism for installing software across a wide range of target platforms.
The Basics of Building Software
This section of the user guide does nothing more than introduce the big ideas with simple one-line examples. You will undoubtedly want to explore these concepts more deeply using online resources. You will quickly outgrow the examples here. We recommend that you master the basics of makefiles as quickly as possible: even the simplest computational research project will benefit enormously from the power and flexibility of a makefile-based build process.
Intel Compilers
Intel is the recommended and default compiler suite on Frontera. Each Intel module also gives you direct access to mkl
without loading an mkl
module; see Intel MKL for more information. Here are simple examples that use the Intel compiler to build an executable from source code:
Compiling a code that uses OpenMP would look like this:
$ icc -qopenmp mycode.c -o myexe # OpenMP
See the published Intel documentation, available both online and in ${TACC_INTEL_DIR}/documentation
, for information on optimization flags and other Intel compiler options.
GNU Compilers
The GNU foundation maintains a number of high quality compilers, including a compiler for C (gcc
), C++ (g++
), and Fortran (gfortran
). The gcc
compiler is the foundation underneath all three, and the term gcc
often means the suite of these three GNU compilers.
Load a gcc
module to access a recent version of the GNU compiler suite. Avoid using the GNU compilers that are available without a gcc
module — those will be older versions based on the "system gcc" that comes as part of the Linux distribution.
Here are simple examples that use the GNU compilers to produce an executable from source code:
$ gcc mycode.c # C source file; executable a.out $ gcc mycode.c -o myexe # C source file; executable myexe $ g++ mycode.cpp -o myexe # C++ source file $ gfortran mycode.f90 -o myexe # Fortran90 source file $ gcc -fopenmp mycode.c -o myexe # OpenMP; GNU flag is different than Intel
Note that some compiler options are the same for both Intel and GNU (e.g. -o
), while others are different (e.g. -qopenmp
vs -fopenmp
). Many options are available in one compiler suite but not the other. See the online GNU documentation for information on optimization flags and other GNU compiler options.
Compiling and Linking as Separate Steps
Building an executable requires two separate steps: (1) compiling (generating a binary object file associated with each source file); and (2) linking (combining those object files into a single executable file that also specifies the libraries that executable needs). The examples in the previous section accomplish these two steps in a single call to the compiler. When building more sophisticated applications or libraries, however, it is often necessary or helpful to accomplish these two steps separately.
Use the -c
("compile") flag to produce object files from source files:
$ icc -c main.c calc.c results.c
Barring errors, this command will produce object files main.o
, calc.o
, and results.o
. Syntax for other compilers Intel and GNU compilers is similar.
You can now link the object files to produce an executable file:
$ icc main.o calc.o results.o -o myexe
The compiler calls a linker utility (usually /bin/ld
) to accomplish this task. Again, syntax for other compilers is similar.
Include and Library Paths
Software often depends on pre-compiled binaries called libraries. When this is true, compiling usually requires using the -I
option to specify paths to so-called header or include files that define interfaces to the procedures and data in those libraries. Similarly, linking often requires using the -L
option to specify paths to the libraries themselves. Typical compile and link lines might look like this:
$ icc -c main.c -I${WORK}/mylib/inc -I${TACC_HDF5_INC} # compile $ icc main.o -o myexe -L${WORK}/mylib/lib -L${TACC_HDF5_LIB} -lmylib -lhdf5 # link
On Frontera, both the hdf5
and phdf5
modules define the environment variables $TACC_HDF5_INC
and $TACC_HDF5_LIB
. Other module files define similar environment variables; see Using Modules for more information.
The details of the linking process vary, and order sometimes matters. Much depends on the type of library: static (.a
suffix; library's binary code becomes part of executable image at link time) versus dynamically-linked shared (.so suffix; library's binary code is not part of executable; it's located and loaded into memory at run time). The link line can use rpath to store in the executable an explicit path to a shared library. In general, however, the LD_LIBRARY_PATH
environment variable specifies the search path for dynamic libraries. For software installed at the system-level, TACC's modules generally modify LD_LIBRARY_PATH
automatically. To see whether and how an executable named myexe
resolves dependencies on dynamically linked libraries, execute ldd myexe
.
A separate section below addresses the Intel Math Kernel Library (MKL).
Compiling and Linking MPI Programs
Intel MPI (module impi
) and MVAPICH2 (module mvapich2
) are the two MPI libraries available on Frontera. After loading an impi
or mvapich2
module, compile and/or link using an mpi wrapper (mpicc
, mpicxx
, mpif90
) in place of the compiler:
$ mpicc mycode.c -o myexe # C source, full build $ mpicc -c mycode.c # C source, compile without linking $ mpicxx mycode.cpp -o myexe # C++ source, full build $ mpif90 mycode.f90 -o myexe # Fortran source, full build
These wrappers call the compiler with the options, include paths, and libraries necessary to produce an MPI executable using the MPI module you're using. To see the effect of a given wrapper, call it with the -show
option:
$ mpicc -show # Show compile line generated by call to mpicc; similarly for other wrappers
Building Third-Party Software
You can discover already installed software using TACC's Software Search tool or execute module spider
or module avail
on the command-line.
You're welcome to download third-party research software and install it in your own account. In most cases you'll want to download the source code and build the software so it's compatible with the Frontera software environment. You can't use yum or any other installation process that requires elevated privileges, but this is almost never necessary. The key is to specify an installation directory for which you have write permissions. Details vary; you should consult the package's documentation and be prepared to experiment. When using the famous three-step autotools build process, the standard approach is to use the PREFIX
environment variable to specify a non-default, user-owned installation directory at the time you execute configure
or make
:
$ export INSTALLDIR=$WORK/apps/t3pio $ ./configure --prefix=$INSTALLDIR $ make $ make install
Other languages, frameworks, and build systems generally have equivalent mechanisms for installing software in user space. In most cases a web search like "Python Linux install local" will get you the information you need.
In Python, a local install will resemble one of the following examples:
$ pip install netCDF4 --user # install netCDF4 package to $HOME/.local $ python3 setup.py install --user # install to $HOME/.local $ pip3 install netCDF4 --prefix=$INSTALLDIR # custom location; add to PYTHONPATH
Similarly in R:
$ module load Rstats # load TACC's default R $ R # launch R > install.packages('devtools') # R will prompt for install location
You may, of course, need to customize the build process in other ways. It's likely, for example, that you'll need to edit a makefile
or other build artifacts to specify Frontera-specific include and library paths or other compiler settings. A good way to proceed is to write a shell script that implements the entire process: definitions of environment variables, module commands, and calls to the build utilities. Include echo
statements with appropriate diagnostics. Run the script until you encounter an error. Research and fix the current problem. Document your experience in the script itself; including dead-ends, alternatives, and lessons learned. Re-run the script to get to the next error, then repeat until done. When you're finished, you'll have a repeatable process that you can archive until it's time to update the software or move to a new machine.
If you wish to share a software package with collaborators, you may need to modify file permissions. See Sharing Files with Collaborators for more information.
Building Frontera
The Intel Math Kernel Library (MKL)
The Intel Math Kernel Library (MKL) is a collection of highly optimized functions implementing some of the most important mathematical kernels used in computational science, including standardized interfaces to:
- BLAS (Basic Linear Algebra Subroutines), a collection of low-level matrix and vector operations like matrix-matrix multiplication
- LAPACK (Linear Algebra PACKage), which includes higher-level linear algebra algorithms like Gaussian Elimination
- FFT (Fast Fourier Transform), including interfaces based on FFTW (Fastest Fourier Transform in the West)
- ScaLAPACK (Scalable LAPACK), BLACS (Basic Linear Algebra Communication Subprograms), Cluster FFT, and other functionality that provide block-based distributed memory (multi-node) versions of selected LAPACK, BLAS, and FFT algorithms;
- Vector Mathematics (VM) functions that implement highly optimized and vectorized versions of special functions like sine and square root.
MKL with Intel Compilers
There is no MKL module for the Intel compilers because you don't need one: the Intel compilers have built-in support for MKL. Unless you have specialized needs, there is no need to specify include paths and libraries explicitly. Instead, using MKL with the Intel modules requires nothing more than compiling and linking with the -mkl
option.; e.g.
$ icc -mkl mycode.c $ ifort -mkl mycode.c
The -mkl
switch is an abbreviated form of -mkl=parallel
, which links your code to the threaded version of MKL. To link to the unthreaded version, use -mkl=sequential
. A third option, -mkl=cluster
, which also links to the unthreaded libraries, is necessary and appropriate only when using ScaLAPACK or other distributed memory packages. For additional information, including advanced linking options, see Intel's MKL documentation and Intel MKL Link Line Advisor.
MKL with GNU Compilers
When using a GNU compiler, load the MKL module before compiling or running your code, then specify explicitly the MKL libraries, library paths, and include paths your application needs. Consult the Intel MKL Link Line Advisor for details. A typical compile/link process on a TACC system will look like this:
$ module load gcc $ module load mkl # available/needed only for GNU compilers $ gcc -fopenmp -I$MKLROOT/include \ -Wl,-L${MKLROOT}/lib/intel64 \ -lmkl_intel_lp64 -lmkl_core \ -lmkl_gnu_thread -lpthread \ -lm -ldl mycode.c
For your convenience the mkl
module file also provides alternative TACC-defined variables like $TACC_MKL_INCLUDE
(equivalent to $MKLROOT/include
). Execute module help mkl
for more information.
MKL with BLAS/LAPACK and Third-Party Software
When your third-party software requires BLAS or LAPACK, you can use MKL to supply this functionality. Replace generic instructions that include link options like -lblas
or -llapack
with the simpler MKL approach described above. There is no need to download and install alternatives like OpenBLAS.
MKL with BLAS/LAPACK and TACC's MATLAB, Python, and R Modules
TACC's MATLAB, Python, and R modules all use threaded (parallel) MKL as their underlying BLAS/LAPACK library. These means that even serial codes written in MATLAB, Python, or R may benefit from MKL's thread-based parallelism. This requires no action on your part other than specifying an appropriate max thread count for MKL.
Controlling Threading in MKL
Any code that calls MKL functions can potentially benefit from MKL's thread-based parallelism; this is true even if your code is not otherwise a parallel application. If you are linking to the threaded MKL (using -mkl
, -mkl=parallel
, or the equivalent explicit link line), you need only specify an appropriate value for the max number of threads available to MKL. You can do this with either of the two environment variables MKL_NUM_THREADS
or OMP_NUM_THREADS
. The environment variable MKL_NUM_THREADS
specifies the max number of threads available to each instance of MKL, and has no effect on non-MKL code. If MKL_NUM_THREADS
is undefined, MKL uses OMP_NUM_THREADS
to determine the max number of threads available to MKL functions. In either case, MKL will attempt to choose an optimal thread count less than or equal to the specified value. Note that OMP_NUM_THREADS
defaults to 1 on TACC systems; if you use the default value you will get no thread-based parallelism from MKL.
If you are running a single serial, unthreaded application (or an unthreaded MPI code involving a single MPI task per node) it is usually best to give MKL as much flexibility as possible by setting the max thread count to the total number of hardware threads on the node (56 on CLX). Of course things are more complicated if you are running more than one process on a node: e.g. multiple serial processes, threaded applications, hybrid MPI-threaded applications, or pure MPI codes running more than one MPI rank per node. See Settings for Calling Intel® Math Kernel Library Routines from Multi-Threaded Applications and related Intel resources for examples of how to manage threading when calling MKL from multiple processes.
Using ScaLAPACK, Cluster FFT, and Other MKL Cluster Capabilities
See Working with the Intel Math Kernel Library Cluster Software and Intel MKL Link Line Advisor for information on linking to the MKL cluster components.
Building for Performance on Frontera
Recommended Compiler
When building software on Frontera, we recommend using the Intel compiler and Intel MPI stack. This will be the default in the early user period, but may change if we determine one of the other MPI stacks provides superior performance.
Architecture-Specific Flags
To compile for CLX only, include -xCORE-AVX512
as a build option. The -x
switch allows you to specify a target architecture. The CLX chips, as well as the Skylake chips (SKX) on Stampede2, support Intel's latest instruction set, CORE-AVX512. You should also consider specifying an optimization level using the -O
flag:
$ icc -xCORE-AVX512 -O3 mycode.c -o myexe # will run only on CLX/SKX $ ifort -xCORE-AVX512 -O3 mycode.f90 -o myexe # will run only on CLX/SKX
It's best to avoid building with -xHost
(a flag that means "optimize for the architecture on which I'm compiling now"). Although this will work on Frontera, since the Frontera login nodes are all CLX nodes, if you build on another system, your binary will be based on whatever architecture you built upon. This may not be the same as the architecture on which you will be running.
Also, you should not use the -fast
flag for the Intel compiler. This flag sets the following options:
-ipo -O3 -no-prec-div -static -fp-model fast=2 -xHost
Frontera software libraries, including the MPI libraries, are installed as shared libraries in most cases. The -static
flag included in -fast
will cause the compile to fail at the link stage. If you’d like to use the other flags, you’ll have to include each option individually.
For information on the performance implications of your choice of build flags, see the sections on Programming and Performance for CLX.
If you use GNU compilers, see GNU x86 Options for information regarding support for CLX.