Applications/Hpl

From HPC
Revision as of 14:15, 7 June 2018 by Pysdlb (talk | contribs)

Jump to: navigation , search

Application Details

  • Description: HPL is a software package that solves a (random) dense linear system in double precision (64 bits) arithmetic on distributed-memory computers. It is freely available implementation of the High Performance Computing Linpack Benchmark.
  • Version: 11.3.1
  • Modules: hpl/intel-2016/11.3.1
  • Licence: Open source (See http://www.netlib.org/benchmark/hpl/copyright.html)

Usage Examples

HPL is a software package that solves a (random) dense linear system in double precision (64 bits) arithmetic on distributed-memory computers. It can thus be regarded as a portable as well as freely available implementation of the High Performance Computing Linpack Benchmark.

The algorithm used by HPL can be summarized by the following keywords: Two-dimensional block-cyclic data distribution - Right-looking variant of the LU factorization with row partial pivoting featuring multiple look-ahead depths - Recursive panel factorization with pivot search and column broadcast combined - Various virtual panel broadcast topologies - bandwidth reducing swap-broadcast algorithm - backward substitution with look-ahead of depth 1.

The HPL package provides a testing and timing program to quantify the accuracy of the obtained solution as well as the time it took to compute it. The best performance achievable by this software on your system depends on a large variety of factors. Nonetheless, with some restrictive assumptions on the interconnection network, the algorithm described here and its attached implementation are scalable in the sense that their parallel efficiency is maintained constant with respect to the per processor memory usage.

The HPL software package requires the availibility on your system of an implementation of the Message Passing Interface MPI (1.1 compliant). An implementation of either the Basic Linear Algebra Subprograms BLAS or the Vector Signal Image Processing Library VSIPL is also needed. Machine-specific as well as generic implementations of MPI, the BLAS and VSIPL are available for a large variety of systems.


Non interactive job

This runs on the scheduler SLURM

Compute Node

In this example we test the compute nodes, note we only specify 1 task per node here as HPL will use all of the cores available.


#!/bin/bash
#SBATCH -J compute-single-node
#SBATCH -N 1
#SBATCH --ntasks-per-node 1
#SBATCH -m cyclic:cyclic
#SBATCH -D /home/user/benchmarks/HPL/single_node/compute
#SBATCH -o %N.%j.%a.out
#SBATCH -e %N.%j.%a.err
#SBATCH -p compute
#SBATCH --exclusive

echo $SLURM_JOB_NODELIST

module purge
module add hpl/intel-2016/11.3.1

export I_MPI_DEBUG=5
export I_MPI_FABRICS=shm:tmi
export I_MPI_FALLBACK=no

xhpl_intel64

And passing it to SLURM

[username@login01 ~]$ sbatch HPLdemo.job
Submitted batch job 889552

GPU Node

In this example we test the GPU nodes, note we only specify 1 task per node here as HPL will use all of the cores available.


#!/bin/bash
#SBATCH -J gpu-single-node
#SBATCH -N 1
#SBATCH --ntasks-per-node 1
#SBATCH -m cyclic:cyclic
#SBATCH -D /home/cvsupport/benchmarks/HPL/single_node/gpu
#SBATCH -o %N.%j.%a.out
#SBATCH -e %N.%j.%a.err
#SBATCH -p gpu
#SBATCH --exclusive

echo $SLURM_JOB_NODELIST

module purge
module add hpl/intel-2016/11.3.1
module add intel/mkl/64/11.3.2

export PSM2_SDMA=0
export I_MPI_DEBUG=5
export I_MPI_FABRICS=shm:tmi
export I_MPI_FALLBACK=no

xhpl_intel64_dynamic

And passing it to SLURM, this time the GPU queue:

[username@login01 ~]$ sbatch HPLgpu-demo.job
Submitted batch job 882552

Further Information


Icon home.png