HPC Challenge Benchmark

From Wikipedia, the free encyclopedia
HPC Challenge Benchmark
Original author(s)Innovative Computing Laboratory, University of Tennessee
Initial release2003 (2003)
Stable release
1.5.0 / March 18, 2016; 8 years ago (2016-03-18)[1]
PlatformCross-platform
LicenseBSD
Websiteicl.cs.utk.edu/hpcc/

HPC Challenge Benchmark combines several benchmarks to test a number of independent attributes of the performance of high-performance computer (HPC) systems. The project has been co-sponsored by the DARPA High Productivity Computing Systems program, the United States Department of Energy and the National Science Foundation.[2]

Context[edit]

The performance of complex applications on HPC systems can depend on a variety of independent performance attributes of the hardware. The HPC Challenge Benchmark is an effort to improve visibility into this multidimensional space by combining the measurement of several of these attributes into a single program.

Although the performance attributes of interest are not specific to any particular computer architecture, the reference implementation of the HPC Challenge Benchmark in C and MPI assumes that the system under test is a cluster of shared memory multiprocessor systems connected by a network. Due to this assumption of a hierarchical system structure most of the tests are run in several different modes of operation. Following the notation used by the benchmark reports, results labeled "single" mean that the test was run on one randomly chosen processor in the system, results labeled "star" mean that an independent copy of the test was run concurrently on each processor in the system, and results labeled "global" mean that all the processors were working in coordination to solve a single problem (with data distributed across the nodes of the system).

Components[edit]

The benchmark currently consists of 7 tests (with the modes of operation indicated for each):

  1. HPL[3] (High Performance LINPACK) – measures performance of a solver for a dense system of linear equations (global).
  2. DGEMM – measures performance for matrix-matrix multiplication (single, star).
  3. STREAM[4] – measures sustained memory bandwidth to/from memory (single, star).
  4. PTRANS – measures the rate at which the system can transpose a large array (global).
  5. RandomAccess – measures the rate of 64-bit updates to randomly selected elements of a large table (single, star, global).
  6. FFT – performs a Fast Fourier Transform on a large one-dimensional vector using the generalized Cooley–Tukey algorithm (single, star, global).
  7. Communication Bandwidth and Latency – MPI-centric performance measurements based on the b_eff[5] bandwidth/latency benchmark.

Performance attributes[edit]

At a high level, the tests are intended to provide coverage of four important attributes of performance: double-precision floating-point arithmetic (DGEMM and HPL), local memory bandwidth (STREAM), network bandwidth for "large" messages (PTRANS, RandomAccess, FFT, b_eff), and network bandwidth for "small" messages (RandomAccess, b_eff). Some of the codes are more complex than others and can have additional performance sensitivities. For example, in some systems HPL performance can be limited by network bandwidth and/or network latency.

Competition[edit]

The annual HPC Challenge Award Competition at the Supercomputing Conference focuses on four of the most challenging benchmarks in the suite:

There are two classes of awards:

  • Class 1: Best performance on a base or optimized run submitted to the HPC Challenge website.[6]
  • Class 2: Most "elegant" implementation of four or five computational kernels including three or more of the HPC Challenge benchmarks.[7]

See also[edit]

References[edit]

  1. ^ "Releases · icl-utk-edu/hpcc". github.com. Retrieved 2021-04-12.
  2. ^ "Cray X1 Supercomputer Has Highest Reported Scores on Government-Sponsored HPC Challenge Benchmark Tests". 2004-06-14. Archived from the original on 2009-03-30. Retrieved 2010-01-22.
  3. ^ "HPL – A Portable Implementation of the High-Performance Linpack Benchmark for Distributed-Memory Computers". Innovative Computing Laboratory, University of Tennessee at Knoxville. Retrieved 2015-06-10.
  4. ^ "STREAM: Sustainable Memory Bandwidth in High Performance Computers". Retrieved 2015-06-10.
  5. ^ "Effective Bandwidth (b_eff) Benchmark". High Performance Computing Center Stuttgart. Retrieved 2015-06-10.
  6. ^ The benchmark is designed to allow replacement of a limited set of functions with more highly optimized versions while remaining a "base" run. Additional (but still limited) modifications are allowed under the category of "optimized" runs.
  7. ^ "HPC Challenge Award Competition". DARPA HPCS Program. Retrieved 2010-01-23.

External links[edit]