SPEChpc™ 2021 Tiny Result

Copyright 2021-2025 Standard Performance Evaluation Corporation

Intel

Hatch: Intel Server D50DNP1SB (Xeon Platinum
8480+)

SPEChpc 2021_tny_base = 86.80

SPEChpc 2021_tny_peak = 95.20

hpc2021 License: 13 Test Date: Apr-2025
Test Sponsor: Intel Hardware Availability: Jan-2023
Tested by: Intel Software Availability: Mar-2025

Benchmark result graphs are available in the PDF report.

Results Table

Benchmark Base Peak
Model Ranks Thrds/Rnk Seconds Ratio Seconds Ratio Seconds Ratio Model Ranks Thrds/Rnk Seconds Ratio Seconds Ratio Seconds Ratio
SPEChpc 2021_tny_base 86.80
SPEChpc 2021_tny_peak 95.20
Results appear in the order in which they were run. Bold underlined text indicates a median measurement.
505.lbm_t TGT 16 1 9.42 2390 7.99 2820 7.92 2840 TGT 16 1 7.91 2840 8.08 2780 7.99 2820
513.soma_t TGT 16 1 17.00 2180 16.90 2180 16.80 2200 TGT 16 1 15.80 2340 15.20 2440 15.20 2440
518.tealeaf_t TGT 16 1 17.80 92.8 17.80 92.7 17.60 93.5 TGT 16 1 17.60 93.9 17.70 93.4 18.10 91.1
519.clvleaf_t TGT 16 1 17.70 93.0 17.60 93.8 18.00 91.5 TGT 16 1 17.70 93.0 17.60 93.8 18.00 91.5
521.miniswp_t TGT 16 1 44.00 36.4 43.50 36.8 42.30 37.8 TGT 16 1 44.20 36.2 42.90 37.3 42.80 37.3
528.pot3d_t TGT 16 1 25.90 82.0 27.70 76.8 26.90 78.9 TGT 16 1 22.50 94.5 22.40 94.8 22.30 95.1
532.sph_exa_t TGT 16 1 39.90 48.8 40.20 48.5 40.00 48.8 TGT 16 1 37.80 51.6 37.80 51.5 37.80 51.6
534.hpgmgfv_t TGT 16 1 50.60 23.2 50.80 23.1 50.60 23.2 TGT 16 1 32.20 36.5 33.60 35.0 32.90 35.7
535.weather_t TGT 16 1 20.10 1600 20.10 1610 20.00 1610 TGT 16 1 19.40 1660 19.50 1650 19.60 1650
Hardware Summary
Type of System: Homogenous Cluster
Compute Node: Intel Server D50DNP1SB (Xeon Platinum 8480+)
Interconnect: Mellanox HDR
Compute Nodes Used: 2
Total Chips: 4
Total Cores: 224
Total Threads: 448
Total Memory: 2 TB
Max. Peak Threads: 1
Software Summary
Compiler: Intel oneAPI Compiler 2025.1.0
MPI Library: Intel MPI Library 2021.15 for Linux OS
Other MPI Info: None
Other Software: None
Base Parallel Model: TGT
Base Ranks Run: 16
Base Threads Run: 1
Peak Parallel Models: TGT
Minimum Peak Ranks: 16
Maximum Peak Ranks: 16
Max. Peak Threads: 1
Min. Peak Threads: 1

Node Description: Intel Server D50DNP1SB (Xeon Platinum 8480+)

Hardware
Number of nodes: 2
Uses of the node: Compute
Vendor: Intel
Model: Intel Server D50DNP1SB (2 x Intel Xeon
Platinum 8480+, 2.0GHz)
CPU Name: Intel Xeon Platinum 8480+
CPU(s) orderable: 1, 2 chips
Chips enabled: 2
Cores enabled: 112
Cores per chip: 56
Threads per core: 2
CPU Characteristics: Turbo Boost Technology up to 3.8 GHz
CPU MHz: 2000
Primary Cache: 32 KB I + 48 KB D on chip per core
Secondary Cache: 2 MB I+D on chip per core
L3 Cache: 105 MB I+D on chip per chip
Other Cache: None
Memory: 1 TB (16x64 GB DDR5 2Rx4 PC5-4800B-R)
Disk Subsystem: 1 x 1 1TB NVMe M.2 INTEL SSDPELKX010T8
Other Hardware: None
Accel Count: 4
Accel Model: Intel Data Center GPU Max 1550
Accel Vendor: Intel
Accel Type: GPU
Accel Connection: PCIe Gen5 x16
Accel ECC enabled: yes
Accel Description: Intel Data Center GPU Max 1550
Adapter: Mellanox ConnectX-6 HDR
Number of Adapters: 1
Slot Type: PCI-Express 4.0 x16
Data Rate: 200Gbit/s
Ports Used: 1
Interconnect Type: Mellanox HDR
Software
Accelerator Driver: 25.05.32567
Adapter: Mellanox ConnectX-6 HDR
Adapter Firmware: 20.38.1900
Operating System: SUSE Linux Enterprise Server 15 SP6
6.4.0-150600.23.42-default
Local File System: lustre
Shared File System: LUSTRE FS
System State: Run level 5
Other Software: None

Interconnect Description: Mellanox HDR

Hardware
Vendor: Mellanox
Model: Mellanox HDR
Switch Model: Mellanox Technologies MT28908 Family
InfiniBand Switch
Number of Switches: 12
Number of Ports: 40
Data Rate: 200 Gbit/s
Firmware: 20.38.1900
Topology: Fat-tree
Primary Use: MPI Traffic, LustreFS traffic
Software

Submit Notes

The config file option 'submit' was used.

General Notes

Environment variables set by runhpc before the start of the run:
LIBOMPTARGET_LEVEL_ZERO_USE_IMMEDIATE_COMMAND_LIST = "all"
I_MPI_FABRICS=shm:ofi
I_MPI_OFFLOAD=1
I_MPI_OFFLOAD_CELL=tile
I_MPI_OFFLOAD_TOPOLIB=level_zero
I_MPI_OFFLOAD_CELL_LIST=0,1,2,3,4,5,6,7
For the following tests src.alt was used in PEAK:
505 513 518 521 528 532 534 535

Platform Notes

 Device Vendor                                   Intel
 Device Version                                  OpenCL 3.0 NEO
 Driver Version                                  25.05.32567
 Base clock                                      900MHz
 Max clock frequency                             1600MHz
 Tiles                                           2
 Slices per Tile                                 1
 Max compute units per Tile                      512
 Sub-slices per slice                            64
 EUs per sub-slice                               8
 Threads per EU                                  8
 Max work item dimensions                        3
 Max work item sizes                             1024x1024x1024
 Max work group size                             1024
 Preferred work group size multiple              32
 Max sub-groups per work group                   64
 Sub-group sizes                                 16, 32
 L1 Cache per EU                                 65536
 L2 cache size                                   427819008
 Global memory size                              137438953472
 Address bits                                    64, Little-Endian

Compiler Version Notes

==============================================================================
 CXXC 532.sph_exa_t(base, peak)
------------------------------------------------------------------------------
Intel(R) oneAPI DPC++/C++ Compiler 2025.1.0 (2025.1.0.20250317)
Target: x86_64-unknown-linux-gnu
Thread model: posix
InstalledDir:
  /lfs/lfs17/mknyazev/intel/2025.1/oneapi/compiler/2025.1/bin/compiler
Configuration file:
  /lfs/lfs17/mknyazev/intel/2025.1/oneapi/compiler/2025.1/bin/compiler/../icpx.cfg
------------------------------------------------------------------------------

==============================================================================
 CC  505.lbm_t(base, peak) 513.soma_t(base, peak) 518.tealeaf_t(base, peak)
      521.miniswp_t(base, peak) 534.hpgmgfv_t(base, peak)
------------------------------------------------------------------------------
Intel(R) oneAPI DPC++/C++ Compiler 2025.1.0 (2025.1.0.20250317)
Target: x86_64-unknown-linux-gnu
Thread model: posix
InstalledDir:
  /lfs/lfs17/mknyazev/intel/2025.1/oneapi/compiler/2025.1/bin/compiler
Configuration file:
  /lfs/lfs17/mknyazev/intel/2025.1/oneapi/compiler/2025.1/bin/compiler/../icx.cfg
------------------------------------------------------------------------------

==============================================================================
 FC  519.clvleaf_t(base, peak) 535.weather_t(base, peak)
------------------------------------------------------------------------------
ifx (IFX) 2025.1.0 20250317
Copyright (C) 1985-2025 Intel Corporation. All rights reserved.
------------------------------------------------------------------------------

==============================================================================
 FC  528.pot3d_t(peak)
------------------------------------------------------------------------------
ifx: command line warning #10034: Unrecognized keyword '-optimistic-collapse'
  for option '-fopenmp'
ifx: command line warning #10157: ignoring option '-W'; argument is of wrong
  type
ifx (IFX) 2025.1.0 20250317
Copyright (C) 1985-2025 Intel Corporation. All rights reserved.
------------------------------------------------------------------------------

==============================================================================
 FC  528.pot3d_t(base)
------------------------------------------------------------------------------
ifx: command line warning #10157: ignoring option '-W'; argument is of wrong
  type
ifx (IFX) 2025.1.0 20250317
Copyright (C) 1985-2025 Intel Corporation. All rights reserved.
------------------------------------------------------------------------------

Base Compiler Invocation

C benchmarks:

 mpiicc -cc=icx 

C++ benchmarks:

 mpiicpc -cxx=icpx 

Fortran benchmarks:

 mpiifort -fc=ifx 

Base Portability Flags

505.lbm_t:  -DUSE_MPI 
513.soma_t:  -DUSE_MPI   -DSPEC_NO_VAR_ARRAY_REDUCE 
518.tealeaf_t:  -DUSE_MPI 
519.clvleaf_t:  -DUSE_MPI 
528.pot3d_t:  -DUSE_MPI 
535.weather_t:  -DUSE_MPI 

Base Optimization Flags

C benchmarks:

 -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:auto   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -DSPEC_COLLAPSE   -fopenmp-optimistic-collapse 

C++ benchmarks:

 -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:auto   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -DSPEC_COLLAPSE 

Fortran benchmarks:

 -DSPEC_COLLAPSE   -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:auto   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -DSPEC_ACCEL_AWARE_MPI   -fopenmp-target-loopopt 

Base Other Flags

Fortran benchmarks:

528.pot3d_t:  -Wno-incompatible-function-pointer-types 

Peak Compiler Invocation

C benchmarks:

 mpiicc -cc=icx 

C++ benchmarks:

 mpiicpc -cxx=icpx 

Fortran benchmarks:

 mpiifort -fc=ifx 

Peak Portability Flags

505.lbm_t:  -DUSE_MPI 
513.soma_t:  -DUSE_MPI   -DSPEC_NO_VAR_ARRAY_REDUCE 
518.tealeaf_t:  -DUSE_MPI 
519.clvleaf_t:  -DUSE_MPI 
528.pot3d_t:  -DUSE_MPI 
535.weather_t:  -DUSE_MPI 

Peak Optimization Flags

C benchmarks:

505.lbm_t:  -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:large   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -DSPEC_COLLAPSE   -fopenmp-optimistic-collapse 
513.soma_t:  -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:auto   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -DSPEC_COLLAPSE   -DSPEC_ACCEL_AWARE_MPI   -fopenmp-optimistic-collapse 
518.tealeaf_t:  -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:auto   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -DSPEC_COLLAPSE   -fopenmp-optimistic-collapse 
521.miniswp_t:  -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:small   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -DSPEC_COLLAPSE   -fopenmp-optimistic-collapse   -fopenmp-targets=spir64_gen="-fp-model=fast"   -fopenmp-target-loopopt 
534.hpgmgfv_t:  Same as 513.soma_t 

C++ benchmarks:

 -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:auto   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -DSPEC_COLLAPSE 

Fortran benchmarks:

519.clvleaf_t:  basepeak = yes 
528.pot3d_t:  -DSPEC_COLLAPSE   -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:large   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -fopenmp-optimistic-collapse   -DSPEC_ACCEL_AWARE_MPI   -fopenmp-target-loopopt 
535.weather_t:  -DSPEC_COLLAPSE   -O3   -xCORE-AVX512   -flto   -mprefer-vector-width=512   -ffast-math   -fiopenmp   -fopenmp-targets=spir64_gen   -ftarget-register-alloc-mode=pvc:auto   -Xopenmp-target-backend '-device pvc -revision_id 0x2f'   -DSPEC_ACCEL_AWARE_MPI   -fopenmp-target-loopopt 

Peak Other Flags

Fortran benchmarks:

528.pot3d_t:  -Wno-incompatible-function-pointer-types 

The flags file that was used to format this result can be browsed at
http://www.spec.org/hpc2021/flags/Intel_compiler_flags.2025-05-22.00.html.

You can also download the XML flags source by saving the following link:
http://www.spec.org/hpc2021/flags/Intel_compiler_flags.2025-05-22.00.xml.