gh200

ARMv8 Neoverse-V2 testing with a Pegatron JIMBO P4352 (00022432 BIOS) and NVIDIA GH200 144G HBM3e 143GB on Ubuntu 24.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2410122-NE-GH200545357
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
a
October 12
  12 Hours, 26 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


gh200OpenBenchmarking.orgPhoronix Test SuiteARMv8 Neoverse-V2 @ 3.47GHz (72 Cores)Pegatron JIMBO P4352 (00022432 BIOS)1 x 480GB LPDDR5-6400MT/s NVIDIA 699-2G530-0236-RC11000GB CT1000T700SSD3NVIDIA GH200 144G HBM3e 143GB2 x Intel X550Ubuntu 24.046.8.0-45-generic-64k (aarch64)NVIDIAOpenCL 3.0 CUDA 12.6.65GCC 13.2.0ext41920x1200ProcessorMotherboardMemoryDiskGraphicsNetworkOSKernelDisplay DriverOpenCLCompilerFile-SystemScreen ResolutionGh200 BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=aarch64-linux-gnu --disable-libquadmath --disable-libquadmath-support --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-fix-cortex-a53-843419 --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-backtrace --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-dIwDw0/gcc-13-13.2.0/debian/tmp-nvptx/usr --enable-plugin --enable-shared --enable-threads=posix --host=aarch64-linux-gnu --program-prefix=aarch64-linux-gnu- --target=aarch64-linux-gnu --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-target-system-zlib=auto --without-cuda-driver -v - Scaling Governor: cppc_cpufreq ondemand (Boost: Disabled)- OpenJDK Runtime Environment (build 21.0.4+7-Ubuntu-1ubuntu224.04)- Python 3.12.3- gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + reg_file_data_sampling: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of __user pointer sanitization + spectre_v2: Not affected + srbds: Not affected + tsx_async_abort: Not affected

gh200onnx: ResNet101_DUC_HDC-12 - CPU - Standardonnx: ResNet101_DUC_HDC-12 - CPU - Parallelonnx: super-resolution-10 - CPU - Standardonnx: super-resolution-10 - CPU - Parallelonnx: ResNet50 v1-12-int8 - CPU - Standardonnx: ResNet50 v1-12-int8 - CPU - Parallelonnx: fcn-resnet101-11 - CPU - Standardonnx: fcn-resnet101-11 - CPU - Parallelonnx: CaffeNet 12-int8 - CPU - Standardonnx: CaffeNet 12-int8 - CPU - Parallelonnx: T5 Encoder - CPU - Standardonnx: T5 Encoder - CPU - Parallelonnx: ZFNet-512 - CPU - Standardonnx: ZFNet-512 - CPU - Parallelonnx: yolov4 - CPU - Standardonnx: yolov4 - CPU - Parallelpyperformance: pickle_pure_pythonpyperformance: asyncio_websocketspyperformance: django_templatepyperformance: asyncio_tcp_sslpyperformance: python_startuppyperformance: regex_compilepyperformance: async_tree_iopyperformance: crypto_pyaespyperformance: json_loadspyperformance: gc_collectpyperformance: xml_etreepyperformance: raytracepyperformance: pathlibpyperformance: nbodypyperformance: floatpyperformance: chaospyperformance: goblender: Pabellon Barcelona - CPU-Onlyblender: Barbershop - CPU-Onlyblender: Fishy Cat - CPU-Onlyblender: Classroom - CPU-Onlyblender: BMW27 - CPU-Onlyxnnpack: QU8MobileNetV3Smallxnnpack: QU8MobileNetV3Largexnnpack: QU8MobileNetV2xnnpack: FP16MobileNetV3Smallxnnpack: FP16MobileNetV3Largexnnpack: FP16MobileNetV2xnnpack: FP32MobileNetV3Smallxnnpack: FP32MobileNetV3Largexnnpack: FP32MobileNetV2mnn: inception-v3mnn: mobilenet-v1-1.0mnn: MobileNetV2_224mnn: SqueezeNetV1.0mnn: resnet-v2-50mnn: squeezenetv1.1mnn: mobilenetV3mnn: nasnetgromacs: water_GMX50_baregromacs: MPI CPU - water_GMX50_barepovray: Trace Timec-ray: 1080p - 16c-ray: 5K - 16c-ray: 4K - 16build2: Time To Compilebuild-llvm: Unix Makefilesbuild-llvm: Ninjabuild-linux-kernel: allmodconfigbuild-linux-kernel: defconfigcompress-7zip: Decompression Ratingcompress-7zip: Compression Ratingcompress-7zip: Decompression Ratingcompress-7zip: Compression Ratingx265: Bosphorus 1080px265: Bosphorus 4Kgraphics-magick: HWB Color Spacegraphics-magick: Noise-Gaussiangraphics-magick: Resizinggraphics-magick: Enhancedgraphics-magick: Sharpengraphics-magick: Rotategraphics-magick: Swirlgraphics-magick: HWB Color Spacegraphics-magick: Noise-Gaussiangraphics-magick: Resizinggraphics-magick: Enhancedgraphics-magick: Sharpengraphics-magick: Rotategraphics-magick: Swirlbyte: Whetstone Doublebyte: System Callbyte: Dhrystone 2byte: Pipesimdjson: DistinctUserIDsimdjson: PartialTweetssimdjson: LargeRandsimdjson: TopTweetsimdjson: Kostyawarpx: Plasma Accelerationwarpx: Uniform Plasmaepoch: Conelczero: Eigenetcpak: Multi-Threaded - ETC2onnx: ResNet101_DUC_HDC-12 - CPU - Standardonnx: ResNet101_DUC_HDC-12 - CPU - Parallelonnx: super-resolution-10 - CPU - Standardonnx: super-resolution-10 - CPU - Parallelonnx: ResNet50 v1-12-int8 - CPU - Standardonnx: ResNet50 v1-12-int8 - CPU - Parallelonnx: fcn-resnet101-11 - CPU - Standardonnx: fcn-resnet101-11 - CPU - Parallelonnx: CaffeNet 12-int8 - CPU - Standardonnx: CaffeNet 12-int8 - CPU - Parallelonnx: T5 Encoder - CPU - Standardonnx: T5 Encoder - CPU - Parallelonnx: ZFNet-512 - CPU - Standardonnx: ZFNet-512 - CPU - Parallelonnx: yolov4 - CPU - Standardonnx: yolov4 - CPU - Parallelstockfish: Chess Benchmarkstockfish: Chess Benchmarklczero: BLASa0.3032220.360196159.06818.6487317.868161.3320.4629430.5829021262.14321.237390.067109.241210.11745.23145.146855.8071120551026.31.4918.782.374854.817.51.0845.821715.564.556.847.498.2154.46381.4573.0278.3738.06108314849458811226840945142696713.6931.7931.5023.39611.3351.8241.1345.0087.1566.0017.7865.19536.20820.35784.786276.926175.030285.13266.71041881939352342052438477512.618.81430217282351171209605656301442359411331657721978.0145868649.34998587529.8202565282.24.164.061.154.143.1120.3805382116.90202076188.20360471.1933298.192776.336.2855353.62523.145236.197262160.111715.640.7916243.111582.560039.151934.7586522.1451194.364172.22558496753168428763OpenBenchmarking.org

ONNX Runtime

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Standarda0.06820.13640.20460.27280.341SE +/- 0.001997, N = 30.3032221. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Parallela0.0810.1620.2430.3240.405SE +/- 0.001260, N = 30.3601961. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: super-resolution-10 - Device: CPU - Executor: Standarda4080120160200SE +/- 1.12, N = 3159.071. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: super-resolution-10 - Device: CPU - Executor: Parallela510152025SE +/- 0.10, N = 318.651. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standarda70140210280350SE +/- 0.61, N = 3317.871. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallela4080120160200SE +/- 0.70, N = 3161.331. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: fcn-resnet101-11 - Device: CPU - Executor: Standarda0.10420.20840.31260.41680.521SE +/- 0.001019, N = 30.4629431. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: fcn-resnet101-11 - Device: CPU - Executor: Parallela0.13120.26240.39360.52480.656SE +/- 0.002945, N = 30.5829021. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: CaffeNet 12-int8 - Device: CPU - Executor: Standarda30060090012001500SE +/- 2.68, N = 31262.141. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallela70140210280350SE +/- 0.27, N = 3321.241. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: T5 Encoder - Device: CPU - Executor: Standarda80160240320400SE +/- 2.61, N = 3390.071. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: T5 Encoder - Device: CPU - Executor: Parallela20406080100SE +/- 0.76, N = 3109.241. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ZFNet-512 - Device: CPU - Executor: Standarda50100150200250SE +/- 2.64, N = 3210.121. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: ZFNet-512 - Device: CPU - Executor: Parallela1020304050SE +/- 0.50, N = 1545.231. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: yolov4 - Device: CPU - Executor: Standarda1.1582.3163.4744.6325.79SE +/- 0.07121, N = 35.146851. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

OpenBenchmarking.orgInferences Per Second, More Is BetterONNX Runtime 1.19Model: yolov4 - Device: CPU - Executor: Parallela1.30662.61323.91985.22646.533SE +/- 0.04903, N = 35.807111. (CXX) g++ options: -O3 -march=native -ffunction-sections -fdata-sections -mtune=native -flto=auto -fno-fat-lto-objects -ldl -lrt

PyPerformance

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: pickle_pure_pythona4080120160200SE +/- 0.33, N = 3205

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: asyncio_websocketsa110220330440550SE +/- 0.33, N = 3510

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: django_templatea612182430SE +/- 0.12, N = 326.3

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: asyncio_tcp_ssla0.33530.67061.00591.34121.6765SE +/- 0.00, N = 31.49

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: python_startupa510152025SE +/- 0.06, N = 318.7

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: regex_compilea20406080100SE +/- 0.12, N = 382.3

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: async_tree_ioa160320480640800SE +/- 2.96, N = 3748

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: crypto_pyaesa1224364860SE +/- 0.03, N = 354.8

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: json_loadsa48121620SE +/- 0.06, N = 317.5

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: gc_collecta0.2430.4860.7290.9721.215SE +/- 0.01, N = 151.08

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: xml_etreea1020304050SE +/- 0.03, N = 345.8

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: raytracea50100150200250SE +/- 0.33, N = 3217

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: pathliba48121620SE +/- 0.03, N = 315.5

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: nbodya1428425670SE +/- 0.09, N = 364.5

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: floata1326395265SE +/- 0.03, N = 356.8

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: chaosa1122334455SE +/- 0.06, N = 347.4

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.11Benchmark: goa20406080100SE +/- 0.07, N = 398.2

Blender

Blender is an open-source 3D creation software project. This test is of Blender's Cycles benchmark with various sample files. GPU computing is supported. This system/blender test profile makes use of the system-supplied Blender. Use pts/blender if wishing to stick to a fixed version of Blender. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: Pabellon Barcelona - Compute: CPU-Onlya306090120150SE +/- 0.37, N = 3154.46

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: Barbershop - Compute: CPU-Onlya80160240320400SE +/- 0.54, N = 3381.45

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: Fishy Cat - Compute: CPU-Onlya1632486480SE +/- 0.44, N = 373.02

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: Classroom - Compute: CPU-Onlya20406080100SE +/- 0.08, N = 378.37

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 4.0.2Blend File: BMW27 - Compute: CPU-Onlya918273645SE +/- 0.04, N = 338.06

XNNPACK

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: QU8MobileNetV3Smalla2004006008001000SE +/- 9.82, N = 310831. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: QU8MobileNetV3Largea30060090012001500SE +/- 8.97, N = 314841. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: QU8MobileNetV2a2004006008001000SE +/- 6.69, N = 39451. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP16MobileNetV3Smalla2004006008001000SE +/- 20.00, N = 38811. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP16MobileNetV3Largea30060090012001500SE +/- 21.31, N = 312261. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP16MobileNetV2a2004006008001000SE +/- 15.62, N = 38401. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP32MobileNetV3Smalla2004006008001000SE +/- 16.38, N = 39451. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP32MobileNetV3Largea30060090012001500SE +/- 6.51, N = 314261. (CXX) g++ options: -O3 -lrt -lm

OpenBenchmarking.orgus, Fewer Is BetterXNNPACK 2cd86bModel: FP32MobileNetV2a2004006008001000SE +/- 8.41, N = 39671. (CXX) g++ options: -O3 -lrt -lm

Mobile Neural Network

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: inception-v3a48121620SE +/- 0.02, N = 313.69MIN: 11.51 / MAX: 42.341. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: mobilenet-v1-1.0a0.40340.80681.21021.61362.017SE +/- 0.005, N = 31.793MIN: 1.34 / MAX: 22.051. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: MobileNetV2_224a0.3380.6761.0141.3521.69SE +/- 0.019, N = 31.502MIN: 1.12 / MAX: 13.521. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: SqueezeNetV1.0a0.76411.52822.29233.05643.8205SE +/- 0.027, N = 33.396MIN: 2.14 / MAX: 29.881. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: resnet-v2-50a3691215SE +/- 0.10, N = 311.34MIN: 8.54 / MAX: 42.161. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: squeezenetv1.1a0.41040.82081.23121.64162.052SE +/- 0.044, N = 31.824MIN: 1.17 / MAX: 20.371. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: mobilenetV3a0.25520.51040.76561.02081.276SE +/- 0.009, N = 31.134MIN: 0.69 / MAX: 11.141. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.9.b11b7037dModel: nasneta1.12682.25363.38044.50725.634SE +/- 0.036, N = 35.008MIN: 4.49 / MAX: 27.911. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -pthread -ldl

GROMACS

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACSInput: water_GMX50_barea246810SE +/- 0.004, N = 37.1561. GROMACS version: 2023.3-Ubuntu_2023.3_1ubuntu3

OpenBenchmarking.orgNs Per Day, More Is BetterGROMACS 2024Implementation: MPI CPU - Input: water_GMX50_barea246810SE +/- 0.003, N = 36.0011. (CXX) g++ options: -O3 -lm

POV-Ray

OpenBenchmarking.orgSeconds, Fewer Is BetterPOV-RayTrace Timea246810SE +/- 0.061, N = 37.7861. POV-Ray 3.7.0.10.unofficial

C-Ray

OpenBenchmarking.orgSeconds, Fewer Is BetterC-Ray 2.0Resolution: 1080p - Rays Per Pixel: 16a1.16892.33783.50674.67565.8445SE +/- 0.003, N = 35.1951. (CC) gcc options: -lpthread -lm

OpenBenchmarking.orgSeconds, Fewer Is BetterC-Ray 2.0Resolution: 5K - Rays Per Pixel: 16a816243240SE +/- 0.02, N = 336.211. (CC) gcc options: -lpthread -lm

OpenBenchmarking.orgSeconds, Fewer Is BetterC-Ray 2.0Resolution: 4K - Rays Per Pixel: 16a510152025SE +/- 0.00, N = 320.361. (CC) gcc options: -lpthread -lm

Build2

OpenBenchmarking.orgSeconds, Fewer Is BetterBuild2 0.17Time To Compilea20406080100SE +/- 0.22, N = 384.79

Timed LLVM Compilation

This test times how long it takes to compile/build the LLVM compiler stack. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed LLVM Compilation 16.0Build System: Unix Makefilesa60120180240300SE +/- 0.32, N = 3276.93

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed LLVM Compilation 16.0Build System: Ninjaa4080120160200SE +/- 1.19, N = 3175.03

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.8Build: allmodconfiga60120180240300SE +/- 2.59, N = 3285.13

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.8Build: defconfiga1530456075SE +/- 0.55, N = 1366.71

7-Zip Compression

OpenBenchmarking.orgMIPS, More Is Better7-Zip CompressionTest: Decompression Ratinga90K180K270K360K450KSE +/- 507.84, N = 34188191. 7-Zip 23.01 (arm64) : Copyright (c) 1999-2023 Igor Pavlov : 2023-06-20

OpenBenchmarking.orgMIPS, More Is Better7-Zip CompressionTest: Compression Ratinga80K160K240K320K400KSE +/- 3097.20, N = 33935231. 7-Zip 23.01 (arm64) : Copyright (c) 1999-2023 Igor Pavlov : 2023-06-20

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 24.05Test: Decompression Ratinga90K180K270K360K450KSE +/- 944.71, N = 34205241. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 24.05Test: Compression Ratinga80K160K240K320K400KSE +/- 4213.31, N = 33847751. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

x265

OpenBenchmarking.orgFrames Per Second, More Is Betterx265Video Input: Bosphorus 1080pa3691215SE +/- 0.18, N = 312.611. x265 [info]: HEVC encoder version 3.5+1-f0c1022b6

OpenBenchmarking.orgFrames Per Second, More Is Betterx265Video Input: Bosphorus 4Ka246810SE +/- 0.03, N = 38.811. x265 [info]: HEVC encoder version 3.5+1-f0c1022b6

GraphicsMagick

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: HWB Color Spacea90180270360450SE +/- 0.67, N = 34301. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Noise-Gaussiana50100150200250SE +/- 1.73, N = 32171. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Resizinga60120180240300SE +/- 1.15, N = 32821. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Enhanceda80160240320400SE +/- 0.33, N = 33511. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Sharpena4080120160200SE +/- 0.33, N = 31711. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Rotatea50100150200250SE +/- 0.88, N = 32091. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagickOperation: Swirla130260390520650SE +/- 5.51, N = 36051. GraphicsMagick 1.3.42 2023-09-23 Q16 http://www.GraphicsMagick.org/

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: HWB Color Spacea140280420560700SE +/- 8.82, N = 36561. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Noise-Gaussiana70140210280350SE +/- 2.18, N = 153011. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Resizinga100200300400500SE +/- 6.06, N = 34421. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Enhanceda80160240320400SE +/- 0.67, N = 33591. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Sharpena90180270360450SE +/- 0.58, N = 34111. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Rotatea70140210280350SE +/- 4.26, N = 33311. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

OpenBenchmarking.orgIterations Per Minute, More Is BetterGraphicsMagick 1.3.43Operation: Swirla140280420560700SE +/- 4.26, N = 36571. (CC) gcc options: -fopenmp -O2 -ljpeg -lSM -lICE -lX11 -lz -lm -lpthread -lgomp

BYTE Unix Benchmark

OpenBenchmarking.orgMWIPS, More Is BetterBYTE Unix Benchmark 5.1.3-gitComputational Test: Whetstone Doublea150K300K450K600K750KSE +/- 19.25, N = 3721978.01. (CC) gcc options: -pedantic -O3 -ffast-math -march=native -mtune=native -lm

OpenBenchmarking.orgLPS, More Is BetterBYTE Unix Benchmark 5.1.3-gitComputational Test: System Calla30M60M90M120M150MSE +/- 15202.21, N = 3145868649.31. (CC) gcc options: -pedantic -O3 -ffast-math -march=native -mtune=native -lm

OpenBenchmarking.orgLPS, More Is BetterBYTE Unix Benchmark 5.1.3-gitComputational Test: Dhrystone 2a1100M2200M3300M4400M5500MSE +/- 2591819.88, N = 34998587529.81. (CC) gcc options: -pedantic -O3 -ffast-math -march=native -mtune=native -lm

OpenBenchmarking.orgLPS, More Is BetterBYTE Unix Benchmark 5.1.3-gitComputational Test: Pipea40M80M120M160M200MSE +/- 32087.94, N = 3202565282.21. (CC) gcc options: -pedantic -O3 -ffast-math -march=native -mtune=native -lm

simdjson

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: DistinctUserIDa0.9361.8722.8083.7444.68SE +/- 0.00, N = 34.161. (CXX) g++ options: -O3 -lrt

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: PartialTweetsa0.91351.8272.74053.6544.5675SE +/- 0.00, N = 34.061. (CXX) g++ options: -O3 -lrt

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: LargeRandoma0.25880.51760.77641.03521.294SE +/- 0.00, N = 31.151. (CXX) g++ options: -O3 -lrt

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: TopTweeta0.93151.8632.79453.7264.6575SE +/- 0.01, N = 34.141. (CXX) g++ options: -O3 -lrt

OpenBenchmarking.orgGB/s, More Is Bettersimdjson 3.10Throughput Test: Kostyaa0.69981.39962.09942.79923.499SE +/- 0.01, N = 33.111. (CXX) g++ options: -O3 -lrt

WarpX

OpenBenchmarking.orgSeconds, Fewer Is BetterWarpX 24.10Input: Plasma Accelerationa510152025SE +/- 0.03, N = 320.381. (CXX) g++ options: -O3

OpenBenchmarking.orgSeconds, Fewer Is BetterWarpX 24.10Input: Uniform Plasmaa48121620SE +/- 0.18, N = 316.901. (CXX) g++ options: -O3

Epoch

OpenBenchmarking.orgSeconds, Fewer Is BetterEpoch 4.19.4Epoch3D Deck: Conea4080120160200SE +/- 2.18, N = 4188.201. (F9X) gfortran options: -O3 -std=f2003 -Jobj -lsdf -lmpi_usempif08 -lmpi_mpifh -lmpi -lopen-rte -lopen-pal -lhwloc -levent_core -levent_pthreads -lm -lz

LeelaChessZero

OpenBenchmarking.orgNodes Per Second, More Is BetterLeelaChessZero 0.31.1Backend: Eigena80160240320400SE +/- 4.26, N = 33601. (CXX) g++ options: -flto -pthread

Etcpak

OpenBenchmarking.orgMpx/s, More Is BetterEtcpak 2.0Benchmark: Multi-Threaded - Configuration: ETC2a100200300400500SE +/- 2.47, N = 3471.191. (CXX) g++ options: -flto -pthread

ONNX Runtime

Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "FasterRCNN-12-int8/FasterRCNN-12-int8.onnx" failed: No such file or directory

Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "FasterRCNN-12-int8/FasterRCNN-12-int8.onnx" failed: No such file or directory

Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "resnet100/resnet100.onnx" failed: No such file or directory

Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "resnet100/resnet100.onnx" failed: No such file or directory

Model: bertsquad-12 - Device: CPU - Executor: Standard

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "bertsquad-12/bertsquad-12.onnx" failed: No such file or directory

Model: bertsquad-12 - Device: CPU - Executor: Parallel

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "bertsquad-12/bertsquad-12.onnx" failed: No such file or directory

Model: GPT-2 - Device: CPU - Executor: Standard

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "GPT2/model.onnx" failed: No such file or directory

Model: GPT-2 - Device: CPU - Executor: Parallel

a: The test quit with a non-zero exit status. E: onnxruntime/onnxruntime/test/onnx/onnx_model_info.cc:45 void OnnxModelInfo::InitOnnxModelInfo(const std::filesystem::__cxx11::path&) open file "GPT2/model.onnx" failed: No such file or directory

Apache Cassandra

Test: Writes

a: The test run did not produce a result.

PostgreSQL

Scaling Factor: 1000 - Clients: 1000 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 800 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 500 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 1000 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 1000 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 800 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1000 - Clients: 500 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 800 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 500 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 1000 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 800 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 100 - Clients: 500 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 1000 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 800 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 500 - Mode: Read Write

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 1000 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 800 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

Scaling Factor: 1 - Clients: 500 - Mode: Read Only

a: The test run did not produce a result. E: ./pgbench: 21: pg_/bin/pgbench: not found

GROMACS

The GROMACS (GROningen MAchine for Chemical Simulations) molecular dynamics package testing with the water_GMX50 data. This test profile allows selecting between CPU and GPU-based GROMACS builds. Learn more via the OpenBenchmarking.org test page.

Implementation: NVIDIA CUDA GPU - Input: water_GMX50_bare

a: The test quit with a non-zero exit status. E: ./gromacs: 5: /cuda-build/run-gromacs: not found

Stockfish

OpenBenchmarking.orgNodes Per Second, More Is BetterStockfishChess Benchmarka13M26M39M52M65MSE +/- 959000.15, N = 15584967531. Stockfish 16 by the Stockfish developers (see AUTHORS file)

OpenBenchmarking.orgNodes Per Second, More Is BetterStockfish 17Chess Benchmarka40M80M120M160M200MSE +/- 6156005.01, N = 151684287631. (CXX) g++ options: -lgcov -lpthread -fno-exceptions -std=c++17 -fno-peel-loops -fno-tracer -pedantic -O3 -funroll-loops -flto -flto-partition=one -flto=jobserver

LeelaChessZero

Backend: BLAS

a: The test quit with a non-zero exit status.

102 Results Shown

ONNX Runtime:
  ResNet101_DUC_HDC-12 - CPU - Standard
  ResNet101_DUC_HDC-12 - CPU - Parallel
  super-resolution-10 - CPU - Standard
  super-resolution-10 - CPU - Parallel
  ResNet50 v1-12-int8 - CPU - Standard
  ResNet50 v1-12-int8 - CPU - Parallel
  fcn-resnet101-11 - CPU - Standard
  fcn-resnet101-11 - CPU - Parallel
  CaffeNet 12-int8 - CPU - Standard
  CaffeNet 12-int8 - CPU - Parallel
  T5 Encoder - CPU - Standard
  T5 Encoder - CPU - Parallel
  ZFNet-512 - CPU - Standard
  ZFNet-512 - CPU - Parallel
  yolov4 - CPU - Standard
  yolov4 - CPU - Parallel
PyPerformance:
  pickle_pure_python
  asyncio_websockets
  django_template
  asyncio_tcp_ssl
  python_startup
  regex_compile
  async_tree_io
  crypto_pyaes
  json_loads
  gc_collect
  xml_etree
  raytrace
  pathlib
  nbody
  float
  chaos
  go
Blender:
  Pabellon Barcelona - CPU-Only
  Barbershop - CPU-Only
  Fishy Cat - CPU-Only
  Classroom - CPU-Only
  BMW27 - CPU-Only
XNNPACK:
  QU8MobileNetV3Small
  QU8MobileNetV3Large
  QU8MobileNetV2
  FP16MobileNetV3Small
  FP16MobileNetV3Large
  FP16MobileNetV2
  FP32MobileNetV3Small
  FP32MobileNetV3Large
  FP32MobileNetV2
Mobile Neural Network:
  inception-v3
  mobilenet-v1-1.0
  MobileNetV2_224
  SqueezeNetV1.0
  resnet-v2-50
  squeezenetv1.1
  mobilenetV3
  nasnet
GROMACS
GROMACS
POV-Ray
C-Ray:
  1080p - 16
  5K - 16
  4K - 16
Build2
Timed LLVM Compilation:
  Unix Makefiles
  Ninja
Timed Linux Kernel Compilation:
  allmodconfig
  defconfig
7-Zip Compression:
  Decompression Rating
  Compression Rating
7-Zip Compression:
  Decompression Rating
  Compression Rating
x265:
  Bosphorus 1080p
  Bosphorus 4K
GraphicsMagick:
  HWB Color Space
  Noise-Gaussian
  Resizing
  Enhanced
  Sharpen
  Rotate
  Swirl
GraphicsMagick:
  HWB Color Space
  Noise-Gaussian
  Resizing
  Enhanced
  Sharpen
  Rotate
  Swirl
BYTE Unix Benchmark:
  Whetstone Double
  System Call
  Dhrystone 2
  Pipe
simdjson:
  DistinctUserID
  PartialTweets
  LargeRand
  TopTweet
  Kostya
WarpX:
  Plasma Acceleration
  Uniform Plasma
Epoch
LeelaChessZero
Etcpak
Stockfish
Stockfish