Intel Core i9-10980XE GCC compiler benchmarking by Michael Larabel for a future article.
GCC 8.5 Processor: Intel Core i9-10980XE @ 4.80GHz (18 Cores / 36 Threads), Motherboard: ASRock X299 Steel Legend (P1.30 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 32GB, Disk: Samsung SSD 970 PRO 512GB, Graphics: NVIDIA NV132 11GB, Audio: Realtek ALC1220, Monitor: ASUS VP28U, Network: Intel I219-V + Intel I211
OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 8.5.0, File-System: ext4, Screen Resolution: 2560x1600
Kernel Notes: Transparent Huge Pages: madviseEnvironment Notes: CXXFLAGS="-O3 -march=native" CFLAGS="-O3 -march=native"Compiler Notes: --disable-multilib --enable-checking=release --enable-languages=c,c++Processor Notes: Scaling Governor: intel_cpufreq schedutil - CPU Microcode: 0x5003102Python Notes: Python 3.9.5Security Notes: itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled
GCC 9.4 OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 9.4.0, File-System: ext4, Screen Resolution: 2560x1600
GCC 10.3 OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 10.3.0, File-System: ext4, Screen Resolution: 2560x1600
GCC 11.1 OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 11.1.0, File-System: ext4, Screen Resolution: 2560x1600
GCC 12.0.0 20210701 OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 12.0.0 20210701, File-System: ext4, Screen Resolution: 2560x1600
GraphicsMagick This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Sharpen GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 70 140 210 280 350 SE +/- 0.33, N = 3 265 197 319 319 317 1. (CC) gcc options: -fopenmp -O3 -march=native -pthread -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread
FinanceBench FinanceBench is a collection of financial program benchmarks with support for benchmarking on the GPU via OpenCL and CPU benchmarking with OpenMP. The FinanceBench test cases are focused on Black-Sholes-Merton Process with Analytic European Option engine, QMC (Sobol) Monte-Carlo method (Equity Option Example), Bonds Fixed-rate bond with flat forward curve, and Repo Securities repurchase agreement. FinanceBench was originally written by the Cavazos Lab at University of Delaware. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better FinanceBench 2016-07-25 Benchmark: Bonds OpenMP GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 16K 32K 48K 64K 80K SE +/- 1061.39, N = 3 SE +/- 5.53, N = 3 SE +/- 48.36, N = 3 SE +/- 36.89, N = 3 SE +/- 11.94, N = 3 76452.78 74820.19 48317.58 48802.76 49799.12 1. (CXX) g++ options: -O3 -march=native -fopenmp
Zstd Compression This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better Zstd Compression 1.5.0 Compression Level: 8, Long Mode - Compression Speed GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 100 200 300 400 500 SE +/- 2.56, N = 13 SE +/- 3.63, N = 3 SE +/- 3.88, N = 15 SE +/- 3.45, N = 3 SE +/- 3.28, N = 15 375.6 335.5 385.5 469.4 370.1 1. (CC) gcc options: -O3 -march=native -pthread -lz
eSpeak-NG Speech Engine This test times how long it takes the eSpeak speech synthesizer to read Project Gutenberg's The Outline of Science and output to a WAV file. This test profile is now tracking the eSpeak-NG version of eSpeak. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better eSpeak-NG Speech Engine 20200907 Text-To-Speech Synthesis GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 8 16 24 32 40 SE +/- 0.23, N = 4 SE +/- 0.19, N = 4 SE +/- 0.18, N = 4 SE +/- 0.16, N = 4 SE +/- 0.09, N = 4 28.11 26.84 27.28 35.09 32.67 1. (CC) gcc options: -O3 -march=native -std=c99
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: ChaCha20Poly1305 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 200 400 600 800 1000 SE +/- 1.36, N = 3 SE +/- 2.79, N = 3 SE +/- 0.73, N = 3 SE +/- 1.05, N = 3 SE +/- 0.15, N = 3 951.15 984.02 781.38 779.38 788.57 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: ChaCha20Poly1305 - Decrypt GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 200 400 600 800 1000 SE +/- 0.08, N = 3 SE +/- 2.38, N = 3 SE +/- 0.47, N = 3 SE +/- 0.89, N = 3 SE +/- 0.79, N = 3 945.45 977.17 775.49 774.65 780.12 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
FinanceBench FinanceBench is a collection of financial program benchmarks with support for benchmarking on the GPU via OpenCL and CPU benchmarking with OpenMP. The FinanceBench test cases are focused on Black-Sholes-Merton Process with Analytic European Option engine, QMC (Sobol) Monte-Carlo method (Equity Option Example), Bonds Fixed-rate bond with flat forward curve, and Repo Securities repurchase agreement. FinanceBench was originally written by the Cavazos Lab at University of Delaware. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better FinanceBench 2016-07-25 Benchmark: Repo OpenMP GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 9K 18K 27K 36K 45K SE +/- 10.60, N = 3 SE +/- 34.41, N = 3 SE +/- 22.03, N = 3 SE +/- 43.71, N = 3 SE +/- 120.35, N = 3 42795.05 42354.90 34223.30 34558.22 35458.66 1. (CXX) g++ options: -O3 -march=native -fopenmp
GraphicsMagick This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Swirl GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 200 400 600 800 1000 SE +/- 0.33, N = 3 SE +/- 1.20, N = 3 SE +/- 0.33, N = 3 752 792 903 924 761 1. (CC) gcc options: -fopenmp -O3 -march=native -pthread -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.1865 0.373 0.5595 0.746 0.9325 SE +/- 0.008442, N = 15 SE +/- 0.003298, N = 3 SE +/- 0.008934, N = 3 SE +/- 0.005568, N = 3 SE +/- 0.005954, N = 8 0.828866 0.679482 0.678626 0.697080 0.680522 MIN: 0.74 MIN: 0.66 MIN: 0.65 MIN: 0.67 MIN: 0.63 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
Smallpt Smallpt is a C++ global illumination renderer written in less than 100 lines of code. Global illumination is done via unbiased Monte Carlo path tracing and there is multi-threading support via the OpenMP library. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Smallpt 1.0 Global Illumination Renderer; 128 Samples GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 2 4 6 8 10 SE +/- 0.011, N = 3 SE +/- 0.007, N = 3 SE +/- 0.017, N = 3 SE +/- 0.004, N = 3 SE +/- 0.033, N = 3 6.045 5.286 5.991 6.201 6.130 1. (CXX) g++ options: -fopenmp -O3 -march=native
Mobile Neural Network MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.2 Model: resnet-v2-50 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 7 14 21 28 35 SE +/- 0.47, N = 15 SE +/- 0.10, N = 3 SE +/- 0.19, N = 3 SE +/- 0.06, N = 3 SE +/- 0.24, N = 3 31.45 27.30 28.27 28.45 28.26 MIN: 24.41 / MAX: 36.24 MIN: 26.86 / MAX: 27.93 MIN: 27.7 / MAX: 28.91 MIN: 27.77 / MAX: 28.83 MIN: 27.6 / MAX: 28.76 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
ViennaCL ViennaCL is an open-source linear algebra library written in C++ and with support for OpenCL and OpenMP. This test profile makes use of ViennaCL's built-in benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GFLOPs/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - dGEMM-NN GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 13 26 39 52 65 SE +/- 0.13, N = 3 SE +/- 0.75, N = 3 SE +/- 0.44, N = 3 SE +/- 0.15, N = 3 SE +/- 0.63, N = 3 56.1 55.5 51.2 51.0 58.7 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: Twofish GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 90 180 270 360 450 SE +/- 0.12, N = 3 SE +/- 0.33, N = 3 SE +/- 0.30, N = 3 SE +/- 1.13, N = 3 SE +/- 0.23, N = 3 414.45 416.18 366.65 367.68 404.16 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
ViennaCL ViennaCL is an open-source linear algebra library written in C++ and with support for OpenCL and OpenMP. This test profile makes use of ViennaCL's built-in benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GFLOPs/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - dGEMM-TN GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 13 26 39 52 65 SE +/- 0.36, N = 3 SE +/- 1.55, N = 2 SE +/- 0.10, N = 3 SE +/- 0.10, N = 3 SE +/- 1.17, N = 3 56.2 56.0 51.9 51.9 58.9 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
OpenBenchmarking.org GFLOPs/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - dGEMM-NT GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 13 26 39 52 65 SE +/- 0.27, N = 3 SE +/- 0.25, N = 3 SE +/- 0.37, N = 3 SE +/- 0.15, N = 3 SE +/- 0.48, N = 3 54.1 54.7 49.8 49.8 56.4 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
OpenBenchmarking.org GFLOPs/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - dGEMM-TT GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 13 26 39 52 65 SE +/- 0.07, N = 3 SE +/- 0.32, N = 3 SE +/- 0.09, N = 3 SE +/- 0.20, N = 3 SE +/- 0.70, N = 2 54.9 55.2 51.5 50.5 57.1 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: Twofish - Decrypt GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 90 180 270 360 450 SE +/- 0.05, N = 3 SE +/- 0.07, N = 3 SE +/- 0.17, N = 3 SE +/- 0.56, N = 3 SE +/- 0.26, N = 3 411.57 420.37 374.71 373.56 411.52 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
GraphicsMagick This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Resizing GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 300 600 900 1200 1500 SE +/- 4.93, N = 3 SE +/- 2.91, N = 3 SE +/- 7.21, N = 3 SE +/- 3.51, N = 3 SE +/- 8.67, N = 3 1617 1444 1607 1571 1585 1. (CC) gcc options: -fopenmp -O3 -march=native -pthread -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread
TNN TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better TNN 0.3 Target: CPU - Model: MobileNet v2 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 80 160 240 320 400 SE +/- 0.34, N = 3 SE +/- 0.29, N = 3 SE +/- 0.17, N = 3 SE +/- 0.19, N = 3 SE +/- 0.31, N = 3 347.44 321.39 318.06 314.56 311.60 MIN: 345.68 / MAX: 356.59 MIN: 319.29 / MAX: 341.28 MIN: 316.44 / MAX: 326.16 MIN: 312.66 / MAX: 328.44 MIN: 309.73 / MAX: 322.67 1. (CXX) g++ options: -O3 -march=native -fopenmp -pthread -fvisibility=hidden -fvisibility=default -rdynamic -ldl
GraphicsMagick This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Rotate GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 200 400 600 800 1000 SE +/- 3.51, N = 3 SE +/- 7.07, N = 15 SE +/- 7.25, N = 15 SE +/- 2.52, N = 3 SE +/- 5.24, N = 3 776 809 794 852 765 1. (CC) gcc options: -fopenmp -O3 -march=native -pthread -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: Blowfish GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 110 220 330 440 550 SE +/- 0.29, N = 3 SE +/- 0.01, N = 3 SE +/- 0.07, N = 3 SE +/- 0.12, N = 3 SE +/- 0.22, N = 3 486.32 491.14 442.27 442.67 486.29 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
Timed MrBayes Analysis This test performs a bayesian analysis of a set of primate genome sequences in order to estimate their phylogeny. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Timed MrBayes Analysis 3.2.7 Primate Phylogeny Analysis GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 30 60 90 120 150 SE +/- 1.65, N = 12 SE +/- 0.45, N = 3 SE +/- 1.02, N = 3 SE +/- 0.32, N = 3 SE +/- 0.58, N = 3 157.33 146.10 145.17 142.93 154.63 1. (CC) gcc options: -mmmx -msse -msse2 -msse3 -mssse3 -msse4.1 -msse4.2 -maes -mavx -mfma -mavx2 -mavx512f -mavx512cd -mavx512vl -mavx512bw -mavx512dq -mrdrnd -mbmi -mbmi2 -madx -mmpx -mabm -O3 -std=c99 -pedantic -march=native -lm
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.1052 0.2104 0.3156 0.4208 0.526 SE +/- 0.002588, N = 3 SE +/- 0.003487, N = 3 SE +/- 0.000391, N = 3 SE +/- 0.004665, N = 3 SE +/- 0.004576, N = 4 0.467594 0.441500 0.425675 0.430676 0.426902 MIN: 0.44 MIN: 0.41 MIN: 0.41 MIN: 0.41 MIN: 0.4 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: Blowfish - Decrypt GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 100 200 300 400 500 SE +/- 0.21, N = 3 SE +/- 0.04, N = 3 SE +/- 0.07, N = 3 SE +/- 0.13, N = 3 SE +/- 0.30, N = 3 475.41 482.78 442.41 439.56 474.80 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
QuantLib QuantLib is an open-source library/framework around quantitative finance for modeling, trading and risk management scenarios. QuantLib is written in C++ with Boost and its built-in benchmark used reports the QuantLib Benchmark Index benchmark score. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MFLOPS, More Is Better QuantLib 1.21 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 600 1200 1800 2400 3000 SE +/- 35.29, N = 3 SE +/- 19.67, N = 3 SE +/- 0.85, N = 3 SE +/- 33.86, N = 4 SE +/- 18.68, N = 3 2568.8 2586.3 2773.6 2749.1 2529.8 1. (CXX) g++ options: -O3 -march=native -rdynamic
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU-v2-v2 - Model: mobilenet-v2 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1.206 2.412 3.618 4.824 6.03 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.04, N = 3 SE +/- 0.10, N = 3 SE +/- 0.01, N = 3 5.10 5.36 4.89 5.06 5.12 MIN: 4.73 / MAX: 10.4 MIN: 4.96 / MAX: 8.4 MIN: 4.72 / MAX: 10.07 MIN: 4.72 / MAX: 10.01 MIN: 4.76 / MAX: 8.89 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
Coremark This is a test of EEMBC CoreMark processor benchmark. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Iterations/Sec, More Is Better Coremark 1.0 CoreMark Size 666 - Iterations Per Second GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 140K 280K 420K 560K 700K SE +/- 1621.16, N = 3 SE +/- 3624.80, N = 3 SE +/- 2406.19, N = 3 SE +/- 1267.23, N = 3 SE +/- 2003.80, N = 3 650499.47 618050.83 601830.26 597455.16 630485.59 1. (CC) gcc options: -O2 -O3 -march=native -lrt" -lrt
Zstd Compression This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better Zstd Compression 1.5.0 Compression Level: 19, Long Mode - Decompression Speed GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 600 1200 1800 2400 3000 SE +/- 4.29, N = 3 SE +/- 10.70, N = 3 SE +/- 2.25, N = 3 SE +/- 14.84, N = 3 SE +/- 2.75, N = 3 3017.6 2819.2 2782.7 2775.7 2876.9 1. (CC) gcc options: -O3 -march=native -pthread -lz
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: CAST-256 - Decrypt GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 30 60 90 120 150 SE +/- 0.03, N = 3 SE +/- 0.01, N = 3 SE +/- 0.28, N = 3 SE +/- 0.50, N = 3 SE +/- 0.06, N = 3 150.88 152.43 140.48 140.85 151.25 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: CAST-256 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 30 60 90 120 150 SE +/- 0.03, N = 3 SE +/- 0.01, N = 3 SE +/- 0.30, N = 3 SE +/- 0.02, N = 3 SE +/- 0.04, N = 3 150.59 152.24 140.31 141.03 151.18 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
Gcrypt Library Libgcrypt is a general purpose cryptographic library developed as part of the GnuPG project. This is a benchmark of libgcrypt's integrated benchmark and is measuring the time to run the benchmark command with a cipher/mac/hash repetition count set for 50 times as simple, high level look at the overall crypto performance of the system under test. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Gcrypt Library 1.9 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 50 100 150 200 250 SE +/- 0.21, N = 3 SE +/- 0.27, N = 3 SE +/- 0.33, N = 3 SE +/- 0.18, N = 3 SE +/- 0.44, N = 3 193.97 194.12 196.22 208.34 193.41 1. (CC) gcc options: -O3 -march=native -fvisibility=hidden -lgpg-error
Mobile Neural Network MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.2 Model: inception-v3 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 8 16 24 32 40 SE +/- 0.28, N = 15 SE +/- 0.10, N = 3 SE +/- 0.43, N = 3 SE +/- 0.42, N = 3 SE +/- 0.42, N = 3 32.41 30.15 30.80 31.28 30.80 MIN: 29.17 / MAX: 33.89 MIN: 29.77 / MAX: 30.53 MIN: 30.14 / MAX: 31.87 MIN: 30.28 / MAX: 31.97 MIN: 30.12 / MAX: 31.88 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: googlenet GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3 6 9 12 15 SE +/- 0.28, N = 3 SE +/- 0.23, N = 3 SE +/- 0.31, N = 3 SE +/- 0.02, N = 3 SE +/- 0.08, N = 3 12.44 12.73 12.21 12.84 13.06 MIN: 12 / MAX: 13.22 MIN: 12.1 / MAX: 19.86 MIN: 11.77 / MAX: 13.02 MIN: 12.68 / MAX: 16.73 MIN: 12.84 / MAX: 14.29 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
TNN TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better TNN 0.3 Target: CPU - Model: SqueezeNet v2 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 16 32 48 64 80 SE +/- 1.05, N = 3 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.02, N = 3 SE +/- 0.07, N = 3 73.94 70.12 69.23 69.80 69.30 MIN: 72.33 / MAX: 77.08 MIN: 69.44 / MAX: 71.63 MIN: 68.59 / MAX: 70.34 MIN: 69.16 / MAX: 70.99 MIN: 68.65 / MAX: 70.61 1. (CXX) g++ options: -O3 -march=native -fopenmp -pthread -fvisibility=hidden -fvisibility=default -rdynamic -ldl
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: efficientnet-b0 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 2 4 6 8 10 SE +/- 0.01, N = 3 SE +/- 0.03, N = 3 SE +/- 0.09, N = 3 SE +/- 0.02, N = 3 SE +/- 0.11, N = 3 6.57 6.69 6.27 6.56 6.62 MIN: 6.28 / MAX: 14.64 MIN: 6.33 / MAX: 10.85 MIN: 6.05 / MAX: 14.37 MIN: 6.27 / MAX: 11.76 MIN: 6.24 / MAX: 24.41 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
Zstd Compression This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better Zstd Compression 1.5.0 Compression Level: 19 - Decompression Speed GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 600 1200 1800 2400 3000 SE +/- 14.28, N = 3 SE +/- 3.24, N = 3 SE +/- 16.86, N = 3 SE +/- 7.69, N = 9 SE +/- 9.76, N = 3 2802.7 2676.8 2773.9 2642.2 2701.9 1. (CC) gcc options: -O3 -march=native -pthread -lz
GraphicsMagick This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: HWB Color Space GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 200 400 600 800 1000 SE +/- 1.33, N = 3 SE +/- 1.15, N = 3 SE +/- 0.88, N = 3 864 903 884 916 864 1. (CC) gcc options: -fopenmp -O3 -march=native -pthread -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: resnet18 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3 6 9 12 15 SE +/- 0.28, N = 3 SE +/- 0.23, N = 3 SE +/- 0.01, N = 3 SE +/- 0.04, N = 3 SE +/- 0.04, N = 3 10.58 10.54 10.93 10.97 11.16 MIN: 10.2 / MAX: 11.57 MIN: 10.19 / MAX: 17.98 MIN: 10.84 / MAX: 11.27 MIN: 10.84 / MAX: 20.49 MIN: 11.03 / MAX: 11.45 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: mnasnet GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1.0845 2.169 3.2535 4.338 5.4225 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 SE +/- 0.05, N = 3 SE +/- 0.03, N = 3 SE +/- 0.10, N = 3 4.73 4.82 4.58 4.72 4.80 MIN: 4.44 / MAX: 11.54 MIN: 4.55 / MAX: 12.4 MIN: 4.39 / MAX: 10.87 MIN: 4.46 / MAX: 10.72 MIN: 4.42 / MAX: 16.22 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
VP9 libvpx Encoding This is a standard video encoding performance test of Google's libvpx library and the vpxenc command for the VP9 video format. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better VP9 libvpx Encoding 1.10.0 Speed: Speed 5 - Input: Bosphorus 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 2 4 6 8 10 SE +/- 0.01, N = 3 SE +/- 0.03, N = 3 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 8.61 8.22 8.65 8.63 8.59 1. (CXX) g++ options: -m64 -lm -lpthread -O3 -march=native -fPIC -U_FORTIFY_SOURCE -std=gnu++11
Mobile Neural Network MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.2 Model: mobilenetV3 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.5427 1.0854 1.6281 2.1708 2.7135 SE +/- 0.015, N = 15 SE +/- 0.012, N = 3 SE +/- 0.011, N = 3 SE +/- 0.032, N = 3 SE +/- 0.011, N = 3 2.297 2.403 2.372 2.412 2.341 MIN: 1.96 / MAX: 2.53 MIN: 2.28 / MAX: 2.54 MIN: 2.25 / MAX: 2.5 MIN: 2.23 / MAX: 2.61 MIN: 2.16 / MAX: 2.53 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
WebP Image Encode This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100, Highest Compression GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 2 4 6 8 10 SE +/- 0.007, N = 3 SE +/- 0.007, N = 3 SE +/- 0.043, N = 3 SE +/- 0.002, N = 3 SE +/- 0.003, N = 3 6.560 6.642 6.548 6.780 6.866 1. (CC) gcc options: -fvisibility=hidden -O3 -march=native -pthread -lm -ljpeg -lpng16
Crypto++ Crypto++ is a C++ class library of cryptographic algorithms. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/second, More Is Better Crypto++ 8.2 Test: Unkeyed Algorithms GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 80 160 240 320 400 SE +/- 0.65, N = 3 SE +/- 0.02, N = 3 SE +/- 0.68, N = 3 SE +/- 0.04, N = 3 SE +/- 0.17, N = 3 375.23 377.78 374.64 360.41 372.58 1. (CXX) g++ options: -O3 -march=native -fPIC -pthread -pipe
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 400 800 1200 1600 2000 SE +/- 1.53, N = 3 SE +/- 0.81, N = 3 SE +/- 1.54, N = 3 SE +/- 3.72, N = 3 SE +/- 1.06, N = 3 1639.78 1573.18 1564.71 1566.25 1566.75 MIN: 1630.86 MIN: 1566.72 MIN: 1557.51 MIN: 1553.36 MIN: 1558.77 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 400 800 1200 1600 2000 SE +/- 2.09, N = 3 SE +/- 1.09, N = 3 SE +/- 0.75, N = 3 SE +/- 2.29, N = 3 SE +/- 0.93, N = 3 1636.96 1573.03 1567.41 1563.05 1565.41 MIN: 1629.18 MIN: 1566.77 MIN: 1561.17 MIN: 1555.64 MIN: 1559.12 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3 6 9 12 15 SE +/- 0.00795, N = 3 SE +/- 0.02091, N = 3 SE +/- 0.02038, N = 3 SE +/- 0.01742, N = 3 SE +/- 0.02137, N = 3 9.79185 9.82858 9.59452 9.41346 9.39062 MIN: 9.62 MIN: 9.53 MIN: 9.4 MIN: 9.27 MIN: 9.29 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
Zstd Compression This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better Zstd Compression 1.5.0 Compression Level: 8 - Decompression Speed GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 700 1400 2100 2800 3500 SE +/- 2.27, N = 5 SE +/- 2.28, N = 3 SE +/- 2.91, N = 3 SE +/- 5.37, N = 3 SE +/- 3.06, N = 3 3436.8 3361.7 3332.7 3351.9 3285.5 1. (CC) gcc options: -O3 -march=native -pthread -lz
Etcpak Etcpack is the self-proclaimed "fastest ETC compressor on the planet" with focused on providing open-source, very fast ETC and S3 texture compression support. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 0.7 Configuration: DXT1 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 300 600 900 1200 1500 SE +/- 0.45, N = 3 SE +/- 1.51, N = 3 SE +/- 1.30, N = 3 SE +/- 0.47, N = 3 SE +/- 0.37, N = 3 1450.56 1445.24 1419.37 1468.97 1484.23 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
Stockfish This is a test of Stockfish, an advanced open-source C++11 chess benchmark that can scale up to 512 CPU threads. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Nodes Per Second, More Is Better Stockfish 13 Total Time GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 11M 22M 33M 44M 55M SE +/- 508293.85, N = 15 SE +/- 562135.06, N = 15 SE +/- 432778.20, N = 8 SE +/- 211973.29, N = 3 SE +/- 623930.99, N = 3 50622552 51140013 50734571 52206963 49942276 1. (CXX) g++ options: -lgcov -m64 -lpthread -O3 -march=native -fno-exceptions -std=c++17 -pedantic -msse -msse3 -mpopcnt -mavx2 -mavx512f -mavx512bw -mavx512vnni -mavx512dq -mavx512vl -msse4.1 -mssse3 -msse2 -mbmi2 -flto -fprofile-use -fno-peel-loops -fno-tracer -flto=jobserver
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.288 0.576 0.864 1.152 1.44 SE +/- 0.00348, N = 3 SE +/- 0.00683, N = 3 SE +/- 0.00255, N = 3 SE +/- 0.00400, N = 3 SE +/- 0.00554, N = 3 1.27991 1.23239 1.22496 1.23012 1.22703 MIN: 1.23 MIN: 1.18 MIN: 1.18 MIN: 1.19 MIN: 1.18 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
Mobile Neural Network MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.2 Model: MobileNetV2_224 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.8447 1.6894 2.5341 3.3788 4.2235 SE +/- 0.040, N = 15 SE +/- 0.069, N = 3 SE +/- 0.058, N = 3 SE +/- 0.018, N = 3 SE +/- 0.090, N = 3 3.594 3.704 3.729 3.754 3.669 MIN: 3.07 / MAX: 4.08 MIN: 3.31 / MAX: 3.95 MIN: 3.41 / MAX: 3.94 MIN: 3.47 / MAX: 3.95 MIN: 3.42 / MAX: 4.19 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
Zstd Compression This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better Zstd Compression 1.5.0 Compression Level: 8, Long Mode - Decompression Speed GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 800 1600 2400 3200 4000 SE +/- 2.52, N = 13 SE +/- 5.55, N = 3 SE +/- 3.04, N = 15 SE +/- 3.35, N = 3 SE +/- 3.49, N = 15 3628.4 3547.4 3531.4 3553.2 3479.0 1. (CC) gcc options: -O3 -march=native -pthread -lz
SVT-VP9 This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: VMAF Optimized - Input: Bosphorus 1080p GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 70 140 210 280 350 SE +/- 4.16, N = 3 SE +/- 2.56, N = 13 SE +/- 3.01, N = 5 SE +/- 2.83, N = 6 SE +/- 1.76, N = 14 295.07 306.08 293.71 297.32 299.97 1. (CC) gcc options: -O3 -fcommon -march=native -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: resnet50 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 5 10 15 20 25 SE +/- 0.29, N = 3 SE +/- 0.30, N = 3 SE +/- 0.24, N = 3 SE +/- 0.24, N = 3 SE +/- 0.29, N = 3 17.59 18.33 17.74 17.77 17.87 MIN: 17.07 / MAX: 18.69 MIN: 17.58 / MAX: 24.57 MIN: 17.09 / MAX: 18.96 MIN: 17.07 / MAX: 28.68 MIN: 17.16 / MAX: 18.62 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
GraphicsMagick This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Noise-Gaussian GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 90 180 270 360 450 SE +/- 0.33, N = 3 SE +/- 0.58, N = 3 SE +/- 0.33, N = 3 SE +/- 0.67, N = 3 387 390 403 403 403 1. (CC) gcc options: -fopenmp -O3 -march=native -pthread -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread
Liquid-DSP LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 2021.01.31 Threads: 36 - Buffer Length: 256 - Filter Length: 57 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 200M 400M 600M 800M 1000M SE +/- 588132.64, N = 3 SE +/- 120554.28, N = 3 SE +/- 1056729.76, N = 3 SE +/- 1013283.99, N = 3 SE +/- 272213.15, N = 3 921670000 916790000 937553333 954536667 940660000 1. (CC) gcc options: -O3 -march=native -pthread -lm -lc -lliquid
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU-v3-v3 - Model: mobilenet-v3 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1.062 2.124 3.186 4.248 5.31 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.05, N = 3 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 4.57 4.72 4.54 4.66 4.64 MIN: 4.37 / MAX: 9.05 MIN: 4.49 / MAX: 7.39 MIN: 4.37 / MAX: 11.47 MIN: 4.46 / MAX: 12.92 MIN: 4.36 / MAX: 10.01 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
Mobile Neural Network MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.2 Model: SqueezeNetV1.0 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1.2877 2.5754 3.8631 5.1508 6.4385 SE +/- 0.039, N = 15 SE +/- 0.056, N = 3 SE +/- 0.079, N = 3 SE +/- 0.020, N = 3 SE +/- 0.083, N = 3 5.540 5.594 5.506 5.723 5.612 MIN: 5.06 / MAX: 6.72 MIN: 5.4 / MAX: 5.85 MIN: 5.22 / MAX: 5.91 MIN: 5.47 / MAX: 6.72 MIN: 5.24 / MAX: 6.01 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
Crypto++ Crypto++ is a C++ class library of cryptographic algorithms. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/second, More Is Better Crypto++ 8.2 Test: Keyed Algorithms GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 160 320 480 640 800 SE +/- 0.35, N = 3 SE +/- 0.10, N = 3 SE +/- 0.24, N = 3 SE +/- 0.20, N = 3 SE +/- 0.48, N = 3 719.76 714.49 714.21 692.99 717.17 1. (CXX) g++ options: -O3 -march=native -fPIC -pthread -pipe
WebP Image Encode This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100, Lossless, Highest Compression GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 9 18 27 36 45 SE +/- 0.13, N = 3 SE +/- 0.11, N = 3 SE +/- 0.01, N = 3 SE +/- 0.00, N = 3 SE +/- 0.01, N = 3 36.87 37.70 37.16 36.47 37.87 1. (CC) gcc options: -fvisibility=hidden -O3 -march=native -pthread -lm -ljpeg -lpng16
dav1d Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org FPS, More Is Better dav1d 0.9.0 Video Input: Chimera 1080p 10-bit GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 50 100 150 200 250 SE +/- 0.34, N = 3 SE +/- 0.39, N = 3 SE +/- 0.48, N = 3 SE +/- 1.18, N = 3 SE +/- 0.50, N = 3 219.69 215.16 221.94 223.06 222.27 -lm - MIN: 156.35 / MAX: 406.23 -lm - MIN: 151.62 / MAX: 411.26 MIN: 157.38 / MAX: 404.98 MIN: 157.45 / MAX: 397.96 MIN: 157.09 / MAX: 436.51 1. (CC) gcc options: -O3 -march=native -pthread
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: squeezenet_ssd GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 4 8 12 16 20 SE +/- 0.06, N = 3 SE +/- 0.36, N = 3 SE +/- 0.13, N = 3 SE +/- 0.08, N = 3 SE +/- 0.05, N = 3 15.08 15.63 15.28 15.23 15.26 MIN: 14.88 / MAX: 21.62 MIN: 15.02 / MAX: 16.85 MIN: 14.89 / MAX: 16.1 MIN: 14.88 / MAX: 17.07 MIN: 14.97 / MAX: 18.92 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
x265 This is a simple test of the x265 encoder run on the CPU with 1080p and 4K options for H.265 video encode performance with x265. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better x265 3.4 Video Input: Bosphorus 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 5 10 15 20 25 SE +/- 0.12, N = 3 SE +/- 0.10, N = 3 SE +/- 0.11, N = 3 SE +/- 0.11, N = 3 SE +/- 0.13, N = 3 21.64 21.53 21.08 21.16 21.84 1. (CXX) g++ options: -O3 -march=native -rdynamic -lpthread -lrt -ldl -lnuma
dav1d Dav1d is an open-source, speedy AV1 video decoder. This test profile times how long it takes to decode sample AV1 video content. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org FPS, More Is Better dav1d 0.9.0 Video Input: Summer Nature 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 40 80 120 160 200 SE +/- 1.65, N = 3 SE +/- 0.89, N = 3 SE +/- 1.97, N = 6 SE +/- 1.13, N = 3 SE +/- 2.16, N = 3 199.89 197.46 194.06 192.95 195.15 -lm - MIN: 143.79 / MAX: 228.44 -lm - MIN: 150.4 / MAX: 226.05 MIN: 131.48 / MAX: 225.93 MIN: 132.83 / MAX: 217.9 MIN: 149.2 / MAX: 222.59 1. (CC) gcc options: -O3 -march=native -pthread
VP9 libvpx Encoding This is a standard video encoding performance test of Google's libvpx library and the vpxenc command for the VP9 video format. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better VP9 libvpx Encoding 1.10.0 Speed: Speed 0 - Input: Bosphorus 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1.107 2.214 3.321 4.428 5.535 SE +/- 0.04, N = 3 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.04, N = 3 SE +/- 0.03, N = 3 4.87 4.75 4.92 4.87 4.84 1. (CXX) g++ options: -m64 -lm -lpthread -O3 -march=native -fPIC -U_FORTIFY_SOURCE -std=gnu++11
ViennaCL ViennaCL is an open-source linear algebra library written in C++ and with support for OpenCL and OpenMP. This test profile makes use of ViennaCL's built-in benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GB/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - sAXPY GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 16 32 48 64 80 SE +/- 0.12, N = 3 SE +/- 1.48, N = 3 SE +/- 0.35, N = 3 SE +/- 0.31, N = 3 SE +/- 0.30, N = 3 71.2 68.8 70.1 70.7 70.3 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
TNN TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better TNN 0.3 Target: CPU - Model: SqueezeNet v1.1 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 60 120 180 240 300 SE +/- 0.07, N = 3 SE +/- 0.07, N = 3 SE +/- 0.19, N = 3 SE +/- 0.63, N = 3 SE +/- 0.10, N = 3 296.14 289.96 287.49 288.91 286.22 MIN: 294.66 / MAX: 298.56 MIN: 288.43 / MAX: 291.61 MIN: 285.88 / MAX: 299.62 MIN: 286.05 / MAX: 294.45 MIN: 285.02 / MAX: 287.82 1. (CXX) g++ options: -O3 -march=native -fopenmp -pthread -fvisibility=hidden -fvisibility=default -rdynamic -ldl
AOM AV1 This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.1 Encoder Mode: Speed 6 Two-Pass - Input: Bosphorus 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1.008 2.016 3.024 4.032 5.04 SE +/- 0.05, N = 3 SE +/- 0.06, N = 3 SE +/- 0.03, N = 10 SE +/- 0.04, N = 3 SE +/- 0.04, N = 3 4.44 4.46 4.40 4.33 4.48 1. (CXX) g++ options: -O3 -march=native -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.1 Encoder Mode: Speed 9 Realtime - Input: Bosphorus 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 7 14 21 28 35 SE +/- 0.07, N = 3 SE +/- 0.04, N = 3 SE +/- 0.04, N = 3 SE +/- 0.08, N = 3 SE +/- 0.15, N = 3 27.24 27.18 28.10 28.07 27.86 1. (CXX) g++ options: -O3 -march=native -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: regnety_400m GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 4 8 12 16 20 SE +/- 0.21, N = 3 SE +/- 0.14, N = 3 SE +/- 0.05, N = 3 SE +/- 0.11, N = 3 SE +/- 0.14, N = 3 13.87 13.77 13.48 13.93 13.63 MIN: 13.18 / MAX: 15.01 MIN: 12.93 / MAX: 14.62 MIN: 13.11 / MAX: 14.04 MIN: 13.17 / MAX: 14.48 MIN: 13.01 / MAX: 14.57 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
Etcpak Etcpack is the self-proclaimed "fastest ETC compressor on the planet" with focused on providing open-source, very fast ETC and S3 texture compression support. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 0.7 Configuration: ETC1 + Dithering GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 70 140 210 280 350 SE +/- 0.09, N = 3 SE +/- 2.72, N = 3 SE +/- 0.05, N = 3 SE +/- 0.20, N = 3 SE +/- 0.38, N = 3 336.11 329.13 325.28 327.33 329.49 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
ViennaCL ViennaCL is an open-source linear algebra library written in C++ and with support for OpenCL and OpenMP. This test profile makes use of ViennaCL's built-in benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GB/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - dGEMV-N GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 16 32 48 64 80 SE +/- 0.07, N = 3 SE +/- 0.12, N = 3 SE +/- 0.31, N = 3 SE +/- 0.09, N = 3 SE +/- 0.03, N = 3 71.8 69.5 71.4 71.7 71.6 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
Opus Codec Encoding Opus is an open audio codec. Opus is a lossy audio compression format designed primarily for interactive real-time applications over the Internet. This test uses Opus-Tools and measures the time required to encode a WAV file to Opus. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Opus Codec Encoding 1.3.1 WAV To Opus Encode GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 2 4 6 8 10 SE +/- 0.021, N = 5 SE +/- 0.011, N = 5 SE +/- 0.011, N = 5 SE +/- 0.013, N = 5 SE +/- 0.031, N = 5 8.241 8.283 8.186 8.456 8.283 1. (CXX) g++ options: -O3 -march=native -fvisibility=hidden -logg -lm
Zstd Compression This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better Zstd Compression 1.5.0 Compression Level: 8 - Compression Speed GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 90 180 270 360 450 SE +/- 4.23, N = 5 SE +/- 4.99, N = 3 SE +/- 5.56, N = 3 SE +/- 5.64, N = 3 SE +/- 4.87, N = 3 429.4 425.7 432.7 419.5 424.0 1. (CC) gcc options: -O3 -march=native -pthread -lz
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.6635 1.327 1.9905 2.654 3.3175 SE +/- 0.01601, N = 3 SE +/- 0.01360, N = 3 SE +/- 0.01797, N = 3 SE +/- 0.01584, N = 3 SE +/- 0.01261, N = 3 2.86008 2.94903 2.94670 2.93428 2.93978 MIN: 2.77 MIN: 2.85 MIN: 2.84 MIN: 2.83 MIN: 2.83 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
Tachyon This is a test of the threaded Tachyon, a parallel ray-tracing system, measuring the time to ray-trace a sample scene. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Tachyon 0.99b6 Total Time GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 11 22 33 44 55 SE +/- 0.08, N = 3 SE +/- 0.08, N = 3 SE +/- 0.18, N = 3 SE +/- 0.26, N = 3 SE +/- 0.08, N = 3 47.88 48.16 49.32 47.87 47.90 1. (CC) gcc options: -m64 -O3 -fomit-frame-pointer -ffast-math -ltachyon -lm -lpthread
Liquid-DSP LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org samples/s, More Is Better Liquid-DSP 2021.01.31 Threads: 32 - Buffer Length: 256 - Filter Length: 57 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 200M 400M 600M 800M 1000M SE +/- 539269.05, N = 3 SE +/- 353836.12, N = 3 SE +/- 4623189.13, N = 3 SE +/- 4781007.56, N = 3 SE +/- 2904171.33, N = 3 930236667 924630000 944433333 951170000 939716667 1. (CC) gcc options: -O3 -march=native -pthread -lm -lc -lliquid
SVT-VP9 This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: PSNR/SSIM Optimized - Input: Bosphorus 1080p GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 70 140 210 280 350 SE +/- 1.42, N = 3 SE +/- 0.21, N = 3 SE +/- 0.68, N = 3 SE +/- 2.35, N = 3 SE +/- 1.79, N = 3 302.25 310.79 303.54 305.72 306.89 1. (CC) gcc options: -O3 -fcommon -march=native -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 200 400 600 800 1000 SE +/- 1.46, N = 3 SE +/- 1.32, N = 3 SE +/- 0.15, N = 3 SE +/- 0.70, N = 3 SE +/- 0.16, N = 3 961.38 939.92 935.38 936.88 938.22 MIN: 955.17 MIN: 933.27 MIN: 931.25 MIN: 931.54 MIN: 930.52 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: blazeface GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.5873 1.1746 1.7619 2.3492 2.9365 SE +/- 0.02, N = 3 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 SE +/- 0.01, N = 3 SE +/- 0.07, N = 3 2.54 2.56 2.54 2.55 2.61 MIN: 2.45 / MAX: 3.32 MIN: 2.5 / MAX: 3.31 MIN: 2.46 / MAX: 3.12 MIN: 2.47 / MAX: 3.17 MIN: 2.47 / MAX: 3.3 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: mobilenet GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 4 8 12 16 20 SE +/- 0.04, N = 3 SE +/- 0.02, N = 3 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 SE +/- 0.06, N = 3 13.94 14.16 13.81 13.79 13.81 MIN: 13.68 / MAX: 22.27 MIN: 13.93 / MAX: 14.76 MIN: 13.64 / MAX: 14.54 MIN: 13.61 / MAX: 14.23 MIN: 13.51 / MAX: 20.36 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.4031 0.8062 1.2093 1.6124 2.0155 SE +/- 0.00782, N = 3 SE +/- 0.00719, N = 3 SE +/- 0.00613, N = 3 SE +/- 0.00508, N = 3 SE +/- 0.00494, N = 3 1.79150 1.76308 1.74471 1.76264 1.75520 MIN: 1.74 MIN: 1.7 MIN: 1.66 MIN: 1.69 MIN: 1.69 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
WebP Image Encode This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Encode Time - Seconds, Fewer Is Better WebP Image Encode 1.1 Encode Settings: Quality 100, Lossless GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 4 8 12 16 20 SE +/- 0.02, N = 3 SE +/- 0.06, N = 3 SE +/- 0.01, N = 3 SE +/- 0.05, N = 3 SE +/- 0.06, N = 3 16.82 17.20 16.89 16.80 17.25 1. (CC) gcc options: -fvisibility=hidden -O3 -march=native -pthread -lm -ljpeg -lpng16
ViennaCL ViennaCL is an open-source linear algebra library written in C++ and with support for OpenCL and OpenMP. This test profile makes use of ViennaCL's built-in benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GB/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - sCOPY GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 11 22 33 44 55 SE +/- 0.17, N = 3 SE +/- 0.27, N = 3 SE +/- 0.09, N = 3 SE +/- 0.07, N = 3 SE +/- 0.12, N = 3 47.1 45.9 46.4 46.7 46.4 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 200 400 600 800 1000 SE +/- 0.63, N = 3 SE +/- 0.50, N = 3 SE +/- 0.46, N = 3 SE +/- 0.60, N = 3 SE +/- 0.69, N = 3 960.13 937.61 936.49 938.04 935.96 MIN: 955.07 MIN: 932.91 MIN: 932.18 MIN: 933.04 MIN: 930.6 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
Timed HMMer Search This test searches through the Pfam database of profile hidden markov models. The search finds the domain structure of Drosophila Sevenless protein. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Timed HMMer Search 3.3.2 Pfam Database Search GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 30 60 90 120 150 SE +/- 0.10, N = 3 SE +/- 0.06, N = 3 SE +/- 0.10, N = 3 SE +/- 0.11, N = 3 SE +/- 0.32, N = 3 126.51 126.20 129.45 126.62 126.69 1. (CC) gcc options: -O3 -march=native -pthread -lhmmer -leasel -lm -lmpi
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: KASUMI GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 20 40 60 80 100 SE +/- 0.12, N = 3 SE +/- 0.01, N = 3 SE +/- 0.25, N = 3 SE +/- 0.07, N = 3 SE +/- 0.06, N = 3 100.41 99.72 100.61 100.70 98.20 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
Ngspice Ngspice is an open-source SPICE circuit simulator. Ngspice was originally based on the Berkeley SPICE electronic circuit simulator. Ngspice supports basic threading using OpenMP. This test profile is making use of the ISCAS 85 benchmark circuits. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Ngspice 34 Circuit: C7552 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 30 60 90 120 150 SE +/- 1.54, N = 3 SE +/- 1.35, N = 3 SE +/- 1.05, N = 3 SE +/- 0.26, N = 3 SE +/- 1.51, N = 3 127.69 126.20 128.32 129.40 126.88 1. (CC) gcc options: -O3 -march=native -fopenmp -lm -lfftw3 -lXaw -lXmu -lXt -lXext -lX11 -lSM -lICE
LAME MP3 Encoding LAME is an MP3 encoder licensed under the LGPL. This test measures the time required to encode a WAV file to MP3 format. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better LAME MP3 Encoding 3.100 WAV To MP3 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 2 4 6 8 10 SE +/- 0.011, N = 3 SE +/- 0.002, N = 3 SE +/- 0.003, N = 3 SE +/- 0.005, N = 3 SE +/- 0.004, N = 3 8.525 8.716 8.599 8.732 8.730 1. (CC) gcc options: -O3 -ffast-math -funroll-loops -fschedule-insns2 -fbranch-count-reg -fforce-addr -pipe -march=native -lm
7-Zip Compression This is a test of 7-Zip using p7zip with its integrated benchmark feature or upstream 7-Zip for the Windows x64 build. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MIPS, More Is Better 7-Zip Compression 16.02 Compress Speed Test GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 20K 40K 60K 80K 100K SE +/- 343.09, N = 3 SE +/- 69.57, N = 3 SE +/- 228.39, N = 3 SE +/- 304.44, N = 3 SE +/- 46.23, N = 3 98298 99711 98493 98149 97426 1. (CXX) g++ options: -pipe -lpthread
Zstd Compression This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better Zstd Compression 1.5.0 Compression Level: 19, Long Mode - Compression Speed GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 10 20 30 40 50 SE +/- 0.13, N = 3 SE +/- 0.15, N = 3 SE +/- 0.15, N = 3 SE +/- 0.07, N = 3 SE +/- 0.19, N = 3 43.5 43.3 43.9 44.3 43.9 1. (CC) gcc options: -O3 -march=native -pthread -lz
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3 6 9 12 15 SE +/- 0.01517, N = 3 SE +/- 0.01311, N = 3 SE +/- 0.00808, N = 3 SE +/- 0.01164, N = 3 SE +/- 0.01276, N = 3 9.55455 9.34766 9.34206 9.35550 9.35095 MIN: 9.5 MIN: 9.29 MIN: 9.28 MIN: 9.29 MIN: 9.29 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
Etcpak Etcpack is the self-proclaimed "fastest ETC compressor on the planet" with focused on providing open-source, very fast ETC and S3 texture compression support. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Mpx/s, More Is Better Etcpak 0.7 Configuration: ETC2 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 40 80 120 160 200 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.01, N = 3 SE +/- 0.08, N = 3 199.10 198.11 197.13 194.76 197.63 1. (CXX) g++ options: -O3 -march=native -std=c++11 -lpthread
PJSIP PJSIP is a free and open source multimedia communication library written in C language implementing standard based protocols such as SIP, SDP, RTP, STUN, TURN, and ICE. It combines signaling protocol (SIP) with rich multimedia framework and NAT traversal functionality into high level API that is portable and suitable for almost any type of systems ranging from desktops, embedded systems, to mobile handsets. This test profile is making use of pjsip-perf with both the client/server on teh system. More details on the PJSIP benchmark at https://www.pjsip.org/high-performance-sip.htm Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Responses Per Second, More Is Better PJSIP 2.11 Method: OPTIONS, Stateless GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 30K 60K 90K 120K 150K SE +/- 1635.40, N = 4 SE +/- 578.10, N = 3 SE +/- 734.59, N = 3 SE +/- 946.90, N = 3 SE +/- 1006.09, N = 3 135903 135302 137966 136444 138222 1. (CC) gcc options: -lstdc++ -lssl -lcrypto -lm -lrt -lpthread -O3 -march=native
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: alexnet GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3 6 9 12 15 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 SE +/- 0.09, N = 3 SE +/- 0.02, N = 3 SE +/- 0.00, N = 3 9.15 9.05 8.99 9.08 9.18 MIN: 9.08 / MAX: 9.59 MIN: 8.96 / MAX: 19.48 MIN: 8.73 / MAX: 9.39 MIN: 9 / MAX: 11.81 MIN: 9.11 / MAX: 9.74 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
PJSIP PJSIP is a free and open source multimedia communication library written in C language implementing standard based protocols such as SIP, SDP, RTP, STUN, TURN, and ICE. It combines signaling protocol (SIP) with rich multimedia framework and NAT traversal functionality into high level API that is portable and suitable for almost any type of systems ranging from desktops, embedded systems, to mobile handsets. This test profile is making use of pjsip-perf with both the client/server on teh system. More details on the PJSIP benchmark at https://www.pjsip.org/high-performance-sip.htm Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Responses Per Second, More Is Better PJSIP 2.11 Method: INVITE GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 700 1400 2100 2800 3500 SE +/- 27.02, N = 3 SE +/- 25.16, N = 15 SE +/- 25.40, N = 15 SE +/- 7.22, N = 3 SE +/- 6.36, N = 3 3252 3307 3304 3240 3281 1. (CC) gcc options: -lstdc++ -lssl -lcrypto -lm -lrt -lpthread -O3 -march=native
C-Ray This is a test of C-Ray, a simple raytracer designed to test the floating-point CPU performance. This test is multi-threaded (16 threads per core), will shoot 8 rays per pixel for anti-aliasing, and will generate a 1600 x 1200 image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better C-Ray 1.1 Total Time - 4K, 16 Rays Per Pixel GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 7 14 21 28 35 SE +/- 0.01, N = 3 SE +/- 0.03, N = 3 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 30.04 29.82 29.97 29.96 30.43 1. (CC) gcc options: -lm -lpthread -O3 -march=native
VOSK Speech Recognition Toolkit VOSK is an open-source offline speech recognition API/toolkit. VOSK supports speech recognition in 17 languages and has a variety of models available and interfaces for different programming languages. This test profile times the speech-to-text process for a roughly three minute audio recording. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better VOSK Speech Recognition Toolkit 0.3.21 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 5 10 15 20 25 SE +/- 0.13, N = 3 SE +/- 0.15, N = 3 SE +/- 0.07, N = 3 SE +/- 0.26, N = 3 SE +/- 0.12, N = 3 20.99 20.72 20.57 20.89 20.75
AOM AV1 This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.1 Encoder Mode: Speed 8 Realtime - Input: Bosphorus 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 5 10 15 20 25 SE +/- 0.06, N = 3 SE +/- 0.04, N = 3 SE +/- 0.09, N = 3 SE +/- 0.04, N = 3 SE +/- 0.01, N = 3 19.34 19.40 19.67 19.64 19.28 1. (CXX) g++ options: -O3 -march=native -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: KASUMI - Decrypt GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 20 40 60 80 100 SE +/- 0.03, N = 3 SE +/- 0.02, N = 3 SE +/- 0.07, N = 3 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 98.56 97.64 96.61 97.00 96.88 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
Himeno Benchmark The Himeno benchmark is a linear solver of pressure Poisson using a point-Jacobi method. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MFLOPS, More Is Better Himeno Benchmark 3.0 Poisson Pressure Solver GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1000 2000 3000 4000 5000 SE +/- 5.70, N = 3 SE +/- 0.61, N = 3 SE +/- 2.82, N = 3 SE +/- 0.64, N = 3 SE +/- 13.08, N = 3 4609.40 4522.73 4580.19 4592.95 4538.66 1. (CC) gcc options: -O3 -march=native -mavx2
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: vgg16 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 8 16 24 32 40 SE +/- 0.49, N = 3 SE +/- 0.37, N = 3 SE +/- 0.53, N = 3 SE +/- 0.52, N = 3 SE +/- 0.47, N = 3 36.01 36.20 36.70 36.56 36.35 MIN: 35.37 / MAX: 37.68 MIN: 35.36 / MAX: 47.25 MIN: 35.5 / MAX: 41.99 MIN: 35.42 / MAX: 58.41 MIN: 35.3 / MAX: 37.7 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
SecureMark SecureMark is an objective, standardized benchmarking framework for measuring the efficiency of cryptographic processing solutions developed by EEMBC. SecureMark-TLS is benchmarking Transport Layer Security performance with a focus on IoT/edge computing. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org marks, More Is Better SecureMark 1.0.4 Benchmark: SecureMark-TLS GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 60K 120K 180K 240K 300K SE +/- 247.29, N = 3 SE +/- 106.27, N = 3 SE +/- 101.43, N = 3 SE +/- 68.80, N = 3 SE +/- 95.66, N = 3 263457 259869 264514 259565 263472 1. (CC) gcc options: -pedantic -O3
GraphicsMagick This is a test of GraphicsMagick with its OpenMP implementation that performs various imaging tests on a sample 6000x4000 pixel JPEG image. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Iterations Per Minute, More Is Better GraphicsMagick 1.3.33 Operation: Enhanced GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 90 180 270 360 450 424 424 429 432 427 1. (CC) gcc options: -fopenmp -O3 -march=native -pthread -lfreetype -ljpeg -lXext -lSM -lICE -lX11 -lz -lm -lpthread
C-Blosc A simple, compressed, fast and persistent data store library for C. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better C-Blosc 2.0 Compressor: blosclz GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3K 6K 9K 12K 15K SE +/- 18.15, N = 3 SE +/- 11.42, N = 3 SE +/- 37.20, N = 3 SE +/- 69.56, N = 3 SE +/- 21.15, N = 3 11802.6 11800.7 11889.4 11926.8 11713.4 1. (CC) gcc options: -std=gnu99 -O3 -pthread -lrt -lm
Mobile Neural Network MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.2 Model: mobilenet-v1-1.0 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.5573 1.1146 1.6719 2.2292 2.7865 SE +/- 0.012, N = 15 SE +/- 0.025, N = 3 SE +/- 0.029, N = 3 SE +/- 0.024, N = 3 SE +/- 0.025, N = 3 2.455 2.433 2.477 2.477 2.467 MIN: 2.23 / MAX: 3.16 MIN: 2.32 / MAX: 2.62 MIN: 2.3 / MAX: 2.74 MIN: 2.3 / MAX: 2.65 MIN: 2.32 / MAX: 2.68 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
AOM AV1 This is a test of the AOMedia AV1 encoder (libaom) developed by AOMedia and Google. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better AOM AV1 3.1 Encoder Mode: Speed 6 Realtime - Input: Bosphorus 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 2 4 6 8 10 SE +/- 0.08, N = 15 SE +/- 0.05, N = 3 SE +/- 0.08, N = 12 SE +/- 0.04, N = 3 SE +/- 0.04, N = 3 7.43 7.51 7.49 7.51 7.38 1. (CXX) g++ options: -O3 -march=native -std=c++11 -U_FORTIFY_SOURCE -lm -lpthread
Crypto++ Crypto++ is a C++ class library of cryptographic algorithms. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/second, More Is Better Crypto++ 8.2 Test: Integer + Elliptic Curve Public Key Algorithms GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1200 2400 3600 4800 6000 SE +/- 6.16, N = 3 SE +/- 1.33, N = 3 SE +/- 5.19, N = 3 SE +/- 1.99, N = 3 SE +/- 1.85, N = 3 5538.55 5519.14 5532.47 5593.13 5503.08 1. (CXX) g++ options: -O3 -march=native -fPIC -pthread -pipe
SVT-VP9 This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-VP9 CPU-based multi-threaded video encoder for the VP9 video format with a sample YUV input video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-VP9 0.3 Tuning: Visual Quality Optimized - Input: Bosphorus 1080p GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 50 100 150 200 250 SE +/- 1.22, N = 3 SE +/- 1.34, N = 3 SE +/- 1.58, N = 3 SE +/- 3.10, N = 3 SE +/- 3.33, N = 3 243.48 247.45 244.56 244.05 246.75 1. (CC) gcc options: -O3 -fcommon -march=native -fPIE -fPIC -fvisibility=hidden -pie -rdynamic -lpthread -lrt -lm
Kvazaar This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 4K - Video Preset: Very Fast GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 5 10 15 20 25 SE +/- 0.04, N = 3 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.06, N = 3 21.20 21.35 21.15 21.01 21.09 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O3 -march=native -lpthread -lm -lrt
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: shufflenet-v2 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1.1363 2.2726 3.4089 4.5452 5.6815 SE +/- 0.01, N = 3 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 SE +/- 0.04, N = 3 SE +/- 0.02, N = 3 4.97 5.05 4.98 5.05 5.02 MIN: 4.8 / MAX: 8.9 MIN: 4.83 / MAX: 14.14 MIN: 4.88 / MAX: 8.6 MIN: 4.88 / MAX: 9.37 MIN: 4.83 / MAX: 15.94 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
SQLite Speedtest This is a benchmark of SQLite's speedtest1 benchmark program with an increased problem size of 1,000. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better SQLite Speedtest 3.30 Timed Time - Size 1,000 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 13 26 39 52 65 SE +/- 0.09, N = 3 SE +/- 0.22, N = 3 SE +/- 0.11, N = 3 SE +/- 0.10, N = 3 SE +/- 0.19, N = 3 57.03 57.61 56.72 57.41 57.30 1. (CC) gcc options: -O3 -march=native -ldl -lz -lpthread
FLAC Audio Encoding This test times how long it takes to encode a sample WAV file to FLAC format five times. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better FLAC Audio Encoding 1.3.2 WAV To FLAC GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 2 4 6 8 10 SE +/- 0.006, N = 5 SE +/- 0.012, N = 5 SE +/- 0.014, N = 5 SE +/- 0.009, N = 5 SE +/- 0.004, N = 5 8.500 8.436 8.379 8.411 8.369 1. (CXX) g++ options: -O3 -march=native -fvisibility=hidden -logg -lm
Zstd Compression This test measures the time needed to compress/decompress a sample file (a FreeBSD disk image - FreeBSD-12.2-RELEASE-amd64-memstick.img) using Zstd compression with options for different compression levels / settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MB/s, More Is Better Zstd Compression 1.5.0 Compression Level: 19 - Compression Speed GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 14 28 42 56 70 SE +/- 0.50, N = 3 SE +/- 0.19, N = 3 SE +/- 0.47, N = 3 SE +/- 0.50, N = 9 SE +/- 0.56, N = 3 60.2 60.3 60.8 61.0 60.1 1. (CC) gcc options: -O3 -march=native -pthread -lz
libjpeg-turbo tjbench tjbench is a JPEG decompression/compression benchmark that is part of libjpeg-turbo, a JPEG image codec library optimized for SIMD instructions on modern CPU architectures. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Megapixels/sec, More Is Better libjpeg-turbo tjbench 2.1.0 Test: Decompression Throughput GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 50 100 150 200 250 SE +/- 0.84, N = 3 SE +/- 0.03, N = 3 SE +/- 0.26, N = 3 SE +/- 0.47, N = 3 SE +/- 0.30, N = 3 220.62 218.81 217.43 218.62 219.54 1. (CC) gcc options: -O3 -march=native -rdynamic
SVT-AV1 This is a benchmark of the SVT-AV1 open-source video encoder/decoder. SVT-AV1 was originally developed by Intel as part of their Open Visual Cloud / Scalable Video Technology (SVT). Development of SVT-AV1 has since moved to the Alliance for Open Media as part of upstream AV1 development. SVT-AV1 is a CPU-based multi-threaded video encoder for the AV1 video format with a sample YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 0.8.7 Encoder Mode: Preset 4 - Input: Bosphorus 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.3058 0.6116 0.9174 1.2232 1.529 SE +/- 0.001, N = 3 SE +/- 0.005, N = 3 SE +/- 0.002, N = 3 SE +/- 0.003, N = 3 SE +/- 0.003, N = 3 1.355 1.340 1.348 1.347 1.359 1. (CXX) g++ options: -O3 -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
OpenBenchmarking.org Frames Per Second, More Is Better SVT-AV1 0.8.7 Encoder Mode: Preset 8 - Input: Bosphorus 4K GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3 6 9 12 15 SE +/- 0.08, N = 3 SE +/- 0.02, N = 3 SE +/- 0.03, N = 3 SE +/- 0.02, N = 3 SE +/- 0.04, N = 3 12.02 11.93 11.97 12.09 12.06 1. (CXX) g++ options: -O3 -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq -pie
ViennaCL ViennaCL is an open-source linear algebra library written in C++ and with support for OpenCL and OpenMP. This test profile makes use of ViennaCL's built-in benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GB/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - sDOT GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 20 40 60 80 100 SE +/- 0.26, N = 3 SE +/- 0.42, N = 3 SE +/- 0.49, N = 3 SE +/- 0.43, N = 3 SE +/- 0.33, N = 3 78.0 77.0 77.1 77.6 77.4 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
Kvazaar This is a test of Kvazaar as a CPU-based H.265 video encoder written in the C programming language and optimized in Assembly. Kvazaar is the winner of the 2016 ACM Open-Source Software Competition and developed at the Ultra Video Group, Tampere University, Finland. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better Kvazaar 2.0 Video Input: Bosphorus 4K - Video Preset: Ultra Fast GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 9 18 27 36 45 SE +/- 0.07, N = 3 SE +/- 0.07, N = 3 SE +/- 0.08, N = 3 SE +/- 0.11, N = 3 SE +/- 0.15, N = 3 40.16 40.13 40.20 40.31 40.64 1. (CC) gcc options: -pthread -ftree-vectorize -fvisibility=hidden -O3 -march=native -lpthread -lm -lrt
PJSIP PJSIP is a free and open source multimedia communication library written in C language implementing standard based protocols such as SIP, SDP, RTP, STUN, TURN, and ICE. It combines signaling protocol (SIP) with rich multimedia framework and NAT traversal functionality into high level API that is portable and suitable for almost any type of systems ranging from desktops, embedded systems, to mobile handsets. This test profile is making use of pjsip-perf with both the client/server on teh system. More details on the PJSIP benchmark at https://www.pjsip.org/high-performance-sip.htm Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Responses Per Second, More Is Better PJSIP 2.11 Method: OPTIONS, Stateful GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1200 2400 3600 4800 6000 SE +/- 24.67, N = 3 SE +/- 16.26, N = 3 SE +/- 23.13, N = 3 SE +/- 53.69, N = 3 SE +/- 8.67, N = 3 5729 5801 5763 5769 5744 1. (CC) gcc options: -lstdc++ -lssl -lcrypto -lm -lrt -lpthread -O3 -march=native
SVT-HEVC This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 1 - Input: Bosphorus 1080p GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 SE +/- 0.01, N = 3 SE +/- 0.00, N = 3 12.91 12.88 12.94 12.96 13.04 1. (CC) gcc options: -O3 -march=native -fPIE -fPIC -O2 -pie -rdynamic -lpthread -lrt
Ngspice Ngspice is an open-source SPICE circuit simulator. Ngspice was originally based on the Berkeley SPICE electronic circuit simulator. Ngspice supports basic threading using OpenMP. This test profile is making use of the ISCAS 85 benchmark circuits. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better Ngspice 34 Circuit: C2670 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 30 60 90 120 150 SE +/- 0.71, N = 3 SE +/- 0.81, N = 3 SE +/- 0.83, N = 3 SE +/- 0.92, N = 3 SE +/- 1.13, N = 3 134.22 134.35 135.43 134.71 133.91 1. (CC) gcc options: -O3 -march=native -fopenmp -lm -lfftw3 -lXaw -lXmu -lXt -lXext -lX11 -lSM -lICE
SVT-HEVC This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 10 - Input: Bosphorus 1080p GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 80 160 240 320 400 SE +/- 0.96, N = 3 SE +/- 0.95, N = 3 SE +/- 0.75, N = 3 SE +/- 1.99, N = 3 SE +/- 1.74, N = 3 376.89 374.62 375.32 375.65 372.69 1. (CC) gcc options: -O3 -march=native -fPIE -fPIC -O2 -pie -rdynamic -lpthread -lrt
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1.2589 2.5178 3.7767 5.0356 6.2945 SE +/- 0.02285, N = 3 SE +/- 0.02371, N = 3 SE +/- 0.02089, N = 3 SE +/- 0.02335, N = 3 SE +/- 0.01959, N = 3 5.59533 5.54062 5.53672 5.54544 5.53334 MIN: 5.45 MIN: 5.4 MIN: 5.38 MIN: 5.4 MIN: 5.38 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3 6 9 12 15 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 SE +/- 0.02, N = 3 SE +/- 0.01, N = 3 10.96 11.02 10.93 10.92 10.92 MIN: 10.74 MIN: 10.79 MIN: 10.77 MIN: 10.74 MIN: 10.74 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
SVT-HEVC This is a test of the Intel Open Visual Cloud Scalable Video Technology SVT-HEVC CPU-based multi-threaded video encoder for the HEVC / H.265 video format with a sample 1080p YUV video file. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Frames Per Second, More Is Better SVT-HEVC 1.5.0 Tuning: 7 - Input: Bosphorus 1080p GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 40 80 120 160 200 SE +/- 0.18, N = 3 SE +/- 0.58, N = 3 SE +/- 0.32, N = 3 SE +/- 0.31, N = 3 SE +/- 0.44, N = 3 190.62 190.03 191.33 191.02 189.88 1. (CC) gcc options: -O3 -march=native -fPIE -fPIC -O2 -pie -rdynamic -lpthread -lrt
ViennaCL ViennaCL is an open-source linear algebra library written in C++ and with support for OpenCL and OpenMP. This test profile makes use of ViennaCL's built-in benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GB/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - dGEMV-T GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 20 40 60 80 100 SE +/- 0.10, N = 3 SE +/- 0.25, N = 3 SE +/- 0.09, N = 3 SE +/- 0.38, N = 3 SE +/- 0.07, N = 3 79.7 79.9 79.5 79.3 79.6 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: AES-256 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 900 1800 2700 3600 4500 SE +/- 0.82, N = 3 SE +/- 0.87, N = 3 SE +/- 7.30, N = 3 SE +/- 2.67, N = 3 SE +/- 4.08, N = 3 3987.08 3998.63 3972.02 3985.27 3999.25 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
GnuPG This test times how long it takes to encrypt a sample file using GnuPG. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better GnuPG 2.2.27 2.7GB Sample File Encryption GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 14 28 42 56 70 SE +/- 0.19, N = 3 SE +/- 0.17, N = 3 SE +/- 0.56, N = 3 SE +/- 0.23, N = 3 SE +/- 0.36, N = 3 64.23 64.24 64.61 64.20 64.42 1. (CC) gcc options: -O3 -march=native
TNN TNN is an open-source deep learning reasoning framework developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better TNN 0.3 Target: CPU - Model: DenseNet GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 800 1600 2400 3200 4000 SE +/- 2.55, N = 3 SE +/- 0.16, N = 3 SE +/- 0.21, N = 3 SE +/- 0.77, N = 3 SE +/- 0.59, N = 3 3527.68 3505.93 3524.75 3508.39 3508.29 MIN: 3508.67 / MAX: 3981.67 MIN: 3487.54 / MAX: 3535.34 MIN: 3509.67 / MAX: 3548.51 MIN: 3486.98 / MAX: 3606.8 MIN: 3489.27 / MAX: 3603.98 1. (CXX) g++ options: -O3 -march=native -fopenmp -pthread -fvisibility=hidden -fvisibility=default -rdynamic -ldl
ViennaCL ViennaCL is an open-source linear algebra library written in C++ and with support for OpenCL and OpenMP. This test profile makes use of ViennaCL's built-in benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GB/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - dAXPY GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 13 26 39 52 65 SE +/- 0.03, N = 3 SE +/- 0.00, N = 3 SE +/- 0.12, N = 3 SE +/- 0.00, N = 3 SE +/- 0.03, N = 3 57.4 57.1 57.2 57.4 57.3 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
OpenBenchmarking.org GB/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - dCOPY GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 9 18 27 36 45 SE +/- 0.15, N = 3 SE +/- 0.00, N = 3 SE +/- 0.03, N = 3 SE +/- 0.00, N = 3 SE +/- 0.03, N = 3 38.1 38.1 38.2 38.3 38.3 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.1189 0.2378 0.3567 0.4756 0.5945 SE +/- 0.003361, N = 3 SE +/- 0.003001, N = 3 SE +/- 0.003225, N = 3 SE +/- 0.003036, N = 3 SE +/- 0.003324, N = 3 0.526798 0.528416 0.527939 0.528133 0.526014 MIN: 0.5 MIN: 0.5 MIN: 0.5 MIN: 0.5 MIN: 0.5 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 0.1038 0.2076 0.3114 0.4152 0.519 SE +/- 0.001062, N = 3 SE +/- 0.000631, N = 3 SE +/- 0.001427, N = 3 SE +/- 0.000687, N = 3 SE +/- 0.001585, N = 3 0.459456 0.461459 0.459660 0.459956 0.460482 MIN: 0.45 MIN: 0.45 MIN: 0.45 MIN: 0.45 MIN: 0.45 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
WavPack Audio Encoding This test times how long it takes to encode a sample WAV file to WavPack format with very high quality settings. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org Seconds, Fewer Is Better WavPack Audio Encoding 5.3 WAV To WavPack GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 3 6 9 12 15 SE +/- 0.00, N = 5 SE +/- 0.00, N = 5 SE +/- 0.00, N = 5 SE +/- 0.00, N = 5 SE +/- 0.02, N = 5 13.38 13.36 13.33 13.35 13.34 1. (CXX) g++ options: -O3 -march=native -rdynamic
Botan Botan is a BSD-licensed cross-platform open-source C++ crypto library "cryptography toolkit" that supports most publicly known cryptographic algorithms. The project's stated goal is to be "the best option for cryptography in C++ by offering the tools necessary to implement a range of practical systems, such as TLS protocol, X.509 certificates, modern AEAD ciphers, PKCS#11 and TPM hardware support, password hashing, and post quantum crypto schemes." Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org MiB/s, More Is Better Botan 2.17.3 Test: AES-256 - Decrypt GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 900 1800 2700 3600 4500 SE +/- 0.39, N = 3 SE +/- 1.28, N = 3 SE +/- 4.17, N = 3 SE +/- 0.79, N = 3 SE +/- 0.87, N = 3 3993.89 3991.20 3993.17 3995.23 3998.48 1. (CXX) g++ options: -fstack-protector -m64 -pthread -lbotan-2 -ldl -lrt
OpenBenchmarking.org FPS, More Is Better libgav1 0.16.3 Video Input: Summer Nature 4K GCC 12.0.0 20210701 7 14 21 28 35 SE +/- 0.01, N = 3 28.24 1. (CXX) g++ options: -O3 -march=native -lpthread -lrt
NCNN NCNN is a high performance neural network inference framework optimized for mobile and other platforms developed by Tencent. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better NCNN 20210525 Target: CPU - Model: yolov4-tiny GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 5 10 15 20 25 SE +/- 0.19, N = 3 SE +/- 0.25, N = 3 SE +/- 1.86, N = 3 SE +/- 0.10, N = 3 SE +/- 0.17, N = 3 21.35 20.84 22.83 21.37 21.26 MIN: 20.42 / MAX: 33.9 MIN: 19.92 / MAX: 24.91 MIN: 20.18 / MAX: 937.4 MIN: 20.44 / MAX: 22.72 MIN: 20 / MAX: 24.4 1. (CXX) g++ options: -O3 -march=native -rdynamic -lgomp -lpthread -pthread
Mobile Neural Network MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better Mobile Neural Network 1.2 Model: squeezenetv1.1 GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 1.037 2.074 3.111 4.148 5.185 SE +/- 0.061, N = 15 SE +/- 0.036, N = 3 SE +/- 0.149, N = 3 SE +/- 0.007, N = 3 SE +/- 0.162, N = 3 4.420 4.564 4.283 4.609 4.271 MIN: 3.98 / MAX: 4.76 MIN: 4.42 / MAX: 4.75 MIN: 3.97 / MAX: 4.71 MIN: 4.51 / MAX: 4.78 MIN: 3.97 / MAX: 4.72 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl
ViennaCL ViennaCL is an open-source linear algebra library written in C++ and with support for OpenCL and OpenMP. This test profile makes use of ViennaCL's built-in benchmarks. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org GB/s, More Is Better ViennaCL 1.7.1 Test: CPU BLAS - dDOT GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 14 28 42 56 70 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.03, N = 3 SE +/- 0.00, N = 3 SE +/- 9.10, N = 3 63.7 63.6 63.7 63.7 54.7 1. (CXX) g++ options: -O3 -march=native -fopenmp -rdynamic -lOpenCL
oneDNN This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI initiative. Learn more via the OpenBenchmarking.org test page.
OpenBenchmarking.org ms, Fewer Is Better oneDNN 2.1.2 Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU GCC 9.4 GCC 8.5 GCC 12.0.0 20210701 GCC 11.1 GCC 10.3 2 4 6 8 10 SE +/- 0.17046, N = 14 SE +/- 0.02602, N = 3 SE +/- 0.03413, N = 3 SE +/- 0.03513, N = 3 SE +/- 0.03159, N = 3 8.10038 7.89032 7.91258 7.90477 7.90163 MIN: 7.58 MIN: 7.58 MIN: 7.63 MIN: 7.56 MIN: 7.61 1. (CXX) g++ options: -O3 -march=native -std=c++11 -fopenmp -msse4.1 -fPIC -pie -lpthread -ldl
GCC 8.5 Processor: Intel Core i9-10980XE @ 4.80GHz (18 Cores / 36 Threads), Motherboard: ASRock X299 Steel Legend (P1.30 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 32GB, Disk: Samsung SSD 970 PRO 512GB, Graphics: NVIDIA NV132 11GB, Audio: Realtek ALC1220, Monitor: ASUS VP28U, Network: Intel I219-V + Intel I211
OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 8.5.0, File-System: ext4, Screen Resolution: 2560x1600
Kernel Notes: Transparent Huge Pages: madviseEnvironment Notes: CXXFLAGS="-O3 -march=native" CFLAGS="-O3 -march=native"Compiler Notes: --disable-multilib --enable-checking=release --enable-languages=c,c++Processor Notes: Scaling Governor: intel_cpufreq schedutil - CPU Microcode: 0x5003102Python Notes: Python 3.9.5Security Notes: itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled
Testing initiated at 3 July 2021 06:26 by user phoronix.
GCC 9.4 Processor: Intel Core i9-10980XE @ 4.80GHz (18 Cores / 36 Threads), Motherboard: ASRock X299 Steel Legend (P1.30 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 32GB, Disk: Samsung SSD 970 PRO 512GB, Graphics: NVIDIA NV132 11GB, Audio: Realtek ALC1220, Monitor: ASUS VP28U, Network: Intel I219-V + Intel I211
OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 9.4.0, File-System: ext4, Screen Resolution: 2560x1600
Kernel Notes: Transparent Huge Pages: madviseEnvironment Notes: CXXFLAGS="-O3 -march=native" CFLAGS="-O3 -march=native"Compiler Notes: --disable-multilib --enable-checking=release --enable-languages=c,c++Processor Notes: Scaling Governor: intel_cpufreq schedutil - CPU Microcode: 0x5003102Python Notes: Python 3.9.5Security Notes: itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled
Testing initiated at 2 July 2021 21:12 by user phoronix.
GCC 10.3 Processor: Intel Core i9-10980XE @ 4.80GHz (18 Cores / 36 Threads), Motherboard: ASRock X299 Steel Legend (P1.30 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 32GB, Disk: Samsung SSD 970 PRO 512GB, Graphics: NVIDIA NV132 11GB, Audio: Realtek ALC1220, Monitor: ASUS VP28U, Network: Intel I219-V + Intel I211
OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 10.3.0, File-System: ext4, Screen Resolution: 2560x1600
Kernel Notes: Transparent Huge Pages: madviseEnvironment Notes: CXXFLAGS="-O3 -march=native" CFLAGS="-O3 -march=native"Compiler Notes: --disable-multilib --enable-checking=release --enable-languages=c,c++Processor Notes: Scaling Governor: intel_cpufreq schedutil - CPU Microcode: 0x5003102Python Notes: Python 3.9.5Security Notes: itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled
Testing initiated at 2 July 2021 11:17 by user phoronix.
GCC 11.1 Processor: Intel Core i9-10980XE @ 4.80GHz (18 Cores / 36 Threads), Motherboard: ASRock X299 Steel Legend (P1.30 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 32GB, Disk: Samsung SSD 970 PRO 512GB, Graphics: NVIDIA NV132 11GB, Audio: Realtek ALC1220, Monitor: ASUS VP28U, Network: Intel I219-V + Intel I211
OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 11.1.0, File-System: ext4, Screen Resolution: 2560x1600
Kernel Notes: Transparent Huge Pages: madviseEnvironment Notes: CXXFLAGS="-O3 -march=native" CFLAGS="-O3 -march=native"Compiler Notes: --disable-multilib --enable-checking=release --enable-languages=c,c++Processor Notes: Scaling Governor: intel_cpufreq schedutil - CPU Microcode: 0x5003102Python Notes: Python 3.9.5Security Notes: itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled
Testing initiated at 2 July 2021 04:57 by user phoronix.
GCC 12.0.0 20210701 Processor: Intel Core i9-10980XE @ 4.80GHz (18 Cores / 36 Threads), Motherboard: ASRock X299 Steel Legend (P1.30 BIOS), Chipset: Intel Sky Lake-E DMI3 Registers, Memory: 32GB, Disk: Samsung SSD 970 PRO 512GB, Graphics: NVIDIA NV132 11GB, Audio: Realtek ALC1220, Monitor: ASUS VP28U, Network: Intel I219-V + Intel I211
OS: Ubuntu 21.04, Kernel: 5.11.0-22-generic (x86_64), Desktop: GNOME Shell 3.38.4, Display Server: X Server + Wayland, Display Driver: nouveau, OpenGL: 4.3 Mesa 21.0.1, Vulkan: 1.0.2, Compiler: GCC 12.0.0 20210701, File-System: ext4, Screen Resolution: 2560x1600
Kernel Notes: Transparent Huge Pages: madviseEnvironment Notes: CXXFLAGS="-O3 -march=native" CFLAGS="-O3 -march=native"Compiler Notes: --disable-multilib --enable-checking=release --enable-languages=c,c++Processor Notes: Scaling Governor: intel_cpufreq schedutil - CPU Microcode: 0x5003102Python Notes: Python 3.9.5Security Notes: itlb_multihit: KVM: Mitigation of VMX disabled + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Enhanced IBRS IBPB: conditional RSB filling + srbds: Not affected + tsx_async_abort: Mitigation of TSX disabled
Testing initiated at 1 July 2021 13:33 by user phoronix.