new new

AMD Ryzen 7 5800X3D 8-Core testing with a ASUS ROG CROSSHAIR VIII HERO (4201 BIOS) and AMD Radeon RX 6700 XT 12GB on Ubuntu 22.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2209116-PTS-NEWNEW6341
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts
Allow Limiting Results To Certain Suite(s)

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Toggle/Hide
Result
Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
A
September 10 2022
  3 Hours, 14 Minutes
B
September 11 2022
  3 Hours, 14 Minutes
C
September 11 2022
  3 Hours, 15 Minutes
Invert Behavior (Only Show Selected Data)
  3 Hours, 14 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


new newOpenBenchmarking.orgPhoronix Test SuiteAMD Ryzen 7 5800X3D 8-Core @ 3.40GHz (8 Cores / 16 Threads)ASUS ROG CROSSHAIR VIII HERO (4201 BIOS)AMD Starship/Matisse32GB1000GB Western Digital WDS100T1X0E-00AFY0AMD Radeon RX 6700 XT 12GB (2855/1000MHz)AMD Navi 21 HDMI AudioASUS MG28URealtek RTL8125 2.5GbE + Intel I211Ubuntu 22.045.17.0-1013-oem (x86_64)GNOME Shell 42.2X Server 1.21.1.3 + Wayland4.6 Mesa 22.2.0-devel (git-a9610ab740) (LLVM 13.0.1 DRM 3.44)1.3.219GCC 11.2.0ext43840x2160ProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDesktopDisplay ServerOpenGLVulkanCompilerFile-SystemScreen ResolutionNew New BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-11-gBFGDP/gcc-11-11.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0xa20120a - BAR1 / Visible vRAM Size: 12272 MB - vBIOS Version: 113-D5121100-101- Python 3.10.4- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling + srbds: Not affected + tsx_async_abort: Not affected

ABCResult OverviewPhoronix Test Suite100%101%102%102%103%NatronGravityMarkMobile Neural NetworkTimed Erlang/OTP CompilationApache CouchDBsrsRANWebP Image EncodeTimed CPython Compilation7-Zip CompressionFLAC Audio EncodingBRL-CADWebP2 Image EncodeTimed PHP CompilationTimed Wasmer CompilationOpenVINOFacebook RocksDBBlenderTimed Node.js Compilation

new newgravitymark: 3840 x 2160 - OpenGLsrsran: OFDM_Testnatron: Spaceshipmnn: nasnetsrsran: 5G PHY_DL_NR Test 52 PRB SISO 64-QAMcouchdb: 100 - 1000 - 30srsran: 5G PHY_DL_NR Test 52 PRB SISO 64-QAMmnn: MobileNetV2_224mnn: squeezenetv1.1gravitymark: 2560 x 1440 - OpenGLopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUgravitymark: 1920 x 1200 - OpenGLsrsran: 4G PHY_DL_Test 100 PRB SISO 256-QAMwebp: Quality 100, Losslessopenvino: Person Detection FP32 - CPUwebp: Quality 100srsran: 4G PHY_DL_Test 100 PRB MIMO 64-QAMgravitymark: 1920 x 1080 - OpenGLsrsran: 4G PHY_DL_Test 100 PRB MIMO 64-QAMrocksdb: Read Rand Write Randopenvino: Person Detection FP32 - CPUsrsran: 4G PHY_DL_Test 100 PRB SISO 256-QAMmnn: inception-v3rocksdb: Read While Writingcouchdb: 100 - 3000 - 30gravitymark: 2560 x 1440 - Vulkanmnn: mobilenetV3webp: Defaultwebp2: Defaultbuild-erlang: Time To Compilecouchdb: 300 - 1000 - 30mnn: mobilenet-v1-1.0build-python: Released Build, PGO + LTO Optimizedsrsran: 4G PHY_DL_Test 100 PRB SISO 64-QAMsrsran: 4G PHY_DL_Test 100 PRB MIMO 256-QAMopenvino: Vehicle Detection FP16 - CPUopenvino: Vehicle Detection FP16 - CPUwebp: Quality 100, Highest Compressiongravitymark: 1920 x 1200 - Vulkangravitymark: 1920 x 1080 - Vulkanopenvino: Face Detection FP16-INT8 - CPUopenvino: Face Detection FP16 - CPUmnn: SqueezeNetV1.0blender: Classroom - CPU-Onlywebp2: Quality 100, Compression Effort 5compress-7zip: Compression Ratingrocksdb: Rand Fillopenvino: Weld Porosity Detection FP16 - CPUcouchdb: 300 - 3000 - 30openvino: Face Detection FP16 - CPUopenvino: Weld Porosity Detection FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUsrsran: 4G PHY_DL_Test 100 PRB MIMO 256-QAMopenvino: Person Detection FP16 - CPUrocksdb: Update Randopenvino: Face Detection FP16-INT8 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUbuild-nodejs: Time To Compileencode-flac: WAV To FLACopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUbrl-cad: VGR Performance Metricrocksdb: Rand Readmnn: resnet-v2-50srsran: 4G PHY_DL_Test 100 PRB SISO 64-QAMcompress-7zip: Decompression Ratingopenvino: Person Detection FP16 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUblender: Pabellon Barcelona - CPU-Onlyopenvino: Vehicle Detection FP16-INT8 - CPUbuild-php: Time To Compileblender: Fishy Cat - CPU-Onlyrocksdb: Seq Fillgravitymark: 3840 x 2160 - Vulkanbuild-wasmer: Time To Compileblender: Barbershop - CPU-Onlyopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUblender: BMW27 - CPU-Onlyrocksdb: Rand Fill Syncbuild-python: Defaultopenvino: Weld Porosity Detection FP16-INT8 - CPUopenvino: Age Gender Recognition Retail 0013 FP16 - CPUwebp2: Quality 100, Lossless Compressionwebp2: Quality 95, Compression Effort 7webp2: Quality 75, Compression Effort 7webp: Quality 100, Lossless, Highest CompressionABC117.715600000036.836158.667.8771.792.088152.10.55173.7185.81.721.7813.26163.5175.2478.919460592244.1953021.8362359942212.221156.80.92620.517.6988.646128.5041.655227.913174.5533.5272.414.684.27182.2183.65.742.344.009356.173.75832671082843276.05411.4961704.7814.4828.06142.54174.21.78579536693.05448.438.91519.59814.62214502.791543425583111216.294479741672235.79.93428.52402.7552.708167.16138283411257.2711337.398387.1214.24135.131866516.592561.570.950.010.060.130.67102.41604000003.16.97715469.70474.81.7912.092149.10.54176.7182.51.691.7513.04162.3177.7476.519776702280.41521.721.8372324387214.838158.80.93920.627.6789.78126.8961.649230.558174.3528271.1214.754.23182.2183.45.792.334.021359.053.72838971076162274.23414.4651711.214.5827.97143.01173.61.79579961690.81448.218.92521.90214.60714566.071538965605749816.229480.7744252235.169.93429.02402.6752.827167.121381278112.257.371339.698388.1414.23135.161865316.601561.870.950.010.060.130.6710216620000037.045156.169.82275.11.8392.132151.90.55173.51831.691.7713.13165174.8484.319652132255.65526.322.1782347519215.289156.60.93620.787.7788.968127.8531.669228.372176.3528.3269.6114.834.26183.91855.762.354.043356.833.73832761074952276.21411.8441699.3814.4827.87143.49173.11.78582445693.98450.238.88520.92614.67114544.51536805587926716.291480.6743092242.729.96427.73401.5752.804166.831380124112.257.2951338.388380.3714.23135.071865516.598561.760.950.010.060.130.67OpenBenchmarking.org

GravityMark

OpenBenchmarking.orgFrames Per Second, More Is BetterGravityMark 1.70Resolution: 3840 x 2160 - Renderer: OpenGLABC306090120150117.7102.4102.0

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSamples / Second, More Is BettersrsRAN 22.04.1Test: OFDM_TestCBA40M80M120M160M200M1662000001604000001560000001. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

Natron

Natron is an open-source, cross-platform compositing software for visual effects (VFX) and motion graphics. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterNatron 2.4.3Input: SpaceshipBCA0.69751.3952.09252.793.48753.13.03.0

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: nasnetABC2468106.8366.9777.045MIN: 6.77 / MAX: 7.56MIN: 6.91 / MAX: 7.76MIN: 6.97 / MAX: 8.191. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 5G PHY_DL_NR Test 52 PRB SISO 64-QAMACB4080120160200158.6156.1154.01. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

Apache CouchDB

This is a bulk insertion benchmark of Apache CouchDB. CouchDB is a document-oriented NoSQL database implemented in Erlang. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterApache CouchDB 3.2.2Bulk Size: 100 - Inserts: 1000 - Rounds: 30ABC163248648067.8069.7069.821. (CXX) g++ options: -std=c++17 -lmozjs-78 -lm -lei -fPIC -MMD

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 5G PHY_DL_NR Test 52 PRB SISO 64-QAMACB2040608010077.075.174.81. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: MobileNetV2_224ABC0.41380.82761.24141.65522.0691.7901.7911.839MIN: 1.77 / MAX: 2.48MIN: 1.77 / MAX: 2.47MIN: 1.81 / MAX: 2.431. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: squeezenetv1.1ABC0.47970.95941.43911.91882.39852.0882.0922.132MIN: 2.05 / MAX: 9.9MIN: 2.05 / MAX: 9.18MIN: 2.1 / MAX: 2.691. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

GravityMark

OpenBenchmarking.orgFrames Per Second, More Is BetterGravityMark 1.70Resolution: 2560 x 1440 - Renderer: OpenGLACB306090120150152.1151.9149.1

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUBAC0.12380.24760.37140.49520.6190.540.550.55MIN: 0.35 / MAX: 8.76MIN: 0.36 / MAX: 2.11MIN: 0.35 / MAX: 1.921. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

GravityMark

OpenBenchmarking.orgFrames Per Second, More Is BetterGravityMark 1.70Resolution: 1920 x 1200 - Renderer: OpenGLBAC4080120160200176.7173.7173.5

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB SISO 256-QAMACB4080120160200185.8183.0182.51. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

WebP Image Encode

This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP Image Encode 1.2.4Encode Settings: Quality 100, LosslessACB0.3870.7741.1611.5481.9351.721.691.691. (CC) gcc options: -fvisibility=hidden -O2 -lm

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Detection FP32 - Device: CPUACB0.40050.8011.20151.6022.00251.781.771.751. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

WebP Image Encode

This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP Image Encode 1.2.4Encode Settings: Quality 100ACB369121513.2613.1313.041. (CC) gcc options: -fvisibility=hidden -O2 -lm

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB MIMO 64-QAMCAB4080120160200165.0163.5162.31. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

GravityMark

OpenBenchmarking.orgFrames Per Second, More Is BetterGravityMark 1.70Resolution: 1920 x 1080 - Renderer: OpenGLBAC4080120160200177.7175.2174.8

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB MIMO 64-QAMCAB100200300400500484.3478.9476.51. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

Facebook RocksDB

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Read Random Write RandomBCA400K800K1200K1600K2000K1977670196521319460591. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Detection FP32 - Device: CPUACB50010001500200025002244.192255.652280.41MIN: 2049.73 / MAX: 2387.93MIN: 2007.08 / MAX: 2440.95MIN: 2123.83 / MAX: 2416.941. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB SISO 256-QAMACB110220330440550530.0526.3521.71. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: inception-v3ABC51015202521.8421.8422.18MIN: 21.69 / MAX: 22.56MIN: 21.68 / MAX: 30.68MIN: 21.99 / MAX: 30.231. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

Facebook RocksDB

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Read While WritingACB500K1000K1500K2000K2500K2359942234751923243871. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Apache CouchDB

This is a bulk insertion benchmark of Apache CouchDB. CouchDB is a document-oriented NoSQL database implemented in Erlang. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterApache CouchDB 3.2.2Bulk Size: 100 - Inserts: 3000 - Rounds: 30ABC50100150200250212.22214.84215.291. (CXX) g++ options: -std=c++17 -lmozjs-78 -lm -lei -fPIC -MMD

GravityMark

OpenBenchmarking.orgFrames Per Second, More Is BetterGravityMark 1.70Resolution: 2560 x 1440 - Renderer: VulkanBAC4080120160200158.8156.8156.6

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: mobilenetV3ACB0.21130.42260.63390.84521.05650.9260.9360.939MIN: 0.91 / MAX: 1.32MIN: 0.92 / MAX: 1.18MIN: 0.93 / MAX: 1.111. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

WebP Image Encode

This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP Image Encode 1.2.4Encode Settings: DefaultCBA51015202520.7820.6220.511. (CC) gcc options: -fvisibility=hidden -O2 -lm

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP2 Image Encode 20220823Encode Settings: DefaultCAB2468107.777.697.671. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

Timed Erlang/OTP Compilation

This test times how long it takes to compile Erlang/OTP. Erlang is a programming language and run-time for massively scalable soft real-time systems with high availability requirements. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Erlang/OTP Compilation 25.0Time To CompileACB2040608010088.6588.9789.78

Apache CouchDB

This is a bulk insertion benchmark of Apache CouchDB. CouchDB is a document-oriented NoSQL database implemented in Erlang. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterApache CouchDB 3.2.2Bulk Size: 300 - Inserts: 1000 - Rounds: 30BCA306090120150126.90127.85128.501. (CXX) g++ options: -std=c++17 -lmozjs-78 -lm -lei -fPIC -MMD

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: mobilenet-v1-1.0BAC0.37550.7511.12651.5021.87751.6491.6551.669MIN: 1.63 / MAX: 2.08MIN: 1.63 / MAX: 2.47MIN: 1.65 / MAX: 2.171. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

Timed CPython Compilation

This test times how long it takes to build the reference Python implementation, CPython, with optimizations and LTO enabled for a release build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed CPython Compilation 3.10.6Build Configuration: Released Build, PGO + LTO OptimizedACB50100150200250227.91228.37230.56

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB SISO 64-QAMCAB4080120160200176.3174.5174.31. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB MIMO 256-QAMACB120240360480600533.5528.3528.01. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16 - Device: CPUABC60120180240300272.40271.12269.611. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16 - Device: CPUABC4812162014.6814.7514.83MIN: 7.85 / MAX: 24.43MIN: 7.83 / MAX: 23.29MIN: 11.21 / MAX: 22.851. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

WebP Image Encode

This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP Image Encode 1.2.4Encode Settings: Quality 100, Highest CompressionACB0.96081.92162.88243.84324.8044.274.264.231. (CC) gcc options: -fvisibility=hidden -O2 -lm

GravityMark

OpenBenchmarking.orgFrames Per Second, More Is BetterGravityMark 1.70Resolution: 1920 x 1200 - Renderer: VulkanCBA4080120160200183.9182.2182.2

OpenBenchmarking.orgFrames Per Second, More Is BetterGravityMark 1.70Resolution: 1920 x 1080 - Renderer: VulkanCAB4080120160200185.0183.6183.4

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Face Detection FP16-INT8 - Device: CPUBCA1.30282.60563.90845.21126.5145.795.765.741. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Face Detection FP16 - Device: CPUCAB0.52881.05761.58642.11522.6442.352.342.331. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: SqueezeNetV1.0ABC0.90971.81942.72913.63884.54854.0094.0214.043MIN: 3.96 / MAX: 11.74MIN: 3.99 / MAX: 4.18MIN: 4.01 / MAX: 4.281. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: Classroom - Compute: CPU-OnlyACB80160240320400356.17356.83359.05

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP2 Image Encode 20220823Encode Settings: Quality 100, Compression Effort 5ACB0.84381.68762.53143.37524.2193.753.733.721. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

7-Zip Compression

This is a test of 7-Zip compression/decompression with its integrated benchmark feature. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 22.01Test: Compression RatingBCA20K40K60K80K100K8389783276832671. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

Facebook RocksDB

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Random FillABC200K400K600K800K1000K1082843107616210749521. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16 - Device: CPUCAB60120180240300276.21276.05274.231. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Apache CouchDB

This is a bulk insertion benchmark of Apache CouchDB. CouchDB is a document-oriented NoSQL database implemented in Erlang. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterApache CouchDB 3.2.2Bulk Size: 300 - Inserts: 3000 - Rounds: 30ACB90180270360450411.50411.84414.471. (CXX) g++ options: -std=c++17 -lmozjs-78 -lm -lei -fPIC -MMD

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Face Detection FP16 - Device: CPUCAB4008001200160020001699.381704.781711.20MIN: 1643.49 / MAX: 1718.37MIN: 1644.25 / MAX: 1725.99MIN: 1648.89 / MAX: 1728.061. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16 - Device: CPUACB4812162014.4814.4814.58MIN: 12.1 / MAX: 24.07MIN: 12.46 / MAX: 23.76MIN: 7.78 / MAX: 22.321. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Machine Translation EN To DE FP16 - Device: CPUABC71421283528.0627.9727.871. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Machine Translation EN To DE FP16 - Device: CPUABC306090120150142.54143.01143.49MIN: 122.84 / MAX: 151.73MIN: 119.9 / MAX: 151.87MIN: 116.71 / MAX: 153.51. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgUE Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB MIMO 256-QAMABC4080120160200174.2173.6173.11. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Detection FP16 - Device: CPUBCA0.40280.80561.20841.61122.0141.791.781.781. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Facebook RocksDB

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Update RandomCBA120K240K360K480K600K5824455799615795361. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Face Detection FP16-INT8 - Device: CPUBAC150300450600750690.81693.05693.98MIN: 678.51 / MAX: 702.72MIN: 678.32 / MAX: 702.57MIN: 679.36 / MAX: 702.471. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUCAB100200300400500450.23448.43448.211. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Vehicle Bike Detection FP16 - Device: CPUCAB2468108.888.918.92MIN: 7.04 / MAX: 16.76MIN: 7.33 / MAX: 13.57MIN: 7.05 / MAX: 12.891. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Timed Node.js Compilation

This test profile times how long it takes to build/compile Node.js itself from source. Node.js is a JavaScript run-time built from the Chrome V8 JavaScript engine while itself is written in C/C++. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Node.js Compilation 18.8Time To CompileACB110220330440550519.60520.93521.90

FLAC Audio Encoding

OpenBenchmarking.orgSeconds, Fewer Is BetterFLAC Audio Encoding 1.4WAV To FLACBAC4812162014.6114.6214.671. (CXX) g++ options: -O3 -fvisibility=hidden -logg -lm

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUBCA3K6K9K12K15K14566.0714544.5014502.791. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

BRL-CAD

BRL-CAD is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgVGR Performance Metric, More Is BetterBRL-CAD 7.32.6VGR Performance MetricABC30K60K90K120K150K1543421538961536801. (CXX) g++ options: -std=c++11 -pipe -fvisibility=hidden -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -pedantic -ldl -lm

Facebook RocksDB

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Random ReadBCA12M24M36M48M60M5605749855879267558311121. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Mobile Neural Network

MNN is the Mobile Neural Network as a highly efficient, lightweight deep learning framework developed by Alibaba. This MNN test profile is building the OpenMP / CPU threaded version for processor benchmarking and not any GPU-accelerated test. MNN does allow making use of AVX-512 extensions. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterMobile Neural Network 2.1Model: resnet-v2-50BCA4812162016.2316.2916.29MIN: 16.02 / MAX: 25.1MIN: 16.15 / MAX: 25.1MIN: 16.1 / MAX: 17.641. (CXX) g++ options: -std=c++11 -O3 -fvisibility=hidden -fomit-frame-pointer -fstrict-aliasing -ffunction-sections -fdata-sections -ffast-math -fno-rtti -fno-exceptions -rdynamic -pthread -ldl

srsRAN

srsRAN is an open-source LTE/5G software radio suite created by Software Radio Systems (SRS). The srsRAN radio suite was formerly known as srsLTE and can be used for building your own software-defined radio (SDR) 4G/5G mobile network. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgeNb Mb/s, More Is BettersrsRAN 22.04.1Test: 4G PHY_DL_Test 100 PRB SISO 64-QAMBCA100200300400500480.7480.6479.01. (CXX) g++ options: -std=c++14 -fno-strict-aliasing -march=native -mfpmath=sse -mavx2 -fvisibility=hidden -O3 -fno-trapping-math -fno-math-errno -ldl -lpthread -lm

7-Zip Compression

This is a test of 7-Zip compression/decompression with its integrated benchmark feature. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMIPS, More Is Better7-Zip Compression 22.01Test: Decompression RatingBCA16K32K48K64K80K7442574309741671. (CXX) g++ options: -lpthread -ldl -O2 -fPIC

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Person Detection FP16 - Device: CPUBAC50010001500200025002235.162235.702242.72MIN: 2017.96 / MAX: 2392.83MIN: 2063.65 / MAX: 2382.7MIN: 2070.83 / MAX: 2391.141. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16-INT8 - Device: CPUABC36912159.939.939.96MIN: 6.68 / MAX: 17.98MIN: 5.6 / MAX: 16.32MIN: 7.83 / MAX: 13.041. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: Pabellon Barcelona - Compute: CPU-OnlyCAB90180270360450427.73428.52429.02

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Vehicle Detection FP16-INT8 - Device: CPUABC90180270360450402.75402.67401.571. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Timed PHP Compilation

This test times how long it takes to build PHP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed PHP Compilation 8.1.9Time To CompileACB122436486052.7152.8052.83

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: Fishy Cat - Compute: CPU-OnlyCBA4080120160200166.83167.12167.16

Facebook RocksDB

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Sequential FillABC300K600K900K1200K1500K1382834138127813801241. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

GravityMark

OpenBenchmarking.orgFrames Per Second, More Is BetterGravityMark 1.70Resolution: 3840 x 2160 - Renderer: VulkanCBA306090120150112.2112.2112.0

Timed Wasmer Compilation

This test times how long it takes to compile Wasmer. Wasmer is written in the Rust programming language and is a WebAssembly runtime implementation that supports WASI and EmScripten. This test profile builds Wasmer with the Cranelift and Singlepast compiler features enabled. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Wasmer Compilation 2.3Time To CompileACB132639526557.2757.3057.371. (CC) gcc options: -m64 -ldl -lgcc_s -lutil -lrt -lpthread -lm -lc -pie -nodefaultlibs

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: Barbershop - Compute: CPU-OnlyACB300600900120015001337.391338.381339.69

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16 - Device: CPUBAC2K4K6K8K10K8388.148387.128380.371. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16-INT8 - Device: CPUBCA4812162014.2314.2314.24MIN: 7.42 / MAX: 22.21MIN: 7.02 / MAX: 17.84MIN: 8.88 / MAX: 22.511. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.3Blend File: BMW27 - Compute: CPU-OnlyCAB306090120150135.07135.13135.16

Facebook RocksDB

OpenBenchmarking.orgOp/s, More Is BetterFacebook RocksDB 7.5.3Test: Random Fill SyncACB4K8K12K16K20K1866518655186531. (CXX) g++ options: -O3 -march=native -pthread -fno-builtin-memcmp -fno-rtti -lpthread

Timed CPython Compilation

This test times how long it takes to build the reference Python implementation, CPython, with optimizations and LTO enabled for a release build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed CPython Compilation 3.10.6Build Configuration: DefaultACB4812162016.5916.6016.60

OpenVINO

This is a test of the Intel OpenVINO, a toolkit around neural networks, using its built-in benchmarking support and analyzing the throughput and latency for various models. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2022.2.devModel: Weld Porosity Detection FP16-INT8 - Device: CPUBCA120240360480600561.87561.76561.571. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2022.2.devModel: Age Gender Recognition Retail 0013 FP16 - Device: CPUABC0.21380.42760.64140.85521.0690.950.950.95MIN: 0.6 / MAX: 9.81MIN: 0.6 / MAX: 9.6MIN: 0.58 / MAX: 9.831. (CXX) g++ options: -fPIC -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -flto -shared

WebP2 Image Encode

This is a test of Google's libwebp2 library with the WebP2 image encode utility and using a sample 6000x4000 pixel JPEG image as the input, similar to the WebP/libwebp test profile. WebP2 is currently experimental and under heavy development as ultimately the successor to WebP. WebP2 supports 10-bit HDR, more efficienct lossy compression, improved lossless compression, animation support, and full multi-threading support compared to WebP. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP2 Image Encode 20220823Encode Settings: Quality 100, Lossless CompressionCBA0.00230.00460.00690.00920.01150.010.010.011. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

OpenBenchmarking.orgMP/s, More Is BetterWebP2 Image Encode 20220823Encode Settings: Quality 95, Compression Effort 7CBA0.01350.0270.04050.0540.06750.060.060.061. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

OpenBenchmarking.orgMP/s, More Is BetterWebP2 Image Encode 20220823Encode Settings: Quality 75, Compression Effort 7CBA0.02930.05860.08790.11720.14650.130.130.131. (CXX) g++ options: -msse4.2 -fno-rtti -O3 -ldl

WebP Image Encode

This is a test of Google's libwebp with the cwebp image encode utility and using a sample 6000x4000 pixel JPEG image as the input. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMP/s, More Is BetterWebP Image Encode 1.2.4Encode Settings: Quality 100, Lossless, Highest CompressionCBA0.15080.30160.45240.60320.7540.670.670.671. (CC) gcc options: -fvisibility=hidden -O2 -lm

88 Results Shown

GravityMark
srsRAN
Natron
Mobile Neural Network
srsRAN
Apache CouchDB
srsRAN
Mobile Neural Network:
  MobileNetV2_224
  squeezenetv1.1
GravityMark
OpenVINO
GravityMark
srsRAN
WebP Image Encode
OpenVINO
WebP Image Encode
srsRAN
GravityMark
srsRAN
Facebook RocksDB
OpenVINO
srsRAN
Mobile Neural Network
Facebook RocksDB
Apache CouchDB
GravityMark
Mobile Neural Network
WebP Image Encode
WebP2 Image Encode
Timed Erlang/OTP Compilation
Apache CouchDB
Mobile Neural Network
Timed CPython Compilation
srsRAN:
  4G PHY_DL_Test 100 PRB SISO 64-QAM
  4G PHY_DL_Test 100 PRB MIMO 256-QAM
OpenVINO:
  Vehicle Detection FP16 - CPU:
    FPS
    ms
WebP Image Encode
GravityMark:
  1920 x 1200 - Vulkan
  1920 x 1080 - Vulkan
OpenVINO:
  Face Detection FP16-INT8 - CPU
  Face Detection FP16 - CPU
Mobile Neural Network
Blender
WebP2 Image Encode
7-Zip Compression
Facebook RocksDB
OpenVINO
Apache CouchDB
OpenVINO:
  Face Detection FP16 - CPU
  Weld Porosity Detection FP16 - CPU
  Machine Translation EN To DE FP16 - CPU
  Machine Translation EN To DE FP16 - CPU
srsRAN
OpenVINO
Facebook RocksDB
OpenVINO:
  Face Detection FP16-INT8 - CPU
  Person Vehicle Bike Detection FP16 - CPU
  Person Vehicle Bike Detection FP16 - CPU
Timed Node.js Compilation
FLAC Audio Encoding
OpenVINO
BRL-CAD
Facebook RocksDB
Mobile Neural Network
srsRAN
7-Zip Compression
OpenVINO:
  Person Detection FP16 - CPU
  Vehicle Detection FP16-INT8 - CPU
Blender
OpenVINO
Timed PHP Compilation
Blender
Facebook RocksDB
GravityMark
Timed Wasmer Compilation
Blender
OpenVINO:
  Age Gender Recognition Retail 0013 FP16 - CPU
  Weld Porosity Detection FP16-INT8 - CPU
Blender
Facebook RocksDB
Timed CPython Compilation
OpenVINO:
  Weld Porosity Detection FP16-INT8 - CPU
  Age Gender Recognition Retail 0013 FP16 - CPU
WebP2 Image Encode:
  Quality 100, Lossless Compression
  Quality 95, Compression Effort 7
  Quality 75, Compression Effort 7
WebP Image Encode