extra tests2

AMD EPYC 9334 32-Core testing with a Supermicro H13SSW (1.1 BIOS) and astdrmfb on AlmaLinux 9.2 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2310249-NE-EXTRATEST98
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts
Allow Limiting Results To Certain Suite(s)

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Additional Graphs

Show Perf Per Core/Thread Calculation Graphs Where Applicable
Show Perf Per Clock Calculation Graphs Where Applicable

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs
Condense Test Profiles With Multiple Version Results Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Toggle/Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
a
October 07 2023
  5 Hours, 38 Minutes
b
October 07 2023
  5 Hours, 6 Minutes
c
October 07 2023
  4 Hours, 48 Minutes
d
October 20 2023
  5 Hours, 58 Minutes
e
October 20 2023
  6 Hours, 40 Minutes
f
October 21 2023
  6 Hours, 19 Minutes
g
October 22 2023
  5 Hours, 59 Minutes
h
October 24 2023
  6 Hours, 2 Minutes
i
October 24 2023
  6 Hours, 8 Minutes
j
October 24 2023
  5 Hours, 52 Minutes
Invert Behavior (Only Show Selected Data)
  5 Hours, 51 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


extra tests2ProcessorMotherboardMemoryDiskGraphicsMonitorOSKernelCompilerFile-SystemScreen Resolutionabcdefghij2 x AMD EPYC 9254 24-Core @ 2.90GHz (48 Cores / 96 Threads)Supermicro H13DSH (1.5 BIOS)24 x 32 GB DDR5-4800MT/s Samsung M321R4GA3BB6-CQKET2 x 1920GB SAMSUNG MZQL21T9HCJR-00A07astdrmfbAlmaLinux 9.25.14.0-284.25.1.el9_2.x86_64 (x86_64)GCC 11.3.1 20221121ext41024x768AMD EPYC 9124 16-Core @ 3.00GHz (16 Cores / 32 Threads)Supermicro H13SSW (1.1 BIOS)12 x 64 GB DDR5-4800MT/s HMCG94MEBRA123NAMD EPYC 9334 32-Core @ 2.70GHz (32 Cores / 64 Threads)DELL E207WFP1680x1050OpenBenchmarking.orgKernel Details- Transparent Huge Pages: alwaysCompiler Details- --build=x86_64-redhat-linux --disable-libunwind-exceptions --enable-__cxa_atexit --enable-bootstrap --enable-cet --enable-checking=release --enable-gnu-indirect-function --enable-gnu-unique-object --enable-host-bind-now --enable-host-pie --enable-initfini-array --enable-languages=c,c++,fortran,lto --enable-link-serialization=1 --enable-multilib --enable-offload-targets=nvptx-none --enable-plugin --enable-shared --enable-threads=posix --mandir=/usr/share/man --with-arch_32=x86-64 --with-arch_64=x86-64-v2 --with-build-config=bootstrap-lto --with-gcc-major-version-only --with-linker-hash-style=gnu --with-tune=generic --without-cuda-driver --without-isl Processor Details- a: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa10113e- b: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa10113e- c: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa10113e- d: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111- e: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111- f: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111- g: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111- h: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111- i: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111- j: Scaling Governor: acpi-cpufreq performance (Boost: Enabled) - CPU Microcode: 0xa101111Java Details- OpenJDK Runtime Environment (Red_Hat-11.0.20.0.8-1) (build 11.0.20+8-LTS)Python Details- Python 3.9.16Security Details- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional IBRS_FW STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected

abcdefghijResult OverviewPhoronix Test Suite100%143%187%230%273%BlenderBRL-CADIntel Open Image DenoiseEmbreeSPECFEM3DOSPRayTiDB Community ServerTimed Linux Kernel CompilationRemhosOpenVINONeural Magic DeepSparseApache CassandraApache HadoopLiquid-DSPSVT-AV1nekRS

extra tests2hadoop: File Status - 50 - 1000000hadoop: Open - 100 - 1000000hadoop: Open - 50 - 1000000deepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Streamopenvino: Handwritten English Recognition FP16-INT8 - CPUdeepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamdeepsparse: ResNet-50, Baseline - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Streamdeepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Streamonednn: IP Shapes 1D - u8s8f32 - CPUdeepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamopenvino: Handwritten English Recognition FP16 - CPUopenvino: Face Detection FP16 - CPUopenvino: Weld Porosity Detection FP16-INT8 - CPUospray: particle_volume/ao/real_timeospray: particle_volume/scivis/real_timeopenvino: Weld Porosity Detection FP16 - CPUdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Streamliquid-dsp: 96 - 256 - 32deepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Streamopenvino: Face Detection FP16-INT8 - CPUdeepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Streamblender: Pabellon Barcelona - CPU-Onlyopenvino: Face Detection Retail FP16-INT8 - CPUblender: Classroom - CPU-Onlyblender: BMW27 - CPU-Onlyopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUblender: Fishy Cat - CPU-Onlyopenvino: Age Gender Recognition Retail 0013 FP16 - CPUopenvino: Person Detection FP32 - CPUspecfem3d: Layered Halfspaceopenvino: Person Detection FP16 - CPUblender: Barbershop - CPU-Onlydeepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Streambrl-cad: VGR Performance Metricembree: Pathtracer - Crownopenvino: Machine Translation EN To DE FP16 - CPUopenvino: Vehicle Detection FP16 - CPUoidn: RTLightmap.hdr.4096x4096 - CPU-Onlyoidn: RT.ldr_alb_nrm.3840x2160 - CPU-Onlyospray: gravity_spheres_volume/dim_512/scivis/real_timeoidn: RT.hdr_alb_nrm.3840x2160 - CPU-Onlyospray: gravity_spheres_volume/dim_512/ao/real_timeembree: Pathtracer ISPC - Crownspecfem3d: Mount St. Helensliquid-dsp: 96 - 256 - 512ospray: gravity_spheres_volume/dim_512/pathtracer/real_timespecfem3d: Homogeneous Halfspaceopenvino: Vehicle Detection FP16-INT8 - CPUembree: Pathtracer - Asian Dragonembree: Pathtracer - Asian Dragon Objopenvino: Person Vehicle Bike Detection FP16 - CPUembree: Pathtracer ISPC - Asian Dragon Objembree: Pathtracer ISPC - Asian Dragonhadoop: File Status - 50 - 100000specfem3d: Water-layered Halfspaceopenvino: Face Detection Retail FP16 - CPUopenvino: Road Segmentation ADAS FP16-INT8 - CPUspecfem3d: Tomographic Modelliquid-dsp: 96 - 256 - 57openvino: Road Segmentation ADAS FP16 - CPUliquid-dsp: 64 - 256 - 512onednn: IP Shapes 1D - bf16bf16bf16 - CPUliquid-dsp: 64 - 256 - 32openvino: Face Detection FP16 - CPUonednn: Convolution Batch Shapes Auto - u8s8f32 - CPUembree: Pathtracer ISPC - Crownonednn: IP Shapes 3D - u8s8f32 - CPUbuild-linux-kernel: defconfigembree: Pathtracer - Crownhadoop: File Status - 100 - 1000000openvino: Face Detection FP16-INT8 - CPUembree: Pathtracer ISPC - Asian Dragon Objembree: Pathtracer - Asian Dragon Objembree: Pathtracer - Asian Dragonopenvino: Person Detection FP16 - CPUembree: Pathtracer ISPC - Asian Dragononednn: Deconvolution Batch shapes_3d - f32 - CPUopenvino: Person Detection FP32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUoidn: RT.ldr_alb_nrm.3840x2160 - CPU-Onlyoidn: RTLightmap.hdr.4096x4096 - CPU-Onlyopenvkl: vklBenchmarkCPU Scalaroidn: RT.hdr_alb_nrm.3840x2160 - CPU-Onlyopenvkl: vklBenchmarkCPU ISPCremhos: Sample Remap Examplehadoop: Open - 100 - 100000onednn: Convolution Batch Shapes Auto - f32 - CPUhadoop: File Status - 100 - 100000onednn: Deconvolution Batch shapes_3d - bf16bf16bf16 - CPUliquid-dsp: 64 - 256 - 57onednn: Convolution Batch Shapes Auto - bf16bf16bf16 - CPUopenvino: Vehicle Detection FP16 - CPUopenvino: Machine Translation EN To DE FP16 - CPUtidb: oltp_read_write - 64tidb: oltp_read_write - 128hadoop: Create - 100 - 100000onednn: Deconvolution Batch shapes_1d - bf16bf16bf16 - CPUopenvino: Vehicle Detection FP16-INT8 - CPUhadoop: Create - 100 - 1000000onednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Training - f32 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUcassandra: Writesonednn: IP Shapes 3D - bf16bf16bf16 - CPUopenvino: Face Detection Retail FP16 - CPUtidb: oltp_read_write - 32hadoop: Open - 50 - 100000openvino: Road Segmentation ADAS FP16-INT8 - CPUopenvino: Person Vehicle Bike Detection FP16 - CPUtidb: oltp_update_non_index - 128liquid-dsp: 32 - 256 - 512openvino: Road Segmentation ADAS FP16 - CPUtidb: oltp_point_select - 64onednn: IP Shapes 1D - f32 - CPUtidb: oltp_point_select - 128onednn: Recurrent Neural Network Inference - u8s8f32 - CPUtidb: oltp_update_non_index - 64tidb: oltp_update_index - 128onednn: Recurrent Neural Network Inference - f32 - CPUonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUkripke: easywave: e2Asean Grid + BengkuluSept2007 Source - 1200tidb: oltp_update_index - 64onednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUsvt-av1: Preset 8 - Bosphorus 4Khadoop: Create - 50 - 100000easywave: e2Asean Grid + BengkuluSept2007 Source - 2400hadoop: Delete - 100 - 100000svt-av1: Preset 12 - Bosphorus 4Ktidb: oltp_point_select - 32svt-av1: Preset 13 - Bosphorus 1080ptidb: oltp_update_index - 32tidb: oltp_point_select - 1svt-av1: Preset 13 - Bosphorus 4Kospray: particle_volume/pathtracer/real_timetidb: oltp_update_non_index - 1hadoop: Create - 50 - 1000000hadoop: Delete - 50 - 100000tidb: oltp_read_write - 1tidb: oltp_update_index - 1svt-av1: Preset 12 - Bosphorus 1080ponednn: IP Shapes 3D - f32 - CPUtidb: oltp_update_non_index - 32liquid-dsp: 32 - 256 - 57tidb: oltp_update_index - 16tidb: oltp_point_select - 16easywave: e2Asean Grid + BengkuluSept2007 Source - 240hadoop: Delete - 100 - 1000000tidb: oltp_read_write - 16tidb: oltp_update_non_index - 16hadoop: Rename - 100 - 1000000hadoop: Rename - 50 - 100000svt-av1: Preset 8 - Bosphorus 1080psvt-av1: Preset 4 - Bosphorus 4Khadoop: Delete - 50 - 1000000hadoop: Rename - 100 - 100000liquid-dsp: 2 - 256 - 512hadoop: Rename - 50 - 1000000nekrs: Kershawnekrs: TurboPipe Periodicsvt-av1: Preset 4 - Bosphorus 1080pliquid-dsp: 1 - 256 - 512liquid-dsp: 1 - 256 - 57deepsparse: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Asynchronous Multi-Streamliquid-dsp: 2 - 256 - 32liquid-dsp: 32 - 256 - 32liquid-dsp: 8 - 256 - 512liquid-dsp: 2 - 256 - 57liquid-dsp: 4 - 256 - 512liquid-dsp: 1 - 256 - 32liquid-dsp: 16 - 256 - 512liquid-dsp: 8 - 256 - 32liquid-dsp: 16 - 256 - 32liquid-dsp: 4 - 256 - 32openvino: Age Gender Recognition Retail 0013 FP16 - CPUliquid-dsp: 16 - 256 - 57openvino: Handwritten English Recognition FP16-INT8 - CPUdeepsparse: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Asynchronous Multi-Streamliquid-dsp: 8 - 256 - 57openvino: Face Detection Retail FP16-INT8 - CPUdeepsparse: ResNet-50, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Streamopenvino: Weld Porosity Detection FP16-INT8 - CPUliquid-dsp: 4 - 256 - 57openvino: Handwritten English Recognition FP16 - CPUdeepsparse: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Asynchronous Multi-Streamopenvino: Weld Porosity Detection FP16 - CPUonednn: Deconvolution Batch shapes_1d - f32 - CPUdeepsparse: NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Streamdeepsparse: BERT-Large, NLP Question Answering - Asynchronous Multi-Streamopenvino: Age Gender Recognition Retail 0013 FP16-INT8 - CPUdeepsparse: BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Streamdeepsparse: NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Streamdeepsparse: NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Streamdeepsparse: ResNet-50, Baseline - Asynchronous Multi-Streamdeepsparse: CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Streamdeepsparse: CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Streamquantlib: Multi-Threadedquantlib: Single-Threadeddeepsparse: CV Detection, YOLOv5s COCO - Asynchronous Multi-Streamopenradioss: Bumper Beamabcdefghij217391321533211261265137.01141244.6949.325839.4951218.146439.4391489.1203485.6725215.6383718.9189158.924322.25051560.0330.415776.9415.98615.95282945.26201.3925300580000068.598856.011417.070680.549837.5866.4226.2120606.3833.2286884.64283.9726.885983804282.55254.88672.463577216254.9017317.222033.170.861.8413.87391.8314.236956.087111.02476577571164000016.346815.105117732873.2460.144953.57332454.0956.485367.337852910126.9850209085882.91842.9112.3129466522559800000748.446225600002207700000393.627.3541886792213.9442.4442.2416.34642016851546419944000005.8937.87909085757407334.17461452480952.035897446082914.234.885110542581000016.02127567159242412812708790.8114364987566163.459104627510.361183614331163.013215.0961328536659107525401212422.994287351192100000125589011438331180957307870522141.2195.2039893275529279010007323911106900000676771000012.477139090005940100035.62877181000118350000010987000011749000052911000394990002160800003075400005942300001538500000.5469974000038.516.90743694300004.864.6508347.66128.2819622000030.72118.750916.26150.5897485.71750.3433.341274.3211605.0388605.75849.3653109.802749.0092111.0131194174817382210204085138.83411239.6749.166639.4681219.526339.4539489.4464488.1264215.9254717.9693159.0596321.18291546.0230.445780.4415.978515.98882986.46201.2528299540000068.661756.061403.06580.769849.0766.6426.24120728.2233.1787359.23284.9928.65210863284.22255.3672.373476851755.3925317.282028.010.861.8413.76661.8314.178356.455111.31849570971814000016.436514.460586982880.5859.912453.81352450.2656.690267.195186206929.4607611975836.27854.5112.1005959472571100000750.496109500002212100000393.2327.241161970213.6242.242.0916.79140485845871620019000005.9137.798018389099374254.16444372566612.056152046948414.034.8942962000015.9813080215972839759274642437191.3224128890827166.378106180542.611178174405166.692214.074131252119738012510427.686289141214200000675158671536950180687212973046138.3385.1499731469348277360007167911240300000675736000012.591140210005929600035.642277019000119030000010808000011401000055588000394860002161500003051100006024700001536900000.5469276000038.6617.06693669300004.854.6476347.21898.2719659000031118.948216.02150.6055487.35990.3433.379874.5553605.7307606.669349.1082109.233248.9703110.91522842521858746839955153.66441237.2947.153539.4464218.516239.4183487.0522489.1106215.647716.1404164.6131321.50821551.6330.435802.6515.987215.97782987.33201.5402299980000068.628756.021418.904180.419845.2766.7226.12123484.2833.0386789.8284.3127.490850157282.67254.72671.255476252955.4037317.332029.790.871.8213.83171.8314.139956.807811.3273597771503000016.53514.8082733652881.1459.792953.69272455.5156.932767.503865789527.0602350795840.53849.312.0409178772564900000757.386226300002206800000393.3727.4081893939213.7942.4342.1916.24340322672992720103000005.937.7978469350754.16440012704802.055963040160614.124.885286542440000015.8314996239106265462332490.4174393773475163.055516.906175654471161.495214.1361381522609058024851189431.8951254800000126816540697031373686682777101143.5455.0499014767159282270007463810826700000675417000012.617142250005751900035.682576924000118480000010914000011855000055165000394530002149100003067600006036500001536700000.5467493000038.7516.88633669900004.864.6348347.37288.2419451000030.89118.780116.02145.2562507.47860.3433.463774.5031605.9183605.876549.0173109.584749.2055111.025181818212484392783191599.2079395.6616.164813.069472.45513.1312163.5559162.850271.9189240.55290.65225955.6132108.909532.5910.472013.775.574695.570011039.6171.137106520000024.483320.03508.087224.153540.88182.997244958.0790.0332002.62106.971.614294327107.02670.87257.272829806421.4811124.12797.640.340.725.453290.725.6074722.58526.7364141182862500006.5874535.5716849081175.6724.687222.26141036.9923.873328.364363291162.4417495852564.78370.5727.3309855881120800000344.672829200001.037491059500000761.591.5582422.39220.6039555.17321.8943600601398.5223.353122.351724.84574.7127.74383.3778274.810.8478050.720.341910.7248730.7615291012.133325917161.9137410933000001.3378910.0164.415533459727579713.059916.79712961643.991641.921642.511978661.028753.114697757803521.577.727376000023.21156752.49408129492849.16334224838.516847.3824099450038.105211080.62823666.9885861798.98105708163.18998149604.986176125898161.854151.9051693721341010101479526.2161.25758262731035000000126221.65711261336480185638120882372118.9464.10711101282102246270008392110318900000793457000010.91126830005266500031.05256705400010471000009959400010565000050258000352280001938500002780300005453600001386000000.4968915000040.415.72463633100004.514.996325.87637.9318893000030.02112.249915.363.81576143.7643493.5960.3533.224273.3058606.101606.577349.0863110.111448.8729111.108132092412048192510041599.1543432.3216.139212.943372.657113.1187162.9298163.136171.9146240.23490.6576155.4634109.0938530.9910.472007.535.541075.563531039.8271.2727106510000024.472520511.4098224.13544.18182.5671.4444933.2790.3132032.06107.2470.189028506107.27670.64257.89429612521.4357123.61793.750.340.725.461530.725.620422.569426.7991434462858800006.582735.0301348891174.624.734322.15771028.6423.939328.314138910562.3251468282562.54373.6427.4598213081117800000342.812818300001.144321057500000761.161.5491122.34220.57579455.09321.9909235627398.9123.528322.291124.828274.527.82933.3843674.540.8444340.720.341900.7248730.8452949852.12576134971.9178110954000001.3386110.0664.685389360145588243.06376.8700571643.9716411639.361957981.054253.114673755248621.47.774213827348000023.321186572.56522129904851.6593388124611849.712841.07823624390038.067212710.63397567.7215861799.41598039162.60896907597.011171175976162.051151.50617087089710060432091490525.1731.2804326285103200000012567702501.65411322536784185578436082237119.3074.11411332783822252070008404110264000000793101000010.984123660005282700030.98676884600010466000009700500010548000050380000353150001960400002777800005451400001386200000.4969292000036.9815.62453579900004.514.9859325.74167.9619123000030.1112.057415.363.84421144.1013494.25750.3533.262573.2609607.913606.75549.0094109.965449.0598111.08881795332130378112210011600.5275431.9416.130713.085372.573613.0934162.8976162.929471.9401240.16420.65318255.5428109.09533.7410.482004.765.57325.555811038.4771.044106530000024.523820.01508.2088223.953548.78181.771.9644968.4390.2631951.64106.7670.542255905107.39667.87257.504629560321.5913124.3791.740.340.725.452270.725.6145422.656626.8731684552859200006.5956335.5350730011180.8524.704722.1491041.8723.935428.323770922061.2817691242539.97369.2626.9737573951120500000343.492830300001.001361057100000760.571.5728222.44070.60083455.14821.77461964637399.2423.504222.267624.88774.4327.8263.3795674.870.8506910.720.341910.7248830.7255235602.130624784691.9142210946000001.3418310.0964.315495660310593823.056746.76705371642.351636.761636.441962871.061443.144714157803521.657.674142427339000023.281190922.49714130389849.3443447024830851.494845.30823659100038.015210670.63032567.3935834397.98799404161.84797368585.3685954160.798151.781697699209699332181483521.5181.2065326695102460000012692701051.657110803361258581581633118.4864.1381111987949125199000825019976450000795579000010.736126810005287900031.02926886100010419000009944100010574000049977000352710001945000002763900005450200001385800000.4969334000037.0115.7213504500004.54.9877324.95687.9718988000029.95112.41315.383.81823143.6922494.22110.3533.275173.2163607.8171606.785249.0712110.000149.0714110.8852203666011074206540221602.5221432.216.067213.07372.687913.0596163.2282162.993771.9003239.51760.647755.4264109.2191538.0110.482006.095.575535.565391039.3770.9275106570000024.460720.05509.139224.123533.64183.2972.0145097.9990.6332008.03107.2469.955609165107.04669.09257.280829552221.5847123.41793.90.340.725.477250.725.6227822.774527.6966313712865300006.6008535.3786000211175.5824.819322.19011031.623.879628.479356179862.8109243762557.66372.2627.7464751621118200000341.362817300001.127231056200000759.921.5511822.42150.6123255.17221.83052049180398.1323.709122.255924.961974.7127.913.3815674.580.8434920.720.341910.7248930.754608292.118134878051.9127410993000001.3356410.0664.775530159944589283.054586.79709221641.41637.371631.991970921.045673.124699354644821.477.744169527407000023.421185492.51441837.5953410724574848.032847.41723717570037.950.62910867.8116068097.529102564160.32296840586.74817135161.324151.68117057270610395031951481528.5331.27918103340000012627699231.64811389536088187358576382237118.4814.14311082880386227270008481010500600000796491000011.016122560005285400031.057468678000104710000010017000010480000049556000352360001946700002774100005430500001384600000.4968207000036.9815.68923578100004.524.9787325.50757.9619075000029.72112.477915.373.82381144.1053495.60330.3533.367473.1867607.1628608.716349.0259109.903748.9837110.975942643932372912330463324.8182810.7132.004225.7096145.477825.7916326.9271327.0192144.1181488.06480.892701109.0251216.23841034.3519.823780.810.798810.78321963.9136.9614206830000047.132337.821003.5171119.36653.8699.5438.468931.3549.159615.88196.0740.329363428197.94352.4507.63457250042.8932234.181481.710.651.3810.59131.3710.852445.475715.03269371151970000012.543819.9615193482252.748.118843.50562180.8346.378554.388892592636.4458715834803.65710.6915.9894919481995500000643.85120200002.09882057500000804.580.77854345.190.3024633.94543.57212352941421.845.923443.844648.906880.7754.22031.7338181.580.4400061.380.653631.3792620.4425263161.147495952381.04119168000000.73446110.7868.2795579104180604961.78177.0973611987.837986.958993.5643245750.704553.327548664516122.57.336581639168000024.841.74203200327568.7475119137126564.12568.64935480800026.106313320.42642699.0995872068.573105263230.026138538726.889242866125223.416192.64518617341110040234801666580.1630.92628336041135290000016965872181.28411178247576237948431083056151.4375.0751134047980825648000851219145900000676127000012.23129090005144300031.489672468000117240000010444000010528000052890000371450002075900002926200005858500001458200000.5374142000039.4415.92493783300004.84.8005337.95688.4620093000030.92116.506416.273.72247146.1191495.22370.3532.752673.8802613.0546613.163948.8812109.792248.8889116781.53357.9110.810819305023829952536783327.1891815.1331.958625.6331145.647425.704326.7757327.7314144.0005486.70681.9419109.4649216.21631036.1119.833777.7510.74310.79041964.61136.684206990000047.349737.57999.4619119.426646.9199.3538.5768945.3248.6959654.98197.6639.833589788193.8351.66508.360457045843.2681233.651488.040.651.3810.58481.3710.882945.400615.19244731852014000012.498719.6196372652264.9348.190843.57362224.0846.363354.506386956537.5052638494848.42709.7315.5902378591997000000642.95123100001.243082052600000804.750.7985445.71060.30153533.86943.81162506266425.6846.087243.767448.71882.4954.18771.7349980.880.4403681.380.653631.3792220.3025555561.150128474581.0431218997000000.73509410.7468.4294261104620590321.788767.0572010985.743991.073987.3553181090.6442523.297525458139522.537.1839320000024.871801791.64478198137564.4545034636141566.178563.33435015120026.393305220.42751298.5735899768.236102041227.865138173728.538237736165227.205192.418487246410224934851656591.3190.93179335650139410000016817864711.28811469247501235418521590827149.0515.0791104368237225910000863119269890000676807000012.259133630005584100031.441172491000117290000010478000010970000052129000371410002097200002924300005838700001459200000.5373580000039.2315.98953772500004.814.8015336.76738.4619864000030.87116.890816.273.65907145.8325494.49530.3532.835473.9404612.9837613.640248.7901109.736348.89281171993365.4110.9833194174828910112515643324.6952812.7632.063925.727145.372625.769326.904327.1855144.0515487.22980.880016109.4068216.5081012.9819.843783.6510.693210.76641964.71136.5656207100000047.333937.481002.8113119.046638.2499.2938.5168895.548.8259505.78196.2639.867423777194.82351.38507.364556906643.3141233.881483.250.651.3810.5741.3710.874345.322715.09991232651944000012.510419.7255866632254.7248.095643.41612194.3646.432754.515775188037.8081130344892.27709.7515.8438953171999500000648.275110700001.949412056500000805.380.76831445.39560.30927833.93443.9335558036425.8846.023243.904348.711582.0954.14731.7350181.50.4401561.380.653631.3792220.363703701.155786849321.0433319223000000.73161810.7768.35105802580051.786917.0872754991.123988.589994.6123164850.714973.267425261728422.537.286406639327000024.671805811.75453197738563.5755030436644569.795566.38634901980025.555306380.4302799.3455934768.51997371224.406137618726.50324366228.765192.701731539560234791660584.120.93600135655136980000016972874121.24511069347414235438716189206149.4515.161135728347226378000885509242080000683536000012.175128990005571500031.502472400000116990000010422000011137000049781000371200002094900002924900005856700001457400000.5373712000039.3415.93893816600004.814.8016336.6758.4520012000031.57116.846416.273.68087145.9015494.00760.3532.808573.777614.2303612.520948.8291109.857348.8934117263.33371.6110.9038OpenBenchmarking.org

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: File Status - Threads: 50 - Files: 1000000cehfdibjga500K1000K1500K2000K2500K2842523209244264391795332181818219305021941748194174820366602173913

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Open - Threads: 100 - Files: 1000000bcajhigedf300K600K900K1200K1500K1738221858742153322891013237293829951107420120481912484391303781

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Open - Threads: 50 - Files: 1000000eidgcbafhj300K600K900K1200K1500K25100425367827831965402268399510204081126126122100112330461251564

Neural Magic DeepSparse

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Streamedfgjhiabc110022003300440055001599.151599.211600.531602.523324.703324.823327.195137.015138.835153.66

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Handwritten English Recognition FP16-INT8 - Device: CPUdfgehjicba30060090012001500395.66431.94432.20432.32810.71812.76815.131237.291239.671244.691. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Streamgfedihjcba112233445516.0716.1316.1416.1631.9632.0032.0647.1549.1749.33

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Streamedgfihjcba91827364512.9413.0713.0713.0925.6325.7125.7339.4539.4739.50

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Streamdfegjhiacb5010015020025072.4672.5772.6672.69145.37145.48145.65218.15218.52219.53

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Streamgfedijhcab91827364513.0613.0913.1213.1325.7025.7725.7939.4239.4439.45

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Streamfegdijhcab110220330440550162.90162.93163.23163.56326.78326.90326.93487.05489.12489.45

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Streamdfgehjiabc110220330440550162.85162.93162.99163.14327.02327.19327.73485.67488.13489.11

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Streamgedfijhacb5010015020025071.9071.9171.9271.94144.00144.05144.12215.64215.65215.93

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Streamgfedijhcba160320480640800239.52240.16240.23240.55486.71487.23488.06716.14717.97718.92

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUihjefdg0.43690.87381.31071.74762.18451.9419000.8927010.8800160.6576100.6531820.6522590.647700MIN: 0.87MIN: 0.79MIN: 0.78MIN: 0.57MIN: 0.57MIN: 0.57MIN: 0.571. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Streamgefdhjiabc408012016020055.4355.4655.5455.61109.03109.41109.46158.92159.06164.61

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Streamdfegihjbca70140210280350108.91109.09109.09109.22216.22216.24216.51321.18321.51322.25

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Handwritten English Recognition FP16 - Device: CPUedfgjhibca30060090012001500530.99532.59533.74538.011012.981034.351036.111546.021551.631560.031. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Face Detection FP16 - Device: CPUdefghijacb71421283510.4710.4710.4810.4819.8219.8319.8430.4130.4330.441. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Weld Porosity Detection FP16-INT8 - Device: CPUfgedihjabc120024003600480060002004.762006.092007.532013.773777.753780.803783.655776.945780.445802.651. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: particle_volume/ao/real_timeefdgjihbac481216205.541075.573205.574695.5755310.6932010.7430010.7988015.9785015.9860015.98720

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: particle_volume/scivis/real_timefegdjhiacb481216205.555815.563535.565395.5700110.7664010.7832010.7904015.9528015.9778015.98880

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Weld Porosity Detection FP16 - Device: CPUfgdehijabc60012001800240030001038.471039.371039.611039.821963.901964.611964.712945.262986.462987.331. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Streamgfdejihbac408012016020070.9371.0471.1471.27136.57136.68136.96201.25201.39201.54

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 96 - Buffer Length: 256 - Filter Length: 32edfghijbca600M1200M1800M2400M3000M10651000001065200000106530000010657000002068300000206990000020710000002995400000299980000030058000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

Neural Magic DeepSparse

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Streamgedfhjiacb153045607524.4624.4724.4824.5247.1347.3347.3568.6068.6368.66

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Face Detection FP16-INT8 - Device: CPUefdgjihacb132639526520.0020.0120.0320.0537.4837.5737.8256.0156.0256.061. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Streamdfgeijhbac30060090012001500508.09508.21509.14511.41999.461002.811003.521403.071417.071418.90

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: Pabellon Barcelona - Compute: CPU-Onlydgefihjbac50100150200250224.15224.12224.10223.95119.42119.30119.0480.7680.5480.41

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Face Detection Retail FP16-INT8 - Device: CPUgdefjihacb2K4K6K8K10K3533.643540.883544.183548.786638.246646.916653.869837.589845.279849.071. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: Classroom - Compute: CPU-Onlygdefhijcba4080120160200183.29182.99182.56181.7099.5499.3599.2966.7266.6466.42

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: BMW27 - Compute: CPU-Onlygdfeijhbac163248648072.0172.0071.9671.4438.5738.5138.4026.2426.2026.12

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUedfgjhiabc30K60K90K120K150K44933.2744958.0744968.4345097.9968895.5068931.3568945.32120606.38120728.22123484.281. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: Fishy Cat - Compute: CPU-Onlygefdhjiabc2040608010090.6390.3190.2690.0349.1048.8248.6933.2233.1733.03

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUfdgejhicab20K40K60K80K100K31951.6432002.6232008.0332032.0659505.7859615.8859654.9886789.8086884.6487359.231. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Person Detection FP32 - Device: CPUfdeghjiacb60120180240300106.76106.90107.24107.24196.07196.26197.66283.97284.31284.991. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Layered Halfspacedfeghjibca163248648071.6170.5470.1969.9640.3339.8739.8328.6527.4926.891. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Person Detection FP16 - Device: CPUdgefijhacb60120180240300107.02107.04107.27107.39193.80194.82197.94282.55282.67284.221. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Blender

OpenBenchmarking.orgSeconds, Fewer Is BetterBlender 3.6Blend File: Barbershop - Compute: CPU-Onlydegfhijbac140280420560700670.87670.64669.09667.87352.40351.66351.38255.30254.88254.72

Neural Magic DeepSparse

OpenBenchmarking.orgitems/sec, More Is BetterNeural Magic DeepSparse 1.5Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Streamdgfejhicba150300450600750257.27257.28257.50257.89507.36507.63508.36671.26672.37672.46

BRL-CAD

BRL-CAD is a cross-platform, open-source solid modeling system with built-in benchmark mode. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgVGR Performance Metric, More Is BetterBRL-CAD 7.36VGR Performance Metricgfedjihcba170K340K510K680K850K2955222956032961252980645690665704585725007625297685177721621. (CXX) g++ options: -std=c++14 -pipe -fvisibility=hidden -fno-strict-aliasing -fno-common -fexceptions -ftemplate-depth-128 -m64 -ggdb3 -O3 -fipa-pta -fstrength-reduce -finline-functions -flto -ltcl8.6 -lregex_brl -lz_brl -lnetpbm -ldl -lm -ltk8.6

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer - Model: Crownedgfhijabc122436486021.4421.4821.5821.5942.8943.2743.3154.9055.3955.40MIN: 21.3 / MAX: 21.78MIN: 21.32 / MAX: 21.8MIN: 21.43 / MAX: 21.89MIN: 21.45 / MAX: 21.84MIN: 42.47 / MAX: 43.88MIN: 42.82 / MAX: 44.23MIN: 42.83 / MAX: 44.44MIN: 53.27 / MAX: 57.28MIN: 54.02 / MAX: 57.64MIN: 53.71 / MAX: 58.99

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Machine Translation EN To DE FP16 - Device: CPUgedfijhabc70140210280350123.41123.61124.12124.30233.65233.88234.18317.22317.28317.331. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Vehicle Detection FP16 - Device: CPUfegdhjibca400800120016002000791.74793.75793.90797.641481.711483.251488.042028.012029.792033.171. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Intel Open Image Denoise

Open Image Denoise is a denoising library for ray-tracing and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.0Run: RTLightmap.hdr.4096x4096 - Device: CPU-Onlydefghijabc0.19580.39160.58740.78320.9790.340.340.340.340.650.650.650.860.860.87

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.0Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Onlydefghijcab0.4140.8281.2421.6562.070.720.720.720.721.381.381.381.821.841.84

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/scivis/real_timefdegjihbca481216205.452275.453295.461535.4772510.5740010.5848010.5913013.7666013.8317013.87390

Intel Open Image Denoise

Open Image Denoise is a denoising library for ray-tracing and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.0Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Onlydefghijabc0.41180.82361.23541.64722.0590.720.720.720.721.371.371.371.831.831.83

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/ao/real_timedfeghjicba481216205.607475.614545.620405.6227810.8524010.8743010.8829014.1399014.1783014.23690

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Crownedfgjihabc132639526522.5722.5922.6622.7745.3245.4045.4856.0956.4656.81MIN: 22.39 / MAX: 22.93MIN: 22.39 / MAX: 22.98MIN: 22.45 / MAX: 22.99MIN: 22.57 / MAX: 23.16MIN: 44.74 / MAX: 46.66MIN: 44.87 / MAX: 46.45MIN: 44.92 / MAX: 46.64MIN: 54.05 / MAX: 59.82MIN: 54.53 / MAX: 59.89MIN: 55.27 / MAX: 59.91

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Mount St. Helensgfedijhcba71421283527.7026.8726.8026.7415.1915.1015.0311.3311.3211.021. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 96 - Buffer Length: 256 - Filter Length: 512efdgjhiacb150M300M450M600M750M2858800002859200002862500002865300005194400005197000005201400007116400007150300007181400001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_timeedfgijhabc481216206.582706.587456.595636.6008512.4987012.5104012.5438016.3468016.4365016.53500

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Homogeneous Halfspacedfgehjiacb81624324035.5735.5435.3835.0319.9619.7319.6215.1114.8114.461. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Vehicle Detection FP16-INT8 - Device: CPUegdfhjiabc60012001800240030001174.601175.581175.671180.852252.702254.722264.932873.242880.582881.141. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer - Model: Asian Dragondfegjhicba132639526524.6924.7024.7324.8248.1048.1248.1959.7959.9160.14MIN: 24.62 / MAX: 24.84MIN: 24.63 / MAX: 24.84MIN: 24.67 / MAX: 24.86MIN: 24.74 / MAX: 25MIN: 47.86 / MAX: 48.8MIN: 47.91 / MAX: 48.91MIN: 47.96 / MAX: 48.63MIN: 58.46 / MAX: 62.03MIN: 58.66 / MAX: 61.96MIN: 58.97 / MAX: 62

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer - Model: Asian Dragon Objfegdjhiacb122436486022.1522.1622.1922.2643.4243.5143.5753.5753.6953.81MIN: 22.07 / MAX: 22.32MIN: 22.08 / MAX: 22.35MIN: 22.12 / MAX: 22.33MIN: 22.18 / MAX: 22.42MIN: 43.14 / MAX: 43.89MIN: 43.26 / MAX: 44.02MIN: 43.34 / MAX: 44.03MIN: 52.17 / MAX: 55.38MIN: 52.63 / MAX: 55.24MIN: 52.72 / MAX: 55.86

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Person Vehicle Bike Detection FP16 - Device: CPUegdfhjibac50010001500200025001028.641031.601036.991041.872180.832194.362224.082450.262454.092455.511. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Embree

Intel Embree is a collection of high-performance ray-tracing kernels for execution on CPUs (and GPUs via SYCL) and supporting instruction sets such as SSE, AVX, AVX2, and AVX-512. Embree also supports making use of the Intel SPMD Program Compiler (ISPC). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Asian Dragon Objdgfeihjabc132639526523.8723.8823.9423.9446.3646.3846.4356.4956.6956.93MIN: 23.78 / MAX: 24.08MIN: 23.79 / MAX: 24.08MIN: 23.84 / MAX: 24.16MIN: 23.84 / MAX: 24.18MIN: 46.13 / MAX: 46.98MIN: 46.09 / MAX: 47.08MIN: 46.17 / MAX: 47.23MIN: 55.29 / MAX: 58.38MIN: 55.42 / MAX: 58.97MIN: 55.56 / MAX: 59.67

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.1Binary: Pathtracer ISPC - Model: Asian Dragonefdghijbac153045607528.3128.3228.3628.4854.3954.5154.5267.2067.3467.50MIN: 28.21 / MAX: 28.56MIN: 28.23 / MAX: 28.55MIN: 28.26 / MAX: 28.59MIN: 28.37 / MAX: 28.69MIN: 54.12 / MAX: 55.15MIN: 54.22 / MAX: 55.08MIN: 54.24 / MAX: 55.1MIN: 65.48 / MAX: 70.41MIN: 65.61 / MAX: 70.54MIN: 65.64 / MAX: 71.17

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: File Status - Threads: 50 - Files: 100000eagdcfjbih200K400K600K800K1000K389105529101561798632911657895709220751880862069869565925926

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Water-layered Halfspacegdefjihbca142842567062.8162.4462.3361.2837.8137.5136.4529.4627.0626.991. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Face Detection Retail FP16 - Device: CPUfgedhijbca130026003900520065002539.972557.662562.542564.784803.654848.424892.275836.275840.535882.911. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Road Segmentation ADAS FP16-INT8 - Device: CPUfdgeijhacb2004006008001000369.26370.57372.26373.64709.73709.75710.69842.91849.30854.511. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

SPECFEM3D

simulates acoustic (fluid), elastic (solid), coupled acoustic/elastic, poroelastic or seismic wave propagation in any type of conforming mesh of hexahedra. This test profile currently relies on CPU-based execution for SPECFEM3D and using a variety of their built-in examples/models for benchmarking. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterSPECFEM3D 4.0Model: Tomographic Modelgedfhjiabc71421283527.7527.4627.3326.9715.9915.8415.5912.3112.1012.041. (F9X) gfortran options: -O2 -fopenmp -std=f2003 -fimplicit-none -fmax-errors=10 -pedantic -pedantic-errors -O3 -finline-functions -lmpi_usempif08 -lmpi_mpifh -lmpi

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 96 - Buffer Length: 256 - Filter Length: 57egfdhijacb600M1200M1800M2400M3000M11178000001118200000112050000011208000001995500000199700000019995000002559800000256490000025711000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenVINO

OpenBenchmarking.orgFPS, More Is BetterOpenVINO 2023.1Model: Road Segmentation ADAS FP16 - Device: CPUgefdihjabc160320480640800341.36342.81343.49344.67642.90643.80648.27748.44750.49757.381. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 64 - Buffer Length: 256 - Filter Length: 512gedfjhibac130M260M390M520M650M2817300002818300002829200002830300005110700005120200005123100006109500006225600006226300001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPUhjiegdf0.47220.94441.41661.88882.3612.098801.949411.243081.144321.127231.037491.00136MIN: 1.29MIN: 1.26MIN: 1.04MIN: 1.07MIN: 0.93MIN: 0.92MIN: 0.921. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 64 - Buffer Length: 256 - Filter Length: 32gfedijhcab500M1000M1500M2000M2500M10562000001057100000105750000010595000002052600000205650000020575000002206800000220770000022121000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Face Detection FP16 - Device: CPUjihdefgacb2004006008001000805.38804.75804.58761.59761.16760.57759.92393.60393.37393.23MIN: 783.22 / MAX: 819.23MIN: 776.93 / MAX: 819.19MIN: 772.52 / MAX: 820.63MIN: 738.34 / MAX: 772.36MIN: 741.99 / MAX: 776.56MIN: 741.4 / MAX: 770.88MIN: 737.63 / MAX: 771.07MIN: 363.29 / MAX: 431.61MIN: 362.57 / MAX: 433.51MIN: 360.87 / MAX: 433.131. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUfdgeihj0.35390.70781.06171.41561.76951.5728201.5582401.5511801.5491100.7985400.7785430.768314MIN: 1.53MIN: 1.51MIN: 1.52MIN: 1.51MIN: 0.7MIN: 0.71MIN: 0.711. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Crownedgfhji102030405022.3422.3922.4222.4445.1945.4045.71MIN: 22.15 / MAX: 22.75MIN: 22.2 / MAX: 22.85MIN: 22.22 / MAX: 22.85MIN: 22.25 / MAX: 22.78MIN: 44.65 / MAX: 46.39MIN: 44.88 / MAX: 46.68MIN: 45.11 / MAX: 47.49

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUgdfejhi0.13780.27560.41340.55120.6890.6123200.6039500.6008340.5757940.3092780.3024600.301535MIN: 0.53MIN: 0.53MIN: 0.53MIN: 0.52MIN: 0.28MIN: 0.27MIN: 0.281. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterTimed Linux Kernel Compilation 6.1Build: defconfigdgfehjicab122436486055.1755.1755.1555.0933.9533.9333.8727.4127.3527.24

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Crownfgdehij102030405021.7721.8321.8921.9943.5743.8143.93MIN: 21.63 / MAX: 22.18MIN: 21.69 / MAX: 22.17MIN: 21.74 / MAX: 22.23MIN: 21.84 / MAX: 22.32MIN: 43.11 / MAX: 44.65MIN: 43.36 / MAX: 45.05MIN: 43.46 / MAX: 45.01

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: File Status - Threads: 100 - Files: 1000000bejdacfghi500K1000K1500K2000K2500K161970235627558036600601188679218939391964637204918023529412506266

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Face Detection FP16-INT8 - Device: CPUjihfedgacb90180270360450425.88425.68421.80399.24398.91398.52398.13213.94213.79213.62MIN: 404.76 / MAX: 434.06MIN: 402.91 / MAX: 432.03MIN: 269.94 / MAX: 598.22MIN: 387.9 / MAX: 408.93MIN: 386.2 / MAX: 407.29MIN: 382.1 / MAX: 404.98MIN: 379.09 / MAX: 404.71MIN: 201.64 / MAX: 242.71MIN: 197.29 / MAX: 236.32MIN: 197.2 / MAX: 235.231. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Asian Dragon Objdfeghji102030405023.3523.5023.5323.7145.9246.0246.09MIN: 23.26 / MAX: 23.57MIN: 23.4 / MAX: 23.74MIN: 23.43 / MAX: 23.73MIN: 23.61 / MAX: 23.93MIN: 45.64 / MAX: 46.53MIN: 45.75 / MAX: 46.57MIN: 45.82 / MAX: 46.6

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Asian Dragon Objgfedihj102030405022.2622.2722.2922.3543.7743.8443.90MIN: 22.18 / MAX: 22.43MIN: 22.2 / MAX: 22.44MIN: 22.22 / MAX: 22.46MIN: 22.28 / MAX: 22.5MIN: 43.51 / MAX: 44.16MIN: 43.64 / MAX: 44.38MIN: 43.66 / MAX: 44.27

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer - Model: Asian Dragonedfgjih112233445524.8324.8524.8924.9648.7148.7248.91MIN: 24.76 / MAX: 24.96MIN: 24.78 / MAX: 25MIN: 24.81 / MAX: 25.06MIN: 24.9 / MAX: 25.13MIN: 48.45 / MAX: 49.47MIN: 48.48 / MAX: 49.3MIN: 48.64 / MAX: 49.47

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Person Detection FP16 - Device: CPUijhgdefacb2040608010082.4982.0980.7774.7174.7174.5074.4342.4442.4342.20MIN: 70.77 / MAX: 94.62MIN: 68.73 / MAX: 91.87MIN: 69.54 / MAX: 95.42MIN: 66.29 / MAX: 79.68MIN: 66.12 / MAX: 81.09MIN: 66.5 / MAX: 80.32MIN: 65.68 / MAX: 83.49MIN: 36.14 / MAX: 61.98MIN: 36.31 / MAX: 62.36MIN: 36.84 / MAX: 61.971. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Embree

OpenBenchmarking.orgFrames Per Second, More Is BetterEmbree 4.3Binary: Pathtracer ISPC - Model: Asian Dragondfegjih122436486027.7427.8327.8327.9154.1554.1954.22MIN: 27.64 / MAX: 27.98MIN: 27.73 / MAX: 28.13MIN: 27.72 / MAX: 28.1MIN: 27.81 / MAX: 28.17MIN: 53.87 / MAX: 54.79MIN: 53.91 / MAX: 54.97MIN: 53.93 / MAX: 54.77

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUegfdjih0.76151.5232.28453.0463.80753.384363.381563.379563.377821.735011.734991.73381MIN: 3.33MIN: 3.33MIN: 3.33MIN: 3.33MIN: 1.64MIN: 1.65MIN: 1.641. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Person Detection FP32 - Device: CPUhjifdgeacb2040608010081.5881.5080.8874.8774.8174.5874.5442.2442.1942.09MIN: 68.74 / MAX: 95.81MIN: 68.9 / MAX: 92.66MIN: 39.72 / MAX: 92.54MIN: 66.72 / MAX: 80.96MIN: 66.88 / MAX: 80.7MIN: 67.63 / MAX: 78.73MIN: 65.97 / MAX: 82.9MIN: 36.59 / MAX: 61.56MIN: 36.21 / MAX: 65.64MIN: 37.13 / MAX: 58.711. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUfdegijh0.19140.38280.57420.76560.9570.8506910.8478050.8444340.8434920.4403680.4401560.440006MIN: 0.83MIN: 0.83MIN: 0.83MIN: 0.83MIN: 0.41MIN: 0.41MIN: 0.411. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Intel Open Image Denoise

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.1Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Onlydefghij0.31050.6210.93151.2421.55250.720.720.720.721.381.381.38

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.1Run: RTLightmap.hdr.4096x4096 - Device: CPU-Onlydefghij0.14630.29260.43890.58520.73150.340.340.340.340.650.650.65

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 2.0.0Benchmark: vklBenchmarkCPU Scalaredfghij80160240320400190191191191363363363MIN: 13 / MAX: 3484MIN: 13 / MAX: 3471MIN: 13 / MAX: 3484MIN: 13 / MAX: 3483MIN: 24 / MAX: 6610MIN: 24 / MAX: 6577MIN: 24 / MAX: 6613

Intel Open Image Denoise

OpenBenchmarking.orgImages / Sec, More Is BetterIntel Open Image Denoise 2.1Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Onlydefghij0.30830.61660.92491.23321.54150.720.720.720.721.371.371.37

OpenVKL

OpenVKL is the Intel Open Volume Kernel Library that offers high-performance volume computation kernels and part of the Intel oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems / Sec, More Is BetterOpenVKL 2.0.0Benchmark: vklBenchmarkCPU ISPCdefgijh2004006008001000487487488489922922926MIN: 36 / MAX: 6949MIN: 36 / MAX: 6956MIN: 36 / MAX: 6952MIN: 36 / MAX: 6969MIN: 67 / MAX: 12374MIN: 67 / MAX: 12356MIN: 67 / MAX: 12416

Remhos

Remhos (REMap High-Order Solver) is a miniapp that solves the pure advection equations that are used to perform monotonic and conservative discontinuous field interpolation (remap) as part of the Eulerian phase in Arbitrary Lagrangian Eulerian (ALE) simulations. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BetterRemhos 1.0Test: Sample Remap Exampleedgfhjibac71421283530.8530.7630.7530.7320.4420.3620.3016.7916.3516.241. (CXX) g++ options: -O3 -std=c++11 -lmfem -lHYPRE -lmetis -lrt -lmpi_cxx -lmpi

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Open - Threads: 100 - Files: 100000ejcbagfhdi120K240K360K480K600K294985370370403226404858420168460829523560526316529101555556

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUdfegjih0.480.961.441.922.42.133322.130622.125702.118131.155781.150121.14749MIN: 2MIN: 1.97MIN: 2.01MIN: 1.99MIN: 1.03MIN: 1MIN: 1.011. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: File Status - Threads: 100 - Files: 100000bfgadhejci200K400K600K800K1000K458716478469487805515464591716595238613497684932729927847458

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPUefdgjih0.43150.8631.29451.7262.15751.917811.914221.913741.912741.043331.043121.04100MIN: 1.88MIN: 1.88MIN: 1.88MIN: 1.88MIN: 0.94MIN: 0.94MIN: 0.941. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 64 - Buffer Length: 256 - Filter Length: 57dfegihjabc400M800M1200M1600M2000M10933000001094600000109540000010993000001899700000191680000019223000001994400000200190000020103000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPUfedgihj0.30190.60380.90571.20761.50951.3418301.3386101.3378901.3356400.7350940.7344610.731618MIN: 1.31MIN: 1.31MIN: 1.31MIN: 1.31MIN: 0.66MIN: 0.66MIN: 0.661. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Vehicle Detection FP16 - Device: CPUhjifgedbca369121510.7810.7710.7410.0910.0610.0610.015.915.905.89MIN: 5.59 / MAX: 21.13MIN: 6 / MAX: 18.16MIN: 5.92 / MAX: 24.44MIN: 5.4 / MAX: 19.17MIN: 5.2 / MAX: 19.38MIN: 5.29 / MAX: 19.07MIN: 5.7 / MAX: 19.52MIN: 4.84 / MAX: 12.9MIN: 4.83 / MAX: 13.4MIN: 4.67 / MAX: 18.41. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Machine Translation EN To DE FP16 - Device: CPUijhgedfacb153045607568.4268.3568.2764.7764.6864.4164.3137.8037.7937.79MIN: 56.13 / MAX: 75.77MIN: 55.82 / MAX: 74.84MIN: 56.41 / MAX: 79.96MIN: 55.8 / MAX: 69.46MIN: 38.02 / MAX: 72.52MIN: 37.44 / MAX: 73.04MIN: 50.85 / MAX: 70.77MIN: 33.35 / MAX: 56.45MIN: 33.29 / MAX: 54.88MIN: 32.97 / MAX: 53.71. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 64efgdcabih20K40K60K80K100K538935495655301553347846979090801839426195579

Test: oltp_read_write - Threads: 64

j: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 128dgefabhij20K40K60K80K100K597275994460145603108575789099104180104620105802

Test: oltp_read_write - Threads: 128

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Create - Threads: 100 - Files: 100000cbadjegifh13K26K39K52K65K35075374254073357971580055882458928590325938260496

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPUedfgijh0.68931.37862.06792.75723.44653.063703.059913.056743.054581.788761.786911.78170MIN: 2.97MIN: 2.96MIN: 2.97MIN: 2.97MIN: 1.65MIN: 1.66MIN: 1.641. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Vehicle Detection FP16-INT8 - Device: CPUhjiegdfacb2468107.097.087.056.806.796.796.764.174.164.16MIN: 4.43 / MAX: 16.88MIN: 4.35 / MAX: 16.86MIN: 4.44 / MAX: 16.57MIN: 4.04 / MAX: 15.37MIN: 3.79 / MAX: 15.41MIN: 3.8 / MAX: 15.48MIN: 4.04 / MAX: 15.47MIN: 3.39 / MAX: 10.07MIN: 3.43 / MAX: 10.26MIN: 3.42 / MAX: 11.21. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Create - Threads: 100 - Files: 1000000cbaefgdijh16K32K48K64K80K44001444374614570057705377092271296720107275473611

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUdefgjhi4008001200160020001643.991643.971642.351641.40991.12987.84985.74MIN: 1588.03MIN: 1590.89MIN: 1586.17MIN: 1589.91MIN: 954.92MIN: 952.67MIN: 949.931. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUdegfijh4008001200160020001641.921641.001637.371636.76991.07988.59986.96MIN: 1584.81MIN: 1595.55MIN: 1584.58MIN: 1585.98MIN: 953.96MIN: 950.96MIN: 949.021. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUdefgjhi4008001200160020001642.511639.361636.441631.99994.61993.56987.36MIN: 1593.16MIN: 1581.93MIN: 1585.81MIN: 1581.62MIN: 960.2MIN: 955.42MIN: 952.161. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Apache Cassandra

This is a benchmark of the Apache Cassandra NoSQL database management system making use of cassandra-stress. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOp/s, More Is BetterApache Cassandra 4.1.3Test: Writesefgdabcjih70K140K210K280K350K195798196287197092197866248095256661270480316485318109324575

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPUfegdjhi0.23880.47760.71640.95521.1941.0614401.0542501.0456701.0287500.7149700.7045500.644252MIN: 0.98MIN: 0.97MIN: 0.98MIN: 0.96MIN: 0.67MIN: 0.66MIN: 0.611. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Face Detection Retail FP16 - Device: CPUhijfgedcba0.7471.4942.2412.9883.7353.323.293.263.143.123.113.112.052.052.03MIN: 2.11 / MAX: 12.58MIN: 1.89 / MAX: 12.75MIN: 2.1 / MAX: 14.3MIN: 1.93 / MAX: 11.65MIN: 1.88 / MAX: 11.92MIN: 1.93 / MAX: 9.72MIN: 1.94 / MAX: 11.57MIN: 1.62 / MAX: 6.96MIN: 1.6 / MAX: 7MIN: 1.66 / MAX: 7.511. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 32edgfacbjih16K32K48K64K80K46737469774699347141589745963061520742527525475486

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Open - Threads: 50 - Files: 100000cabgedfijh140K280K420K560K700K401606460829469484546448552486578035578035581395617284645161

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Road Segmentation ADAS FP16-INT8 - Device: CPUjihfdgeacb51015202522.5322.5322.5021.6521.5721.4721.4014.2314.1214.03MIN: 18.74 / MAX: 31.08MIN: 19.09 / MAX: 30.15MIN: 13.76 / MAX: 30.22MIN: 19.48 / MAX: 24.27MIN: 19.5 / MAX: 24.76MIN: 17.62 / MAX: 28.13MIN: 19.07 / MAX: 25.3MIN: 11.51 / MAX: 25.86MIN: 11.51 / MAX: 26.04MIN: 11.59 / MAX: 26.041. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Person Vehicle Bike Detection FP16 - Device: CPUegdfhjibca2468107.777.747.707.677.337.287.184.894.884.88MIN: 5.42 / MAX: 16.35MIN: 6.06 / MAX: 12.66MIN: 5.51 / MAX: 16.06MIN: 5.32 / MAX: 16.6MIN: 5.45 / MAX: 15.94MIN: 5.53 / MAX: 15.78MIN: 4.98 / MAX: 16.11MIN: 3.93 / MAX: 13.44MIN: 3.9 / MAX: 14.94MIN: 3.95 / MAX: 16.051. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 128fgeacjh14K28K42K56K70K41424416954213851105528656406665816

Test: oltp_update_non_index - Threads: 128

b: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

d: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

i: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 32 - Buffer Length: 256 - Filter Length: 512fedghijcab90M180M270M360M450M2733900002734800002737600002740700003916800003932000003932700004244000004258100004296200001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Road Segmentation ADAS FP16 - Device: CPUihjgefdabc61218243024.8724.8424.6723.4223.3223.2823.2016.0215.9815.83MIN: 17 / MAX: 33.34MIN: 16.93 / MAX: 33.96MIN: 20.15 / MAX: 37.89MIN: 20.46 / MAX: 32.43MIN: 19.49 / MAX: 30.99MIN: 15.73 / MAX: 30.77MIN: 15.1 / MAX: 31.6MIN: 12.5 / MAX: 33.94MIN: 12.74 / MAX: 33.34MIN: 12.38 / MAX: 32.971. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 64dgefabij40K80K120K160K200K115675118549118657119092127567130802180179180581

Test: oltp_point_select - Threads: 64

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

h: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUegfdjhi0.57721.15441.73162.30882.8862.565222.514412.497142.494081.754531.742031.64478MIN: 2.32MIN: 2.3MIN: 2.26MIN: 2.3MIN: 1.52MIN: 1.51MIN: 1.421. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 128defcabjih40K80K120K160K200K129492129904130389149962159242159728197738198137200327

Test: oltp_point_select - Threads: 128

g: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUefdghij2004006008001000851.66849.34849.16837.60568.75564.45563.58MIN: 809.45MIN: 805.8MIN: 806.44MIN: 796.61MIN: 546.79MIN: 545.57MIN: 543.651. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 64egdfcbajih11K22K33K44K55K33881341073422434470391063975941281503045034651191

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 128gefcabijh8K16K24K32K40K245742461124830265462708727464361413664437126

Test: oltp_update_index - Threads: 128

d: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUfegdjih2004006008001000851.49849.71848.03838.52569.80566.18564.12MIN: 807.97MIN: 805.98MIN: 807.34MIN: 796.3MIN: 548.08MIN: 544.56MIN: 545.131. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUgdfehji2004006008001000847.42847.38845.31841.08568.65566.39563.33MIN: 806.72MIN: 806.33MIN: 803.78MIN: 798.46MIN: 547.26MIN: 542.7MIN: 544.041. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Kripke

Kripke is a simple, scalable, 3D Sn deterministic particle transport code. Its primary purpose is to research how data layout, programming paradigms and architectures effect the implementation and performance of Sn transport. Kripke is developed by LLNL. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgThroughput FoM, More Is BetterKripke 1.2.6efgdjih80M160M240M320M400M2362439002365910002371757002409945003490198003501512003548080001. (CXX) g++ options: -O3 -fopenmp -ldl

a: The test quit with a non-zero exit status.

b: The test quit with a non-zero exit status.

c: The test quit with a non-zero exit status.

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 1200defgihj91827364538.1138.0738.0237.9526.3926.1125.561. (CXX) g++ options: -O3 -fopenmp

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 64fdecbijh7K14K21K28K35K2106721108212712332424371305223063831332

Test: oltp_update_index - Threads: 64

a: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

g: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUefgdjih0.14260.28520.42780.57040.7130.6339750.6303250.6291080.6282360.4302700.4275120.426426MIN: 0.6MIN: 0.6MIN: 0.6MIN: 0.6MIN: 0.38MIN: 0.39MIN: 0.381. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 8 - Input: Bosphorus 4Kdfegcabihj2040608010066.9967.3967.7267.8190.4290.8191.3298.5799.1099.351. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Create - Threads: 50 - Files: 100000bacfdehijg13K26K39K52K65K41288436494393758343586175861758720589975934760680

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 2400edfghji2040608010099.4298.9897.9997.5368.5768.5268.241. (CXX) g++ options: -O3 -fopenmp

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Delete - Threads: 100 - Files: 100000cabjefighd20K40K60K80K100K734758756690827973719803999404102041102564105263105708

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 12 - Input: Bosphorus 4Kgfecdabjih50100150200250160.32161.85162.61163.06163.19163.46166.38224.41227.87230.031. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 32gefdabjih30K60K90K120K150K96840969079736898149104627106180137618138173138538

Test: oltp_point_select - Threads: 32

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 13 - Input: Bosphorus 1080pacbfgedjhi160320480640800510.36516.91542.61585.37586.75597.01604.99726.50726.89728.541. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 32egcdbaihj5K10K15K20K25K171171713517565176121781718361237732428624366

Test: oltp_update_index - Threads: 32

f: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 1abcdfehi1300260039005200650043314405447158985954597661256165

Test: oltp_point_select - Threads: 1

g: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

j: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 13 - Input: Bosphorus 4Kfgcdeabhij50100150200250160.80161.32161.50161.85162.05163.01166.69223.42227.21228.771. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

OSPRay

Intel OSPRay is a portable ray-tracing engine for high-performance, high-fidelity scientific visualizations. OSPRay builds off Intel's Embree and Intel SPMD Program Compiler (ISPC) components as part of the oneAPI rendering toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgItems Per Second, More Is BetterOSPRay 2.12Benchmark: particle_volume/pathtracer/real_timeegfdihjbca50100150200250151.51151.68151.78151.91192.40192.65192.70214.07214.14215.10

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 1bacdfgeih400800120016002000131213281381169316971705170818481861

Test: oltp_update_non_index - Threads: 1

j: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Create - Threads: 50 - Files: 1000000bcafedigjh16K32K48K64K80K52119522605366569920708977213472464727067315373411

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Delete - Threads: 50 - Files: 100000bcajfhedig20K40K60K80K100K7380190580910759560296993100402100604101010102249103950

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 1cbagefjhi7001400210028003500248525102540319532093218347934803485

Test: oltp_read_write - Threads: 1

d: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 1cadgfeijh400800120016002000118912121479148114831490165616601666

Test: oltp_update_index - Threads: 1

b: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 12 - Input: Bosphorus 1080pabcfedghji130260390520650422.99427.69431.90521.52525.17526.22528.53580.16584.12591.321. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUegdfjih0.28810.57620.86431.15241.44051.2804301.2791801.2575801.2065300.9360010.9317930.926283MIN: 1.24MIN: 1.24MIN: 1.21MIN: 1.18MIN: 0.86MIN: 0.86MIN: 0.851. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 32defabijh8K16K24K32K40K2627326285266952873528914356503565536041

Test: oltp_update_non_index - Threads: 32

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

g: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 32 - Buffer Length: 256 - Filter Length: 57fegdabchji300M600M900M1200M1500M10246000001032000000103340000010350000001192100000121420000012548000001352900000136980000013941000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_index - Threads: 16aedgcfihj4K8K12K16K20K125581256712622126271268112692168171696516972

Test: oltp_update_index - Threads: 16

b: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_point_select - Threads: 16cbgfeihj20K40K60K80K100K6540667515699237010570250864718721887412

Test: oltp_point_select - Threads: 16

a: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

d: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

easyWave

The easyWave software allows simulating tsunami generation and propagation in the context of early warning systems. EasyWave supports making use of OpenMP for CPU multi-threading and there are also GPU ports available but not currently incorporated as part of this test profile. The easyWave tsunami generation software is run with one of the example/reference input files for measuring the CPU execution time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgSeconds, Fewer Is BettereasyWave r34Input: e2Asean Grid + BengkuluSept2007 Source - Time: 240fdegihj0.37280.74561.11841.49121.8641.6571.6571.6541.6481.2881.2841.2451. (CXX) g++ options: -O3 -fopenmp

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Delete - Threads: 100 - Files: 1000000bacjfhdegi20K40K60K80K100K867159011497031110693110803111782112613113225113895114692

TiDB Community Server

This is a PingCAP TiDB Community Server benchmark facilitated using the sysbench OLTP database benchmarks. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_read_write - Threads: 16gfdebcajih10K20K30K40K50K36088361253648036784369503736838331474144750147576

OpenBenchmarking.orgQueries Per Second, More Is BetterTiDB Community Server 7.3Test: oltp_update_non_index - Threads: 16baedgijh5K10K15K20K25K1806818095185571856318735235412354323794

Test: oltp_update_non_index - Threads: 16

c: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

f: The test quit with a non-zero exit status. E: FATAL: Thread initialization failed!

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Rename - Threads: 100 - Files: 1000000cbadheigfj20K40K60K80K100K66827721297307881208843108436085215857638581587161

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Rename - Threads: 50 - Files: 100000abcfegdhji20K40K60K80K100K70522730467710181633822378223782372830568920690827

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 8 - Input: Bosphorus 1080pgfdebacijh306090120150118.48118.49118.95119.31138.34141.22143.55149.05149.45151.441. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 4 - Input: Bosphorus 4Kdefgchibja1.17072.34143.51214.68285.85354.1074.1144.1384.1435.0495.0755.0795.1495.1605.2031. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Delete - Threads: 50 - Files: 1000000cbaigdfehj20K40K60K80K100K901479731498932110436110828111012111198113327113404113572

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Rename - Threads: 100 - Files: 100000cbafhgdije20K40K60K80K100K67159693487552979491798088038682102823728347283822

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 2 - Buffer Length: 256 - Filter Length: 512gdfehijbac6M12M18M24M30M227270002462700025199000252070002564800025910000263780002773600027901000282270001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

Apache Hadoop

This is a benchmark of the Apache Hadoop making use of its built-in name-node throughput benchmark (NNThroughputBenchmark). Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps per sec, More Is BetterApache Hadoop 3.3.6Operation: Rename - Threads: 50 - Files: 1000000bacfdeghij20K40K60K80K100K71679732397463882501839218404184810851218631188550

nekRS

nekRS is an open-source Navier Stokes solver based on the spectral element method. NekRS supports both CPU and GPU/accelerator support though this test profile is currently configured for CPU execution. NekRS is part of Nek5000 of the Mathematics and Computer Science MCS at Argonne National Laboratory. This nekRS benchmark is primarily relevant to large core count HPC servers and otherwise may be very time consuming on smaller systems. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgflops/rank, More Is BetternekRS 23.0Input: Kershawhjifedgcab2000M4000M6000M8000M10000M91459000009242080000926989000099764500001026400000010318900000105006000001082670000011106900000112403000001. (CXX) g++ options: -fopenmp -O2 -march=native -mtune=native -ftree-vectorize -rdynamic -lmpi_cxx -lmpi

OpenBenchmarking.orgflops/rank, More Is BetternekRS 23.0Input: TurboPipe Periodiccbhaijedfg2000M4000M6000M8000M10000M67541700006757360000676127000067677100006768070000683536000079310100007934570000795579000079649100001. (CXX) g++ options: -fopenmp -O2 -march=native -mtune=native -ftree-vectorize -rdynamic -lmpi_cxx -lmpi

SVT-AV1

OpenBenchmarking.orgFrames Per Second, More Is BetterSVT-AV1 1.7Encoder Mode: Preset 4 - Input: Bosphorus 1080pfdegjhiabc369121510.7410.9110.9811.0212.1812.2312.2612.4812.5912.621. (CXX) g++ options: -march=native -mno-avx -mavx2 -mavx512f -mavx512bw -mavx512dq

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 1 - Buffer Length: 256 - Filter Length: 512gefdjhiabc3M6M9M12M15M122560001236600012681000126830001289900012909000133630001390900014021000142250001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 1 - Buffer Length: 256 - Filter Length: 57hdegfjicba13M26M39M52M65M514430005266500052827000528540005287900055715000558410005751900059296000594010001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Sentiment Analysis, 80% Pruned Quantized BERT Base Uncased - Scenario: Asynchronous Multi-Streamcbajhigdfe81624324035.6835.6435.6331.5031.4931.4431.0631.0531.0330.99

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 2 - Buffer Length: 256 - Filter Length: 32dgefjhicba17M34M51M68M85M670540006867800068846000688610007240000072468000724910007692400077019000771810001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 32 - Buffer Length: 256 - Filter Length: 32fedgjhiacb300M600M900M1200M1500M10419000001046600000104710000010471000001169900000117240000011729000001183500000118480000011903000001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 8 - Buffer Length: 256 - Filter Length: 512efdgjhibca20M40M60M80M100M9700500099441000995940001001700001042200001044400001047800001080800001091400001098700001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 2 - Buffer Length: 256 - Filter Length: 57ghedfijbac30M60M90M120M150M1048000001052800001054800001056500001057400001097000001113700001140100001174900001185500001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 4 - Buffer Length: 256 - Filter Length: 512gjfdeihacb12M24M36M48M60M495560004978100049977000502580005038000052129000528900005291100055165000555880001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 1 - Buffer Length: 256 - Filter Length: 32dgfejihcba8M16M24M32M40M352280003523600035271000353150003712000037141000371450003945300039486000394990001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 16 - Buffer Length: 256 - Filter Length: 512dfgehjicab50M100M150M200M250M1938500001945000001946700001960400002075900002094900002097200002149100002160800002161500001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 8 - Buffer Length: 256 - Filter Length: 32fgedijhbca70M140M210M280M350M2763900002774100002777800002780300002924300002924900002926200003051100003067600003075400001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 16 - Buffer Length: 256 - Filter Length: 32gfedijhabc130M260M390M520M650M5430500005450200005451400005453600005838700005856700005858500005942300006024700006036500001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 4 - Buffer Length: 256 - Filter Length: 32gfdejhicba30M60M90M120M150M1384600001385800001386000001386200001457400001458200001459200001536700001536900001538500001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Age Gender Recognition Retail 0013 FP16 - Device: CPUcbajihgfed0.12150.2430.36450.4860.60750.540.540.540.530.530.530.490.490.490.49MIN: 0.45 / MAX: 5.03MIN: 0.45 / MAX: 7.81MIN: 0.45 / MAX: 7.64MIN: 0.31 / MAX: 7.18MIN: 0.31 / MAX: 10.06MIN: 0.31 / MAX: 10.06MIN: 0.3 / MAX: 8.84MIN: 0.3 / MAX: 8.2MIN: 0.3 / MAX: 9.07MIN: 0.3 / MAX: 9.281. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 16 - Buffer Length: 256 - Filter Length: 57cgdbefaijh160M320M480M640M800M6749300006820700006891500006927600006929200006933400006997400007358000007371200007414200001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Handwritten English Recognition FP16-INT8 - Device: CPUdhjicbafge91827364540.4039.4439.3439.2338.7538.6638.5037.0136.9836.98MIN: 26.93 / MAX: 74.83MIN: 33.14 / MAX: 45.28MIN: 25.19 / MAX: 46.89MIN: 34.71 / MAX: 47.63MIN: 37.46 / MAX: 43.52MIN: 37.22 / MAX: 43.52MIN: 36.77 / MAX: 44.23MIN: 32.25 / MAX: 43.6MIN: 32.61 / MAX: 41.91MIN: 32.02 / MAX: 44.781. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, BERT base uncased SST2, Sparse INT8 - Scenario: Asynchronous Multi-Streambacijhdfge4812162017.0716.9116.8915.9915.9415.9215.7215.7215.6915.62

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 8 - Buffer Length: 256 - Filter Length: 57fgedbcaihj80M160M240M320M400M3504500003578100003579900003633100003669300003669900003694300003772500003783300003816600001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Face Detection Retail FP16-INT8 - Device: CPUcabjihgedf1.09352.1873.28054.3745.46754.864.864.854.814.814.804.524.514.514.50MIN: 4.34 / MAX: 12.27MIN: 4.23 / MAX: 12.81MIN: 4.25 / MAX: 12.86MIN: 3.23 / MAX: 14.45MIN: 3.23 / MAX: 15.04MIN: 3.23 / MAX: 14.95MIN: 2.77 / MAX: 13.57MIN: 2.96 / MAX: 16.06MIN: 2.98 / MAX: 13.05MIN: 2.98 / MAX: 13.861. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: ResNet-50, Sparse INT8 - Scenario: Asynchronous Multi-Streamdfegjihabc1.12412.24823.37234.49645.62054.99604.98774.98594.97874.80164.80154.80054.65084.64764.6348

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: CV Segmentation, 90% Pruned YOLACT Pruned - Scenario: Asynchronous Multi-Streamacbhijdegf80160240320400347.66347.37347.22337.96336.77336.68325.88325.74325.51324.96

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Weld Porosity Detection FP16-INT8 - Device: CPUihjabcfged2468108.468.468.458.288.278.247.977.967.967.93MIN: 4.49 / MAX: 17.8MIN: 4.67 / MAX: 18MIN: 4.46 / MAX: 17.31MIN: 7.44 / MAX: 23.35MIN: 7.37 / MAX: 25.18MIN: 7.62 / MAX: 23.32MIN: 4.37 / MAX: 16.86MIN: 4.19 / MAX: 14.2MIN: 4.19 / MAX: 16.59MIN: 4.2 / MAX: 16.921. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Liquid-DSP

LiquidSDR's Liquid-DSP is a software-defined radio (SDR) digital signal processing library. This test profile runs a multi-threaded benchmark of this SDR/DSP library focused on embedded platform usage. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgsamples/s, More Is BetterLiquid-DSP 1.6Threads: 4 - Buffer Length: 256 - Filter Length: 57dfgecabijh40M80M120M160M200M1889300001898800001907500001912300001945100001962200001965900001986400002001200002009300001. (CC) gcc options: -O3 -pthread -lm -lc -lliquid

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Handwritten English Recognition FP16 - Device: CPUjbhciaedfg71421283531.5731.0030.9230.8930.8730.7230.1030.0229.9529.72MIN: 20.39 / MAX: 39.22MIN: 29.59 / MAX: 36.33MIN: 25.94 / MAX: 41.77MIN: 29.48 / MAX: 36.29MIN: 20.13 / MAX: 42.34MIN: 29.51 / MAX: 35.07MIN: 22.61 / MAX: 39.15MIN: 18.78 / MAX: 38.72MIN: 19.01 / MAX: 38.08MIN: 19.46 / MAX: 38.991. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Question Answering, BERT base uncased SQuaD 12layer Pruned90 - Scenario: Asynchronous Multi-Streambcaijhgfde306090120150118.95118.78118.75116.89116.85116.51112.48112.41112.25112.06

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Weld Porosity Detection FP16 - Device: CPUjihacbfged4812162016.2716.2716.2716.2616.0216.0215.3815.3715.3615.36MIN: 8.44 / MAX: 25.48MIN: 8.5 / MAX: 25.86MIN: 8.92 / MAX: 25.52MIN: 14.71 / MAX: 28.14MIN: 14.63 / MAX: 33.79MIN: 14.41 / MAX: 30.55MIN: 7.99 / MAX: 24MIN: 7.99 / MAX: 23.98MIN: 8.02 / MAX: 23.81MIN: 8.08 / MAX: 24.341. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of the Intel oneAPI toolkit. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 3.3Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUegfdhji0.86491.72982.59473.45964.32453.844213.823813.818233.815763.722473.680873.65907MIN: 3.27MIN: 3.29MIN: 3.25MIN: 3.26MIN: 2.83MIN: 2.85MIN: 2.811. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -pie -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, BERT base uncased SST2 - Scenario: Asynchronous Multi-Streambahjicgedf306090120150150.61150.59146.12145.90145.83145.26144.11144.10143.76143.69

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: BERT-Large, NLP Question Answering - Scenario: Asynchronous Multi-Streamcghiefjdba110220330440550507.48495.60495.22494.50494.26494.22494.01493.60487.36485.72

OpenVINO

OpenBenchmarking.orgms, Fewer Is BetterOpenVINO 2023.1Model: Age Gender Recognition Retail 0013 FP16-INT8 - Device: CPUjihgfedcba0.07880.15760.23640.31520.3940.350.350.350.350.350.350.350.340.340.34MIN: 0.22 / MAX: 8.35MIN: 0.22 / MAX: 8.62MIN: 0.21 / MAX: 8.91MIN: 0.23 / MAX: 8.63MIN: 0.23 / MAX: 9.15MIN: 0.23 / MAX: 8.84MIN: 0.23 / MAX: 9.09MIN: 0.29 / MAX: 7.09MIN: 0.29 / MAX: 10.87MIN: 0.29 / MAX: 7.331. (CXX) g++ options: -fsigned-char -ffunction-sections -fdata-sections -O3 -fno-strict-overflow -fwrapv -pie -ldl

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: BERT-Large, NLP Question Answering, Sparse INT8 - Scenario: Asynchronous Multi-Streamcbgafedijh81624324033.4633.3833.3733.3433.2833.2633.2232.8432.8132.75

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Text Classification, DistilBERT mnli - Scenario: Asynchronous Multi-Streambcaihjdefg2040608010074.5674.5074.3273.9473.8873.7873.3173.2673.2273.19

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Document Classification, oBERT base uncased on IMDB - Scenario: Asynchronous Multi-Streamjhiefgdcba130260390520650614.23613.05612.98607.91607.82607.16606.10605.92605.73605.04

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: NLP Token Classification, BERT base uncased conll2003 - Scenario: Asynchronous Multi-Streamihjgfebdca130260390520650613.64613.16612.52608.72606.79606.76606.67606.58605.88605.76

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: ResNet-50, Baseline - Scenario: Asynchronous Multi-Streamabdfgcehji112233445549.3749.1149.0949.0749.0349.0249.0148.8848.8348.79

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: CV Detection, YOLOv5s COCO, Sparse INT8 - Scenario: Asynchronous Multi-Streamdfegjahicb20406080100110.11110.00109.97109.90109.86109.80109.79109.74109.58109.23

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: CV Classification, ResNet-50 ImageNet - Scenario: Asynchronous Multi-Streamcfeagbjihd112233445549.2149.0749.0649.0148.9848.9748.8948.8948.8948.87

QuantLib

QuantLib is an open-source library/framework around quantitative finance for modeling, trading and risk management scenarios. QuantLib is written in C++ with Boost and its built-in benchmark used reports the QuantLib Benchmark Index benchmark score. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMFLOPS, More Is BetterQuantLib 1.32Configuration: Multi-Threadedhij30K60K90K120K150K116781.5117199.0117263.31. (CXX) g++ options: -O3 -march=native -fPIE -pie

OpenBenchmarking.orgMFLOPS, More Is BetterQuantLib 1.32Configuration: Single-Threadedhij70014002100280035003357.93365.43371.61. (CXX) g++ options: -O3 -march=native -fPIE -pie

Neural Magic DeepSparse

OpenBenchmarking.orgms/batch, Fewer Is BetterNeural Magic DeepSparse 1.5Model: CV Detection, YOLOv5s COCO - Scenario: Asynchronous Multi-Streamdecaigbjfh20406080100111.11111.09111.03111.01110.98110.98110.92110.90110.89110.81

Timed Linux Kernel Compilation

This test times how long it takes to build the Linux kernel in a default configuration (defconfig) for the architecture being tested or alternatively an allmodconfig for building all possible kernel modules for the build. Learn more via the OpenBenchmarking.org test page.

Build: allmodconfig

a: The test quit with a non-zero exit status.

b: The test quit with a non-zero exit status.

c: The test quit with a non-zero exit status.

d: The test quit with a non-zero exit status.

e: The test quit with a non-zero exit status.

f: The test quit with a non-zero exit status.

g: The test quit with a non-zero exit status.

h: The test quit with a non-zero exit status.

i: The test quit with a non-zero exit status.

j: The test quit with a non-zero exit status.

OpenRadioss

OpenRadioss is an open-source AGPL-licensed finite element solver for dynamic event analysis OpenRadioss is based on Altair Radioss and open-sourced in 2022. This open-source finite element solver is benchmarked with various example models available from https://www.openradioss.org/models/ and https://github.com/OpenRadioss/ModelExchange/tree/main/Examples. This test is currently using a reference OpenRadioss binary build offered via GitHub. Learn more via the OpenBenchmarking.org test page.

Model: INIVOL and Fluid Structure Interaction Drop Container

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

h: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

i: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

j: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

Model: Rubber O-Ring Seal Installation

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

h: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

i: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

j: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

Model: Bird Strike on Windshield

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

h: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

i: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

j: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

Model: Cell Phone Drop Test

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

h: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

i: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

j: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

Model: Chrysler Neon 1M

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

h: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

i: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

j: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

Model: Bumper Beam

a: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

b: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

c: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

d: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

e: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

f: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

g: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

h: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

i: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

j: The test run did not produce a result. E: ./engine_linux64_gf_ompi: error while loading shared libraries: libmpi.so.40: cannot open shared object file: No such file or directory

207 Results Shown

Apache Hadoop:
  File Status - 50 - 1000000
  Open - 100 - 1000000
  Open - 50 - 1000000
Neural Magic DeepSparse
OpenVINO
Neural Magic DeepSparse:
  BERT-Large, NLP Question Answering - Asynchronous Multi-Stream
  NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream
  CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Stream
  NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream
  CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream
  ResNet-50, Baseline - Asynchronous Multi-Stream
  CV Detection, YOLOv5s COCO - Asynchronous Multi-Stream
  BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Stream
oneDNN
Neural Magic DeepSparse:
  NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Stream
  NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream
OpenVINO:
  Handwritten English Recognition FP16 - CPU
  Face Detection FP16 - CPU
  Weld Porosity Detection FP16-INT8 - CPU
OSPRay:
  particle_volume/ao/real_time
  particle_volume/scivis/real_time
OpenVINO
Neural Magic DeepSparse
Liquid-DSP
Neural Magic DeepSparse
OpenVINO
Neural Magic DeepSparse
Blender
OpenVINO
Blender:
  Classroom - CPU-Only
  BMW27 - CPU-Only
OpenVINO
Blender
OpenVINO:
  Age Gender Recognition Retail 0013 FP16 - CPU
  Person Detection FP32 - CPU
SPECFEM3D
OpenVINO
Blender
Neural Magic DeepSparse
BRL-CAD
Embree
OpenVINO:
  Machine Translation EN To DE FP16 - CPU
  Vehicle Detection FP16 - CPU
Intel Open Image Denoise:
  RTLightmap.hdr.4096x4096 - CPU-Only
  RT.ldr_alb_nrm.3840x2160 - CPU-Only
OSPRay
Intel Open Image Denoise
OSPRay
Embree
SPECFEM3D
Liquid-DSP
OSPRay
SPECFEM3D
OpenVINO
Embree:
  Pathtracer - Asian Dragon
  Pathtracer - Asian Dragon Obj
OpenVINO
Embree:
  Pathtracer ISPC - Asian Dragon Obj
  Pathtracer ISPC - Asian Dragon
Apache Hadoop
SPECFEM3D
OpenVINO:
  Face Detection Retail FP16 - CPU
  Road Segmentation ADAS FP16-INT8 - CPU
SPECFEM3D
Liquid-DSP
OpenVINO
Liquid-DSP
oneDNN
Liquid-DSP
OpenVINO
oneDNN
Embree
oneDNN
Timed Linux Kernel Compilation
Embree
Apache Hadoop
OpenVINO
Embree:
  Pathtracer ISPC - Asian Dragon Obj
  Pathtracer - Asian Dragon Obj
  Pathtracer - Asian Dragon
OpenVINO
Embree
oneDNN
OpenVINO
oneDNN
Intel Open Image Denoise:
  RT.ldr_alb_nrm.3840x2160 - CPU-Only
  RTLightmap.hdr.4096x4096 - CPU-Only
OpenVKL
Intel Open Image Denoise
OpenVKL
Remhos
Apache Hadoop
oneDNN
Apache Hadoop
oneDNN
Liquid-DSP
oneDNN
OpenVINO:
  Vehicle Detection FP16 - CPU
  Machine Translation EN To DE FP16 - CPU
TiDB Community Server:
  oltp_read_write - 64
  oltp_read_write - 128
Apache Hadoop
oneDNN
OpenVINO
Apache Hadoop
oneDNN:
  Recurrent Neural Network Training - bf16bf16bf16 - CPU
  Recurrent Neural Network Training - f32 - CPU
  Recurrent Neural Network Training - u8s8f32 - CPU
Apache Cassandra
oneDNN
OpenVINO
TiDB Community Server
Apache Hadoop
OpenVINO:
  Road Segmentation ADAS FP16-INT8 - CPU
  Person Vehicle Bike Detection FP16 - CPU
TiDB Community Server
Liquid-DSP
OpenVINO
TiDB Community Server
oneDNN
TiDB Community Server
oneDNN
TiDB Community Server:
  oltp_update_non_index - 64
  oltp_update_index - 128
oneDNN:
  Recurrent Neural Network Inference - f32 - CPU
  Recurrent Neural Network Inference - bf16bf16bf16 - CPU
Kripke
easyWave
TiDB Community Server
oneDNN
SVT-AV1
Apache Hadoop
easyWave
Apache Hadoop
SVT-AV1
TiDB Community Server
SVT-AV1
TiDB Community Server:
  oltp_update_index - 32
  oltp_point_select - 1
SVT-AV1
OSPRay
TiDB Community Server
Apache Hadoop:
  Create - 50 - 1000000
  Delete - 50 - 100000
TiDB Community Server:
  oltp_read_write - 1
  oltp_update_index - 1
SVT-AV1
oneDNN
TiDB Community Server
Liquid-DSP
TiDB Community Server:
  oltp_update_index - 16
  oltp_point_select - 16
easyWave
Apache Hadoop
TiDB Community Server:
  oltp_read_write - 16
  oltp_update_non_index - 16
Apache Hadoop:
  Rename - 100 - 1000000
  Rename - 50 - 100000
SVT-AV1:
  Preset 8 - Bosphorus 1080p
  Preset 4 - Bosphorus 4K
Apache Hadoop:
  Delete - 50 - 1000000
  Rename - 100 - 100000
Liquid-DSP
Apache Hadoop
nekRS:
  Kershaw
  TurboPipe Periodic
SVT-AV1
Liquid-DSP:
  1 - 256 - 512
  1 - 256 - 57
Neural Magic DeepSparse
Liquid-DSP:
  2 - 256 - 32
  32 - 256 - 32
  8 - 256 - 512
  2 - 256 - 57
  4 - 256 - 512
  1 - 256 - 32
  16 - 256 - 512
  8 - 256 - 32
  16 - 256 - 32
  4 - 256 - 32
OpenVINO
Liquid-DSP
OpenVINO
Neural Magic DeepSparse
Liquid-DSP
OpenVINO
Neural Magic DeepSparse:
  ResNet-50, Sparse INT8 - Asynchronous Multi-Stream
  CV Segmentation, 90% Pruned YOLACT Pruned - Asynchronous Multi-Stream
OpenVINO
Liquid-DSP
OpenVINO
Neural Magic DeepSparse
OpenVINO
oneDNN
Neural Magic DeepSparse:
  NLP Text Classification, BERT base uncased SST2 - Asynchronous Multi-Stream
  BERT-Large, NLP Question Answering - Asynchronous Multi-Stream
OpenVINO
Neural Magic DeepSparse:
  BERT-Large, NLP Question Answering, Sparse INT8 - Asynchronous Multi-Stream
  NLP Text Classification, DistilBERT mnli - Asynchronous Multi-Stream
  NLP Document Classification, oBERT base uncased on IMDB - Asynchronous Multi-Stream
  NLP Token Classification, BERT base uncased conll2003 - Asynchronous Multi-Stream
  ResNet-50, Baseline - Asynchronous Multi-Stream
  CV Detection, YOLOv5s COCO, Sparse INT8 - Asynchronous Multi-Stream
  CV Classification, ResNet-50 ImageNet - Asynchronous Multi-Stream
QuantLib:
  Multi-Threaded
  Single-Threaded
Neural Magic DeepSparse