9600X AMD Ryzen 5 9600X 6-Core testing with a ASUS ROG STRIX X670E-E GAMING WIFI (2308 BIOS) and AMD Radeon RX 7900 GRE 16GB on Ubuntu 24.04 via the Phoronix Test Suite. a: Processor: AMD Ryzen 5 9600X 6-Core @ 5.48GHz (6 Cores / 12 Threads), Motherboard: ASUS ROG STRIX X670E-E GAMING WIFI (2308 BIOS), Chipset: AMD Device 14d8, Memory: 2 x 16GB DDR5-6000MT/s G Skill F5-6000J3038F16G, Disk: 2000GB Corsair MP700 PRO, Graphics: AMD Radeon RX 7900 GRE 16GB, Audio: AMD Navi 31 HDMI/DP, Monitor: DELL U2723QE, Network: Intel I225-V + Intel Wi-Fi 6E OS: Ubuntu 24.04, Kernel: 6.10.0-phx (x86_64), Desktop: GNOME Shell 46.0, Display Server: X Server + Wayland, OpenGL: 4.6 Mesa 24.2~git2406040600.8112d4~oibaf~n (git-8112d44 2024-06-04 noble-oibaf-ppa) (LLVM 17.0.6 DRM 3.57), Compiler: GCC 13.2.0, File-System: ext4, Screen Resolution: 3840x2160 b: Processor: AMD Ryzen 5 9600X 6-Core @ 5.48GHz (6 Cores / 12 Threads), Motherboard: ASUS ROG STRIX X670E-E GAMING WIFI (2308 BIOS), Chipset: AMD Device 14d8, Memory: 2 x 16GB DDR5-6000MT/s G Skill F5-6000J3038F16G, Disk: 2000GB Corsair MP700 PRO, Graphics: AMD Radeon RX 7900 GRE 16GB, Audio: AMD Navi 31 HDMI/DP, Monitor: DELL U2723QE, Network: Intel I225-V + Intel Wi-Fi 6E OS: Ubuntu 24.04, Kernel: 6.10.0-phx (x86_64), Desktop: GNOME Shell 46.0, Display Server: X Server + Wayland, OpenGL: 4.6 Mesa 24.2~git2406040600.8112d4~oibaf~n (git-8112d44 2024-06-04 noble-oibaf-ppa) (LLVM 17.0.6 DRM 3.57), Compiler: GCC 13.2.0, File-System: ext4, Screen Resolution: 3840x2160 SVT-AV1 2.2 Encoder Mode: Preset 3 - Input: Beauty 4K 10-bit Frames Per Second > Higher Is Better a . 0.869 |==================================================================== b . 0.870 |==================================================================== Blender 4.2 Blend File: Barbershop - Compute: CPU-Only Seconds < Lower Is Better a . 1128.79 |================================================================== b . 1129.22 |================================================================== SVT-AV1 2.2 Encoder Mode: Preset 3 - Input: Bosphorus 4K Frames Per Second > Higher Is Better a . 5.597 |==================================================================== b . 5.611 |==================================================================== Whisperfile 20Aug24 Model Size: Medium Seconds < Lower Is Better a . 424.17 |=================================================================== b . 423.65 |=================================================================== Blender 4.2 Blend File: Pabellon Barcelona - Compute: CPU-Only Seconds < Lower Is Better a . 357.14 |=================================================================== b . 356.68 |=================================================================== LeelaChessZero 0.31.1 Backend: Eigen Nodes Per Second > Higher Is Better a . 296 |====================================================================== b . 277 |================================================================== ACES DGEMM 1.0 Sustained Floating-Point Rate GFLOP/s > Higher Is Better a . 144.48 |=================================================================== b . 144.97 |=================================================================== LeelaChessZero 0.31.1 Backend: BLAS Nodes Per Second > Higher Is Better a . 395 |====================================================================== b . 372 |================================================================== SVT-AV1 2.2 Encoder Mode: Preset 5 - Input: Beauty 4K 10-bit Frames Per Second > Higher Is Better a . 3.765 |==================================================================== b . 3.777 |==================================================================== Blender 4.2 Blend File: Classroom - Compute: CPU-Only Seconds < Lower Is Better a . 317.03 |=================================================================== b . 316.58 |=================================================================== XNNPACK 2cd86b Model: QU8MobileNetV3Small us < Lower Is Better a . 290 |====================================================================== b . 290 |====================================================================== XNNPACK 2cd86b Model: QU8MobileNetV3Large us < Lower Is Better a . 604 |====================================================================== b . 605 |====================================================================== XNNPACK 2cd86b Model: QU8MobileNetV2 us < Lower Is Better a . 589 |===================================================================== b . 594 |====================================================================== XNNPACK 2cd86b Model: FP16MobileNetV3Small us < Lower Is Better a . 478 |===================================================================== b . 482 |====================================================================== XNNPACK 2cd86b Model: FP16MobileNetV3Large us < Lower Is Better a . 1198 |===================================================================== b . 1193 |===================================================================== XNNPACK 2cd86b Model: FP16MobileNetV2 us < Lower Is Better a . 1252 |===================================================================== b . 1251 |===================================================================== XNNPACK 2cd86b Model: FP32MobileNetV3Small us < Lower Is Better a . 301 |====================================================================== b . 302 |====================================================================== XNNPACK 2cd86b Model: FP32MobileNetV3Large us < Lower Is Better a . 1040 |===================================================================== b . 1025 |==================================================================== XNNPACK 2cd86b Model: FP32MobileNetV2 us < Lower Is Better a . 846 |====================================================================== b . 849 |====================================================================== BYTE Unix Benchmark 5.1.3-git Computational Test: Whetstone Double MWIPS > Higher Is Better a . 149214.7 |================================================================= b . 149060.0 |================================================================= BYTE Unix Benchmark 5.1.3-git Computational Test: Pipe LPS > Higher Is Better a . 44538458.4 |=============================================================== b . 44695866.6 |=============================================================== BYTE Unix Benchmark 5.1.3-git Computational Test: System Call LPS > Higher Is Better a . 43927689.4 |=============================================================== b . 43955081.6 |=============================================================== BYTE Unix Benchmark 5.1.3-git Computational Test: Dhrystone 2 LPS > Higher Is Better a . 781175751.0 |============================================================== b . 780130318.1 |============================================================== SVT-AV1 2.2 Encoder Mode: Preset 8 - Input: Beauty 4K 10-bit Frames Per Second > Higher Is Better a . 5.113 |==================================================================== b . 5.081 |==================================================================== SVT-AV1 2.2 Encoder Mode: Preset 3 - Input: Bosphorus 1080p Frames Per Second > Higher Is Better a . 19.80 |==================================================================== b . 19.78 |==================================================================== OSPRay 3.2 Benchmark: particle_volume/scivis/real_time Items Per Second > Higher Is Better a . 4.75488 |================================================================== b . 4.75153 |================================================================== GROMACS Input: water_GMX50_bare Ns Per Day > Higher Is Better a . 1.060 |==================================================================== b . 1.056 |==================================================================== Build2 0.17 Time To Compile Seconds < Lower Is Better a . 163.78 |=================================================================== b . 164.21 |=================================================================== Blender 4.2 Blend File: Junkshop - Compute: CPU-Only Seconds < Lower Is Better a . 159.46 |=================================================================== b . 160.26 |=================================================================== Blender 4.2 Blend File: Fishy Cat - Compute: CPU-Only Seconds < Lower Is Better a . 159.34 |=================================================================== b . 160.43 |=================================================================== Whisperfile 20Aug24 Model Size: Small Seconds < Lower Is Better a . 156.75 |=================================================================== b . 156.83 |=================================================================== OSPRay 3.2 Benchmark: particle_volume/pathtracer/real_time Items Per Second > Higher Is Better a . 184.85 |=================================================================== b . 183.84 |=================================================================== SVT-AV1 2.2 Encoder Mode: Preset 13 - Input: Beauty 4K 10-bit Frames Per Second > Higher Is Better a . 8.898 |==================================================================== b . 8.887 |==================================================================== OSPRay 3.2 Benchmark: particle_volume/ao/real_time Items Per Second > Higher Is Better a . 4.74195 |================================================================== b . 4.72912 |================================================================== Blender 4.2 Blend File: BMW27 - Compute: CPU-Only Seconds < Lower Is Better a . 113.20 |=================================================================== b . 112.43 |=================================================================== SVT-AV1 2.2 Encoder Mode: Preset 5 - Input: Bosphorus 4K Frames Per Second > Higher Is Better a . 21.90 |==================================================================== b . 21.72 |=================================================================== GROMACS 2024 Implementation: MPI CPU - Input: water_GMX50_bare Ns Per Day > Higher Is Better a . 1.741 |==================================================================== b . 1.743 |==================================================================== Intel Open Image Denoise 2.3 Run: RTLightmap.hdr.4096x4096 - Device: CPU-Only Images / Sec > Higher Is Better a . 0.33 |===================================================================== b . 0.33 |===================================================================== simdjson 3.10 Throughput Test: Kostya GB/s > Higher Is Better a . 7.58 |===================================================================== b . 7.52 |==================================================================== Stockfish 17 Chess Benchmark Nodes Per Second > Higher Is Better a . 26172950 |================================================================= b . 21488082 |===================================================== simdjson 3.10 Throughput Test: LargeRandom GB/s > Higher Is Better a . 2.06 |===================================================================== b . 2.05 |===================================================================== OSPRay 3.2 Benchmark: gravity_spheres_volume/dim_512/scivis/real_time Items Per Second > Higher Is Better a . 3.74914 |================================================================== b . 3.74053 |================================================================== OSPRay 3.2 Benchmark: gravity_spheres_volume/dim_512/ao/real_time Items Per Second > Higher Is Better a . 3.83541 |================================================================== b . 3.84506 |================================================================== OSPRay 3.2 Benchmark: gravity_spheres_volume/dim_512/pathtracer/real_time Items Per Second > Higher Is Better a . 4.37639 |================================================================== b . 4.36447 |================================================================== simdjson 3.10 Throughput Test: DistinctUserID GB/s > Higher Is Better a . 13.37 |==================================================================== b . 12.73 |================================================================= simdjson 3.10 Throughput Test: PartialTweets GB/s > Higher Is Better a . 12.42 |==================================================================== b . 12.39 |==================================================================== simdjson 3.10 Throughput Test: TopTweet GB/s > Higher Is Better a . 13.1 |==================================================================== b . 13.2 |===================================================================== Stockfish Chess Benchmark Nodes Per Second > Higher Is Better a . 20554668 |================================================================= b . 20386602 |================================================================ ONNX Runtime 1.19 Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 1154.28 |================================================================== b . 1157.24 |================================================================== ONNX Runtime 1.19 Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 0.866341 |================================================================= b . 0.864063 |================================================================= ONNX Runtime 1.19 Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 765.88 |=================================================================== b . 763.76 |=================================================================== ONNX Runtime 1.19 Model: ResNet101_DUC_HDC-12 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 1.30568 |================================================================== b . 1.30931 |================================================================== ONNX Runtime 1.19 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 619.25 |=================================================================== b . 611.51 |================================================================== ONNX Runtime 1.19 Model: fcn-resnet101-11 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 1.61485 |================================================================= b . 1.63528 |================================================================== GraphicsMagick Operation: Sharpen Iterations Per Minute > Higher Is Better a . 19 |======================================================================= b . 19 |======================================================================= ONNX Runtime 1.19 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 356.00 |=================================================================== b . 354.36 |=================================================================== ONNX Runtime 1.19 Model: fcn-resnet101-11 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 2.80898 |================================================================== b . 2.82195 |================================================================== GraphicsMagick Operation: Enhanced Iterations Per Minute > Higher Is Better a . 50 |======================================================================= b . 50 |======================================================================= ONNX Runtime 1.19 Model: GPT-2 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 19.32 |==================================================================== b . 18.91 |=================================================================== ONNX Runtime 1.19 Model: GPT-2 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 51.75 |=================================================================== b . 52.89 |==================================================================== ONNX Runtime 1.19 Model: bertsquad-12 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 78.63 |=================================================================== b . 79.62 |==================================================================== ONNX Runtime 1.19 Model: bertsquad-12 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 12.72 |==================================================================== b . 12.56 |=================================================================== ONNX Runtime 1.19 Model: GPT-2 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 7.11087 |================================================================== b . 7.09276 |================================================================== ONNX Runtime 1.19 Model: GPT-2 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 140.54 |=================================================================== b . 140.90 |=================================================================== ONNX Runtime 1.19 Model: ZFNet-512 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 18.19 |==================================================================== b . 17.49 |================================================================= ONNX Runtime 1.19 Model: ZFNet-512 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 54.97 |================================================================= b . 57.16 |==================================================================== GraphicsMagick Operation: Noise-Gaussian Iterations Per Minute > Higher Is Better a . 68 |======================================================================= b . 67 |====================================================================== ONNX Runtime 1.19 Model: yolov4 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 117.70 |================================================================== b . 118.82 |=================================================================== ONNX Runtime 1.19 Model: yolov4 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 8.49570 |================================================================== b . 8.41617 |================================================================= ONNX Runtime 1.19 Model: ZFNet-512 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 8.81821 |================================================================== b . 8.71089 |================================================================= ONNX Runtime 1.19 Model: ZFNet-512 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 113.38 |================================================================== b . 114.77 |=================================================================== ONNX Runtime 1.19 Model: yolov4 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 66.05 |==================================================================== b . 65.28 |=================================================================== ONNX Runtime 1.19 Model: yolov4 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 15.14 |=================================================================== b . 15.32 |==================================================================== ONNX Runtime 1.19 Model: bertsquad-12 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 37.08 |==================================================================== b . 36.87 |==================================================================== ONNX Runtime 1.19 Model: bertsquad-12 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 26.97 |==================================================================== b . 27.12 |==================================================================== ONNX Runtime 1.19 Model: T5 Encoder - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 15.12 |=================================================================== b . 15.42 |==================================================================== ONNX Runtime 1.19 Model: T5 Encoder - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 66.13 |==================================================================== b . 64.84 |=================================================================== ONNX Runtime 1.19 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 42.73 |==================================================================== b . 41.52 |================================================================== ONNX Runtime 1.19 Model: ArcFace ResNet-100 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 23.40 |================================================================== b . 24.08 |==================================================================== ONNX Runtime 1.19 Model: T5 Encoder - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 6.11301 |================================================================= b . 6.18885 |================================================================== ONNX Runtime 1.19 Model: T5 Encoder - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 163.54 |=================================================================== b . 161.52 |================================================================== ONNX Runtime 1.19 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 19.20 |==================================================================== b . 19.22 |==================================================================== ONNX Runtime 1.19 Model: ArcFace ResNet-100 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 52.08 |==================================================================== b . 52.04 |==================================================================== ONNX Runtime 1.19 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 16.80 |=================================================================== b . 16.95 |==================================================================== ONNX Runtime 1.19 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 59.50 |==================================================================== b . 58.99 |=================================================================== ONNX Runtime 1.19 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 25.60 |==================================================================== b . 25.77 |==================================================================== ONNX Runtime 1.19 Model: Faster R-CNN R-50-FPN-int8 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 39.06 |==================================================================== b . 38.81 |==================================================================== GraphicsMagick Operation: Resizing Iterations Per Minute > Higher Is Better a . 283 |====================================================================== b . 235 |========================================================== ONNX Runtime 1.19 Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 5.76164 |================================================================== b . 5.79732 |================================================================== ONNX Runtime 1.19 Model: CaffeNet 12-int8 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 173.53 |=================================================================== b . 172.47 |=================================================================== ONNX Runtime 1.19 Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 1.14962 |================================================================== b . 1.15518 |================================================================== ONNX Runtime 1.19 Model: CaffeNet 12-int8 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 869.33 |=================================================================== b . 865.22 |=================================================================== GraphicsMagick Operation: HWB Color Space Iterations Per Minute > Higher Is Better a . 257 |====================================================================== b . 252 |===================================================================== ONNX Runtime 1.19 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 7.61827 |================================================================== b . 7.51917 |================================================================= ONNX Runtime 1.19 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 131.25 |================================================================== b . 132.98 |=================================================================== GraphicsMagick Operation: Rotate Iterations Per Minute > Higher Is Better a . 212 |====================================================================== b . 200 |================================================================== ONNX Runtime 1.19 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 1.86390 |================================================================== b . 1.87362 |================================================================== ONNX Runtime 1.19 Model: ResNet50 v1-12-int8 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 536.30 |=================================================================== b . 533.49 |=================================================================== ONNX Runtime 1.19 Model: super-resolution-10 - Device: CPU - Executor: Parallel Inference Time Cost (ms) < Lower Is Better a . 14.73 |==================================================================== b . 14.82 |==================================================================== ONNX Runtime 1.19 Model: super-resolution-10 - Device: CPU - Executor: Parallel Inferences Per Second > Higher Is Better a . 67.89 |==================================================================== b . 67.48 |==================================================================== ONNX Runtime 1.19 Model: super-resolution-10 - Device: CPU - Executor: Standard Inference Time Cost (ms) < Lower Is Better a . 4.86562 |================================================================== b . 4.83625 |================================================================== ONNX Runtime 1.19 Model: super-resolution-10 - Device: CPU - Executor: Standard Inferences Per Second > Higher Is Better a . 205.49 |=================================================================== b . 206.73 |=================================================================== GraphicsMagick Operation: Swirl Iterations Per Minute > Higher Is Better a . 126 |====================================================================== b . 125 |===================================================================== Mobile Neural Network 2.9.b11b7037d Model: inception-v3 ms < Lower Is Better a . 11.17 |==================================================================== b . 10.86 |================================================================== Mobile Neural Network 2.9.b11b7037d Model: mobilenet-v1-1.0 ms < Lower Is Better a . 1.218 |==================================================================== b . 1.214 |==================================================================== Mobile Neural Network 2.9.b11b7037d Model: MobileNetV2_224 ms < Lower Is Better a . 1.079 |==================================================================== b . 1.082 |==================================================================== Mobile Neural Network 2.9.b11b7037d Model: SqueezeNetV1.0 ms < Lower Is Better a . 1.491 |=================================================================== b . 1.510 |==================================================================== Mobile Neural Network 2.9.b11b7037d Model: resnet-v2-50 ms < Lower Is Better a . 6.904 |==================================================================== b . 6.837 |=================================================================== Mobile Neural Network 2.9.b11b7037d Model: squeezenetv1.1 ms < Lower Is Better a . 0.950 |==================================================================== b . 0.926 |================================================================== Mobile Neural Network 2.9.b11b7037d Model: mobilenetV3 ms < Lower Is Better a . 0.587 |==================================================================== b . 0.556 |================================================================ Mobile Neural Network 2.9.b11b7037d Model: nasnet ms < Lower Is Better a . 4.357 |==================================================================== b . 4.352 |==================================================================== SVT-AV1 2.2 Encoder Mode: Preset 5 - Input: Bosphorus 1080p Frames Per Second > Higher Is Better a . 71.96 |==================================================================== b . 71.82 |==================================================================== Etcpak 2.0 Benchmark: Multi-Threaded - Configuration: ETC2 Mpx/s > Higher Is Better a . 291.96 |================================================================== b . 294.20 |=================================================================== SVT-AV1 2.2 Encoder Mode: Preset 8 - Input: Bosphorus 4K Frames Per Second > Higher Is Better a . 50.95 |==================================================================== b . 50.65 |==================================================================== Intel Open Image Denoise 2.3 Run: RT.hdr_alb_nrm.3840x2160 - Device: CPU-Only Images / Sec > Higher Is Better a . 0.69 |===================================================================== b . 0.69 |===================================================================== Intel Open Image Denoise 2.3 Run: RT.ldr_alb_nrm.3840x2160 - Device: CPU-Only Images / Sec > Higher Is Better a . 0.69 |===================================================================== b . 0.69 |===================================================================== LZ4 Compression 1.10 Compression Level: 9 - Decompression Speed MB/s > Higher Is Better a . 6446.9 |=================================================================== b . 6478.6 |=================================================================== LZ4 Compression 1.10 Compression Level: 9 - Compression Speed MB/s > Higher Is Better a . 60.49 |==================================================================== b . 60.47 |==================================================================== LZ4 Compression 1.10 Compression Level: 1 - Decompression Speed MB/s > Higher Is Better a . 6687.6 |=================================================================== b . 6700.2 |=================================================================== LZ4 Compression 1.10 Compression Level: 1 - Compression Speed MB/s > Higher Is Better a . 1033.44 |================================================================== b . 1029.67 |================================================================== LZ4 Compression 1.10 Compression Level: 12 - Decompression Speed MB/s > Higher Is Better a . 6644.3 |=================================================================== b . 6643.3 |=================================================================== LZ4 Compression 1.10 Compression Level: 12 - Compression Speed MB/s > Higher Is Better a . 20.74 |==================================================================== b . 20.70 |==================================================================== LZ4 Compression 1.10 Compression Level: 3 - Decompression Speed MB/s > Higher Is Better a . 6167.8 |=================================================================== b . 6174.2 |=================================================================== LZ4 Compression 1.10 Compression Level: 3 - Compression Speed MB/s > Higher Is Better a . 168.65 |=================================================================== b . 168.61 |=================================================================== LZ4 Compression 1.10 Compression Level: 2 - Decompression Speed MB/s > Higher Is Better a . 5923.8 |=================================================================== b . 5922.1 |=================================================================== LZ4 Compression 1.10 Compression Level: 2 - Compression Speed MB/s > Higher Is Better a . 474.96 |=================================================================== b . 476.47 |=================================================================== POV-Ray Trace Time Seconds < Lower Is Better a . 38.06 |==================================================================== b . 38.07 |==================================================================== Z3 Theorem Prover 4.12.1 SMT File: 2.smt2 Seconds < Lower Is Better a . 36.99 |================================================================== b . 38.08 |==================================================================== Whisperfile 20Aug24 Model Size: Tiny Seconds < Lower Is Better a . 32.88 |==================================================================== b . 32.88 |==================================================================== x265 Video Input: Bosphorus 4K Frames Per Second > Higher Is Better a . 18.3 |==================================================================== b . 18.6 |===================================================================== Y-Cruncher 0.8.5 Pi Digits To Calculate: 1B Seconds < Lower Is Better a . 23.42 |=================================================================== b . 23.65 |==================================================================== SVT-AV1 2.2 Encoder Mode: Preset 8 - Input: Bosphorus 1080p Frames Per Second > Higher Is Better a . 171.00 |=================================================================== b . 170.25 |=================================================================== Z3 Theorem Prover 4.12.1 SMT File: 1.smt2 Seconds < Lower Is Better a . 16.40 |=============================================================== b . 17.81 |==================================================================== SVT-AV1 2.2 Encoder Mode: Preset 13 - Input: Bosphorus 4K Frames Per Second > Higher Is Better a . 167.00 |=================================================================== b . 166.59 |=================================================================== Y-Cruncher 0.8.5 Pi Digits To Calculate: 500M Seconds < Lower Is Better a . 10.088 |=================================================================== b . 9.946 |================================================================== x265 Video Input: Bosphorus 1080p Frames Per Second > Higher Is Better a . 96.00 |=================================================================== b . 97.25 |==================================================================== SVT-AV1 2.2 Encoder Mode: Preset 13 - Input: Bosphorus 1080p Frames Per Second > Higher Is Better a . 706.48 |=================================================================== b . 706.92 |===================================================================