Desktop machine learning

AMD Ryzen 9 3900X 12-Core testing with a MSI X570-A PRO (MS-7C37) v3.0 (H.70 BIOS) and NVIDIA GeForce RTX 3060 12GB on Ubuntu 23.10 via the Phoronix Test Suite.

HTML result view exported from: https://openbenchmarking.org/result/2405015-VPA1-DESKTOP46&sro&grw.

Desktop machine learningProcessorMotherboardChipsetMemoryDiskGraphicsAudioMonitorNetworkOSKernelDisplay ServerDisplay DriverOpenCLCompilerFile-SystemScreen Resolutionmanticmantic-no-omit-framepointernobleAMD Ryzen 9 3900X 12-Core @ 3.80GHz (12 Cores / 24 Threads)MSI X570-A PRO (MS-7C37) v3.0 (H.70 BIOS)AMD Starship/Matisse2 x 16GB DDR4-3200MT/s F4-3200C16-16GVK2000GB Seagate ST2000DM006-2DM1 + 2000GB Western Digital WD20EZAZ-00G + 500GB Samsung SSD 860 + 8002GB Seagate ST8000DM004-2CX1 + 1000GB CT1000BX500SSD1 + 512GB TS512GESD310CNVIDIA GeForce RTX 3060 12GBNVIDIA GA104 HD AudioDELL P2314HRealtek RTL8111/8168/8411Ubuntu 23.106.5.0-9-generic (x86_64)X Server 1.21.1.7NVIDIAOpenCL 3.0 CUDA 12.2.146GCC 13.2.0 + CUDA 12.2ext41920x1080NVIDIA GeForce RTX 3060DELL P2314H + U32J59xRealtek RTL8111/8168/8211/8411Ubuntu 24.046.8.0-31-generic (x86_64)GCC 13.2.0OpenBenchmarking.orgKernel Details- Transparent Huge Pages: madviseCompiler Details- mantic: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-XYspKM/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - mantic-no-omit-framepointer: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-bootstrap --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-link-serialization=2 --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-b9QCDx/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-b9QCDx/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-build-config=bootstrap-lto-lean --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - noble: --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-cet --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,d,fortran,objc,obj-c++,m2 --enable-libphobos-checking=release --enable-libstdcxx-backtrace --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-defaulted --enable-offload-targets=nvptx-none=/build/gcc-13-uJ7kn6/gcc-13-13.2.0/debian/tmp-nvptx/usr,amdgcn-amdhsa=/build/gcc-13-uJ7kn6/gcc-13-13.2.0/debian/tmp-gcn/usr --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v Processor Details- Scaling Governor: acpi-cpufreq schedutil (Boost: Enabled) - CPU Microcode: 0x8701013Python Details- mantic: Python 3.11.6- mantic-no-omit-framepointer: Python 3.11.6- noble: Python 3.12.3Security Details- mantic: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Mitigation of untrained return thunk; SMT enabled with STIBP protection + spec_rstack_overflow: Mitigation of safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected - mantic-no-omit-framepointer: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Mitigation of untrained return thunk; SMT enabled with STIBP protection + spec_rstack_overflow: Mitigation of safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines IBPB: conditional STIBP: always-on RSB filling PBRSB-eIBRS: Not affected + srbds: Not affected + tsx_async_abort: Not affected - noble: gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + reg_file_data_sampling: Not affected + retbleed: Mitigation of untrained return thunk; SMT enabled with STIBP protection + spec_rstack_overflow: Mitigation of Safe RET + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Retpolines; IBPB: conditional; STIBP: always-on; RSB filling; PBRSB-eIBRS: Not affected; BHI: Not affected + srbds: Not affected + tsx_async_abort: Not affected Environment Details- mantic-no-omit-framepointer: CXXFLAGS=-fno-omit-frame-pointer QMAKE_CFLAGS=-fno-omit-frame-pointer CFLAGS=-fno-omit-frame-pointer CFLAGS_OVERRIDE=-fno-omit-frame-pointer QMAKE_CXXFLAGS=-fno-omit-frame-pointer FFLAGS=-fno-omit-frame-pointer - noble: CXXFLAGS="-fno-omit-frame-pointer -frecord-gcc-switches -O2" QMAKE_CFLAGS="-fno-omit-frame-pointer -frecord-gcc-switches -O2" CFLAGS="-fno-omit-frame-pointer -frecord-gcc-switches -O2" CFLAGS_OVERRIDE="-fno-omit-frame-pointer -frecord-gcc-switches -O2" QMAKE_CXXFLAGS="-fno-omit-frame-pointer -frecord-gcc-switches -O2" FFLAGS="-fno-omit-frame-pointer -frecord-gcc-switches -O2"

Desktop machine learningscikit-learn: GLMscikit-learn: SAGAscikit-learn: Treescikit-learn: Lassoscikit-learn: Sparsifyscikit-learn: Plot Wardscikit-learn: MNIST Datasetscikit-learn: Plot Neighborsscikit-learn: SGD Regressionscikit-learn: SGDOneClassSVMscikit-learn: Isolation Forestscikit-learn: Text Vectorizersscikit-learn: Plot Hierarchicalscikit-learn: Plot OMP vs. LARSscikit-learn: Feature Expansionsscikit-learn: LocalOutlierFactorscikit-learn: TSNE MNIST Datasetscikit-learn: Isotonic / Logisticscikit-learn: Plot Incremental PCAscikit-learn: Hist Gradient Boostingscikit-learn: Sample Without Replacementscikit-learn: Covertype Dataset Benchmarkscikit-learn: Hist Gradient Boosting Adultscikit-learn: Isotonic / Perturbed Logarithmscikit-learn: Hist Gradient Boosting Threadingscikit-learn: 20 Newsgroups / Logistic Regressionscikit-learn: Plot Polynomial Kernel Approximationscikit-learn: Hist Gradient Boosting Categorical Onlyscikit-learn: Kernel PCA Solvers / Time vs. N Samplesscikit-learn: Kernel PCA Solvers / Time vs. N Componentsscikit-learn: Sparse Rand Projections / 100 Iterationsnumpy: pytorch: CPU - 1 - ResNet-50pytorch: CPU - 1 - ResNet-152pytorch: CPU - 16 - ResNet-50pytorch: CPU - 32 - ResNet-50pytorch: CPU - 64 - ResNet-50pytorch: CPU - 16 - ResNet-152pytorch: CPU - 256 - ResNet-50pytorch: CPU - 32 - ResNet-152pytorch: CPU - 512 - ResNet-50pytorch: CPU - 64 - ResNet-152pytorch: CPU - 256 - ResNet-152pytorch: CPU - 512 - ResNet-152pytorch: CPU - 1 - Efficientnet_v2_lpytorch: CPU - 16 - Efficientnet_v2_lpytorch: CPU - 32 - Efficientnet_v2_lpytorch: CPU - 64 - Efficientnet_v2_lpytorch: CPU - 256 - Efficientnet_v2_lpytorch: CPU - 512 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 1 - ResNet-50pytorch: NVIDIA CUDA GPU - 1 - ResNet-152pytorch: NVIDIA CUDA GPU - 16 - ResNet-50pytorch: NVIDIA CUDA GPU - 32 - ResNet-50pytorch: NVIDIA CUDA GPU - 64 - ResNet-50pytorch: NVIDIA CUDA GPU - 16 - ResNet-152pytorch: NVIDIA CUDA GPU - 256 - ResNet-50pytorch: NVIDIA CUDA GPU - 32 - ResNet-152pytorch: NVIDIA CUDA GPU - 512 - ResNet-50pytorch: NVIDIA CUDA GPU - 64 - ResNet-152pytorch: NVIDIA CUDA GPU - 256 - ResNet-152pytorch: NVIDIA CUDA GPU - 512 - ResNet-152pytorch: NVIDIA CUDA GPU - 1 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 16 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 32 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 64 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 256 - Efficientnet_v2_lpytorch: NVIDIA CUDA GPU - 512 - Efficientnet_v2_lpyhpc: CPU - Numpy - 16384 - Equation of Statepyhpc: CPU - Numpy - 16384 - Isoneutral Mixingpyhpc: CPU - Numpy - 65536 - Equation of Statepyhpc: CPU - Numpy - 65536 - Isoneutral Mixingpyhpc: GPU - Numpy - 16384 - Equation of Statepyhpc: GPU - Numpy - 16384 - Isoneutral Mixingpyhpc: GPU - Numpy - 65536 - Equation of Statepyhpc: GPU - Numpy - 65536 - Isoneutral Mixingpyhpc: CPU - Numpy - 262144 - Equation of Statepyhpc: CPU - Numpy - 262144 - Isoneutral Mixingpyhpc: GPU - Numpy - 262144 - Equation of Statepyhpc: GPU - Numpy - 262144 - Isoneutral Mixingpyhpc: CPU - Numpy - 1048576 - Equation of Statepyhpc: CPU - Numpy - 1048576 - Isoneutral Mixingpyhpc: CPU - Numpy - 4194304 - Equation of Statepyhpc: CPU - Numpy - 4194304 - Isoneutral Mixingpyhpc: GPU - Numpy - 1048576 - Equation of Statepyhpc: GPU - Numpy - 1048576 - Isoneutral Mixingpyhpc: GPU - Numpy - 4194304 - Equation of Statepyhpc: GPU - Numpy - 4194304 - Isoneutral Mixingpyperformance: gopyperformance: 2to3pyperformance: chaospyperformance: floatpyperformance: nbodypyperformance: pathlibpyperformance: raytracepyperformance: json_loadspyperformance: crypto_pyaespyperformance: regex_compilepyperformance: python_startuppyperformance: django_templatepyperformance: pickle_pure_pythonpybench: Total For Average Test Timesmanticmantic-no-omit-framepointernoble293.598868.01848.338511.848127.28257.82465.763147.752106.315379.739289.37160.814211.28691.499131.27753.464236.8651470.80631.006109.984158.262376.145103.4971788.259110.21541.519150.73218.57972.54137.242613.547426.2832.3612.7224.2824.2924.249.8824.429.8424.139.889.779.877.315.635.635.625.615.61210.8873.91200.30199.46201.4173.01202.7274.15203.1871.8171.7472.3139.3538.9537.7137.8837.3637.430.0030.0090.0150.0320.0030.0090.0150.0330.0610.1310.0620.1310.2630.6191.4022.6700.2630.6311.4222.66212922162.867.476.219.726219.565.11167.6128.5259774295.096873.82252.969509.537125.44257.54565.877142.451107.527382.611336.37263.875208.39192.582133.09256.754236.7861471.83431.057111.255161.460370.694105.6471828.300110.37441.728150.37618.86572.90937.889631.071428.6132.5412.7824.3824.3524.409.9324.3710.0024.289.919.919.807.325.645.645.655.645.65211.4672.27200.17202.68205.9572.24203.2273.36201.1473.6572.9173.7537.2936.1037.1637.2436.6037.220.0030.0080.0150.0320.0020.0080.0150.0330.0580.1320.0580.1280.2620.6181.4052.6260.2600.6221.4112.62013122463.666.977.120.227420.866.61207.6429.5263790269.806869.36947.033345.400125.06956.13265.416142.15978.880385.383314.03466.393207.10468.172133.14454.288285.8231684.54630.617117.407179.638381.447112.7131963.772111.55441.914145.36319.93270.02237.107663.953430.8332.3412.8924.4324.1224.199.8824.339.8124.309.879.869.877.315.595.595.605.615.600.0030.0090.0160.0330.0030.0080.0150.0340.0600.1330.0610.1360.2610.6311.4362.7200.2620.6301.4462.66812122.88.76839OpenBenchmarking.org

Scikit-Learn

Benchmark: GLM

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: GLMmanticmantic-no-omit-framepointernoble60120180240300SE +/- 1.06, N = 3SE +/- 1.07, N = 3SE +/- 0.93, N = 3293.60295.10269.81-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: SAGA

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: SAGAmanticmantic-no-omit-framepointernoble2004006008001000SE +/- 8.69, N = 6SE +/- 5.60, N = 3SE +/- 10.35, N = 3868.02873.82869.37-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Tree

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Treemanticmantic-no-omit-framepointernoble1224364860SE +/- 0.59, N = 4SE +/- 0.48, N = 15SE +/- 0.52, N = 348.3452.9747.03-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Lasso

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Lassomanticmantic-no-omit-framepointernoble110220330440550SE +/- 3.22, N = 3SE +/- 3.50, N = 3SE +/- 1.37, N = 3511.85509.54345.40-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Sparsify

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Sparsifymanticmantic-no-omit-framepointernoble306090120150SE +/- 1.36, N = 5SE +/- 1.28, N = 5SE +/- 0.65, N = 3127.28125.44125.07-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Plot Ward

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Plot Wardmanticmantic-no-omit-framepointernoble1326395265SE +/- 0.21, N = 3SE +/- 0.22, N = 3SE +/- 0.20, N = 357.8257.5556.13-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: MNIST Dataset

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: MNIST Datasetmanticmantic-no-omit-framepointernoble1530456075SE +/- 0.82, N = 4SE +/- 0.47, N = 3SE +/- 0.67, N = 365.7665.8865.42-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Plot Neighbors

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Plot Neighborsmanticmantic-no-omit-framepointernoble306090120150SE +/- 1.34, N = 7SE +/- 0.59, N = 3SE +/- 1.09, N = 3147.75142.45142.16-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: SGD Regression

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: SGD Regressionmanticmantic-no-omit-framepointernoble20406080100SE +/- 1.06, N = 6SE +/- 0.49, N = 3SE +/- 0.05, N = 3106.32107.5378.88-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: SGDOneClassSVM

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: SGDOneClassSVMmanticmantic-no-omit-framepointernoble80160240320400SE +/- 4.18, N = 3SE +/- 3.48, N = 7SE +/- 3.55, N = 3379.74382.61385.38-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Isolation Forest

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Isolation Forestmanticmantic-no-omit-framepointernoble70140210280350SE +/- 1.30, N = 3SE +/- 51.04, N = 9SE +/- 2.83, N = 3289.37336.37314.03-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Text Vectorizers

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Text Vectorizersmanticmantic-no-omit-framepointernoble1530456075SE +/- 0.19, N = 3SE +/- 0.08, N = 3SE +/- 0.32, N = 360.8163.8866.39-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Plot Hierarchical

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Plot Hierarchicalmanticmantic-no-omit-framepointernoble50100150200250SE +/- 0.75, N = 3SE +/- 0.42, N = 3SE +/- 2.35, N = 3211.29208.39207.10-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Plot OMP vs. LARS

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Plot OMP vs. LARSmanticmantic-no-omit-framepointernoble20406080100SE +/- 0.08, N = 3SE +/- 0.44, N = 3SE +/- 0.03, N = 391.5092.5868.17-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Feature Expansions

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Feature Expansionsmanticmantic-no-omit-framepointernoble306090120150SE +/- 0.86, N = 3SE +/- 1.22, N = 3SE +/- 1.21, N = 3131.28133.09133.14-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: LocalOutlierFactor

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: LocalOutlierFactormanticmantic-no-omit-framepointernoble1326395265SE +/- 0.18, N = 3SE +/- 0.74, N = 15SE +/- 0.02, N = 353.4656.7554.29-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: TSNE MNIST Dataset

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: TSNE MNIST Datasetmanticmantic-no-omit-framepointernoble60120180240300SE +/- 0.44, N = 3SE +/- 0.54, N = 3SE +/- 0.91, N = 3236.87236.79285.82-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Isotonic / Logistic

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Isotonic / Logisticmanticmantic-no-omit-framepointernoble400800120016002000SE +/- 12.29, N = 3SE +/- 14.46, N = 3SE +/- 9.43, N = 31470.811471.831684.55-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Plot Incremental PCA

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Plot Incremental PCAmanticmantic-no-omit-framepointernoble714212835SE +/- 0.03, N = 3SE +/- 0.07, N = 3SE +/- 0.06, N = 331.0131.0630.62-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Hist Gradient Boosting

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Hist Gradient Boostingmanticmantic-no-omit-framepointernoble306090120150SE +/- 0.22, N = 3SE +/- 0.25, N = 3SE +/- 0.17, N = 3109.98111.26117.41-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Sample Without Replacement

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Sample Without Replacementmanticmantic-no-omit-framepointernoble4080120160200SE +/- 0.60, N = 3SE +/- 0.62, N = 3SE +/- 2.21, N = 3158.26161.46179.64-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Covertype Dataset Benchmark

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Covertype Dataset Benchmarkmanticmantic-no-omit-framepointernoble80160240320400SE +/- 4.88, N = 3SE +/- 3.40, N = 3SE +/- 2.58, N = 3376.15370.69381.45-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Hist Gradient Boosting Adult

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Hist Gradient Boosting Adultmanticmantic-no-omit-framepointernoble306090120150SE +/- 0.70, N = 3SE +/- 0.59, N = 3SE +/- 0.52, N = 3103.50105.65112.71-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Isotonic / Perturbed Logarithm

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Isotonic / Perturbed Logarithmmanticmantic-no-omit-framepointernoble400800120016002000SE +/- 24.41, N = 3SE +/- 16.46, N = 3SE +/- 1.48, N = 31788.261828.301963.77-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Hist Gradient Boosting Threading

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Hist Gradient Boosting Threadingmanticmantic-no-omit-framepointernoble20406080100SE +/- 0.13, N = 3SE +/- 0.15, N = 3SE +/- 0.13, N = 3110.22110.37111.55-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: 20 Newsgroups / Logistic Regression

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: 20 Newsgroups / Logistic Regressionmanticmantic-no-omit-framepointernoble1020304050SE +/- 0.19, N = 3SE +/- 0.24, N = 3SE +/- 0.12, N = 341.5241.7341.91-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Plot Polynomial Kernel Approximation

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Plot Polynomial Kernel Approximationmanticmantic-no-omit-framepointernoble306090120150SE +/- 1.22, N = 3SE +/- 1.20, N = 3SE +/- 1.46, N = 3150.73150.38145.36-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Hist Gradient Boosting Categorical Only

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Hist Gradient Boosting Categorical Onlymanticmantic-no-omit-framepointernoble510152025SE +/- 0.06, N = 3SE +/- 0.12, N = 3SE +/- 0.10, N = 318.5818.8719.93-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Kernel PCA Solvers / Time vs. N Samples

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Kernel PCA Solvers / Time vs. N Samplesmanticmantic-no-omit-framepointernoble1632486480SE +/- 0.05, N = 3SE +/- 0.16, N = 3SE +/- 0.44, N = 372.5472.9170.02-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Kernel PCA Solvers / Time vs. N Components

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Kernel PCA Solvers / Time vs. N Componentsmanticmantic-no-omit-framepointernoble918273645SE +/- 0.21, N = 3SE +/- 0.36, N = 3SE +/- 0.43, N = 337.2437.8937.11-O21. (F9X) gfortran options: -O0

Scikit-Learn

Benchmark: Sparse Random Projections / 100 Iterations

OpenBenchmarking.orgSeconds, Fewer Is BetterScikit-Learn 1.2.2Benchmark: Sparse Random Projections / 100 Iterationsmanticmantic-no-omit-framepointernoble140280420560700SE +/- 3.80, N = 3SE +/- 7.06, N = 4SE +/- 4.34, N = 3613.55631.07663.95-O21. (F9X) gfortran options: -O0

Numpy Benchmark

OpenBenchmarking.orgScore, More Is BetterNumpy Benchmarkmanticmantic-no-omit-framepointernoble90180270360450SE +/- 1.20, N = 3SE +/- 0.90, N = 3SE +/- 1.01, N = 3426.28428.61430.83

PyTorch

Device: CPU - Batch Size: 1 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 1 - Model: ResNet-50manticmantic-no-omit-framepointernoble816243240SE +/- 0.11, N = 3SE +/- 0.16, N = 3SE +/- 0.17, N = 332.3632.5432.34MIN: 31.89 / MAX: 32.7MIN: 31.64 / MAX: 32.94MIN: 28.9 / MAX: 32.83

PyTorch

Device: CPU - Batch Size: 1 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 1 - Model: ResNet-152manticmantic-no-omit-framepointernoble3691215SE +/- 0.03, N = 3SE +/- 0.04, N = 3SE +/- 0.05, N = 312.7212.7812.89MIN: 11.99 / MAX: 12.8MIN: 11.9 / MAX: 12.9MIN: 12.36 / MAX: 13.05

PyTorch

Device: CPU - Batch Size: 16 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 16 - Model: ResNet-50manticmantic-no-omit-framepointernoble612182430SE +/- 0.05, N = 3SE +/- 0.16, N = 3SE +/- 0.01, N = 324.2824.3824.43MIN: 20.22 / MAX: 24.56MIN: 22.2 / MAX: 24.87MIN: 22.57 / MAX: 24.72

PyTorch

Device: CPU - Batch Size: 32 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 32 - Model: ResNet-50manticmantic-no-omit-framepointernoble612182430SE +/- 0.10, N = 3SE +/- 0.16, N = 3SE +/- 0.06, N = 324.2924.3524.12MIN: 22.24 / MAX: 24.66MIN: 23.67 / MAX: 24.87MIN: 22.33 / MAX: 24.46

PyTorch

Device: CPU - Batch Size: 64 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 64 - Model: ResNet-50manticmantic-no-omit-framepointernoble612182430SE +/- 0.04, N = 3SE +/- 0.15, N = 3SE +/- 0.11, N = 324.2424.4024.19MIN: 23.59 / MAX: 24.49MIN: 21.6 / MAX: 24.8MIN: 22.75 / MAX: 24.73

PyTorch

Device: CPU - Batch Size: 16 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 16 - Model: ResNet-152manticmantic-no-omit-framepointernoble3691215SE +/- 0.04, N = 3SE +/- 0.01, N = 3SE +/- 0.02, N = 39.889.939.88MIN: 9.31 / MAX: 10.01MIN: 9.39 / MAX: 10.01MIN: 9.15 / MAX: 9.98

PyTorch

Device: CPU - Batch Size: 256 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 256 - Model: ResNet-50manticmantic-no-omit-framepointernoble612182430SE +/- 0.03, N = 3SE +/- 0.11, N = 3SE +/- 0.06, N = 324.4224.3724.33MIN: 20.15 / MAX: 24.74MIN: 23.76 / MAX: 24.81MIN: 22.79 / MAX: 24.66

PyTorch

Device: CPU - Batch Size: 32 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 32 - Model: ResNet-152manticmantic-no-omit-framepointernoble3691215SE +/- 0.05, N = 3SE +/- 0.09, N = 3SE +/- 0.04, N = 39.8410.009.81MIN: 9.6 / MAX: 9.98MIN: 8.09 / MAX: 10.27MIN: 9.42 / MAX: 9.93

PyTorch

Device: CPU - Batch Size: 512 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 512 - Model: ResNet-50manticmantic-no-omit-framepointernoble612182430SE +/- 0.02, N = 3SE +/- 0.08, N = 3SE +/- 0.14, N = 324.1324.2824.30MIN: 23.58 / MAX: 24.41MIN: 22.31 / MAX: 24.53MIN: 22.45 / MAX: 24.75

PyTorch

Device: CPU - Batch Size: 64 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 64 - Model: ResNet-152manticmantic-no-omit-framepointernoble3691215SE +/- 0.03, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 39.889.919.87MIN: 8.8 / MAX: 9.98MIN: 8.69 / MAX: 10.08MIN: 8.61 / MAX: 9.96

PyTorch

Device: CPU - Batch Size: 256 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 256 - Model: ResNet-152manticmantic-no-omit-framepointernoble3691215SE +/- 0.07, N = 3SE +/- 0.04, N = 3SE +/- 0.03, N = 39.779.919.86MIN: 9.17 / MAX: 10MIN: 9.19 / MAX: 10.05MIN: 8.69 / MAX: 9.99

PyTorch

Device: CPU - Batch Size: 512 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 512 - Model: ResNet-152manticmantic-no-omit-framepointernoble3691215SE +/- 0.02, N = 3SE +/- 0.07, N = 3SE +/- 0.03, N = 39.879.809.87MIN: 9.09 / MAX: 9.96MIN: 9.12 / MAX: 9.98MIN: 9.21 / MAX: 10

PyTorch

Device: CPU - Batch Size: 1 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 1 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointernoble246810SE +/- 0.00, N = 3SE +/- 0.02, N = 3SE +/- 0.00, N = 37.317.327.31MIN: 7.16 / MAX: 7.34MIN: 7.23 / MAX: 7.38MIN: 7.07 / MAX: 7.36

PyTorch

Device: CPU - Batch Size: 16 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 16 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointernoble1.2692.5383.8075.0766.345SE +/- 0.02, N = 3SE +/- 0.01, N = 3SE +/- 0.02, N = 35.635.645.59MIN: 5.39 / MAX: 5.71MIN: 5.45 / MAX: 5.68MIN: 5.31 / MAX: 5.65

PyTorch

Device: CPU - Batch Size: 32 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 32 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointernoble1.2692.5383.8075.0766.345SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.00, N = 35.635.645.59MIN: 5.31 / MAX: 5.68MIN: 5.52 / MAX: 5.69MIN: 5.46 / MAX: 5.64

PyTorch

Device: CPU - Batch Size: 64 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 64 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointernoble1.27132.54263.81395.08526.3565SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 35.625.655.60MIN: 5.35 / MAX: 5.66MIN: 5.45 / MAX: 5.7MIN: 5.32 / MAX: 5.64

PyTorch

Device: CPU - Batch Size: 256 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 256 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointernoble1.2692.5383.8075.0766.345SE +/- 0.02, N = 3SE +/- 0.01, N = 3SE +/- 0.02, N = 35.615.645.61MIN: 5.44 / MAX: 5.65MIN: 5.29 / MAX: 5.68MIN: 5.46 / MAX: 5.67

PyTorch

Device: CPU - Batch Size: 512 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: CPU - Batch Size: 512 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointernoble1.27132.54263.81395.08526.3565SE +/- 0.01, N = 3SE +/- 0.02, N = 3SE +/- 0.01, N = 35.615.655.60MIN: 5.45 / MAX: 5.66MIN: 5.36 / MAX: 5.93MIN: 5.37 / MAX: 5.66

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 1 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 1 - Model: ResNet-50manticmantic-no-omit-framepointer50100150200250SE +/- 2.67, N = 3SE +/- 1.46, N = 15210.88211.46MIN: 195.21 / MAX: 218.16MIN: 192.13 / MAX: 223.01

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 1 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 1 - Model: ResNet-152manticmantic-no-omit-framepointer1632486480SE +/- 0.56, N = 3SE +/- 0.96, N = 373.9172.27MIN: 68.9 / MAX: 75.9MIN: 68.86 / MAX: 76.62

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 16 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 16 - Model: ResNet-50manticmantic-no-omit-framepointer4080120160200SE +/- 0.25, N = 3SE +/- 0.96, N = 3200.30200.17MIN: 182.88 / MAX: 202.36MIN: 183.43 / MAX: 203.55

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 32 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 32 - Model: ResNet-50manticmantic-no-omit-framepointer4080120160200SE +/- 1.06, N = 3SE +/- 2.52, N = 4199.46202.68MIN: 182.77 / MAX: 206.03MIN: 182.69 / MAX: 211.53

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 64 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 64 - Model: ResNet-50manticmantic-no-omit-framepointer50100150200250SE +/- 0.58, N = 3SE +/- 1.98, N = 3201.41205.95MIN: 184.02 / MAX: 203.68MIN: 186.96 / MAX: 210.21

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 16 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 16 - Model: ResNet-152manticmantic-no-omit-framepointer1632486480SE +/- 0.96, N = 3SE +/- 0.20, N = 373.0172.24MIN: 68.06 / MAX: 75.3MIN: 68.36 / MAX: 73.14

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 256 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 256 - Model: ResNet-50manticmantic-no-omit-framepointer4080120160200SE +/- 1.76, N = 3SE +/- 1.21, N = 3202.72203.22MIN: 183.1 / MAX: 207.93MIN: 185.88 / MAX: 206.71

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 32 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 32 - Model: ResNet-152manticmantic-no-omit-framepointer1632486480SE +/- 0.96, N = 3SE +/- 0.74, N = 374.1573.36MIN: 68.27 / MAX: 75.61MIN: 68.19 / MAX: 74.63

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 512 - Model: ResNet-50

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 512 - Model: ResNet-50manticmantic-no-omit-framepointer4080120160200SE +/- 1.69, N = 3SE +/- 0.33, N = 3203.18201.14MIN: 183.76 / MAX: 207.98MIN: 183.61 / MAX: 202.73

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 64 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 64 - Model: ResNet-152manticmantic-no-omit-framepointer1632486480SE +/- 0.44, N = 3SE +/- 0.66, N = 371.8173.65MIN: 67.31 / MAX: 72.89MIN: 68.88 / MAX: 75.03

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 256 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 256 - Model: ResNet-152manticmantic-no-omit-framepointer1632486480SE +/- 0.24, N = 3SE +/- 0.83, N = 371.7472.91MIN: 67.87 / MAX: 72.6MIN: 68 / MAX: 75.45

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 512 - Model: ResNet-152

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 512 - Model: ResNet-152manticmantic-no-omit-framepointer1632486480SE +/- 0.94, N = 3SE +/- 0.50, N = 372.3173.75MIN: 67.38 / MAX: 74.62MIN: 68.91 / MAX: 75.15

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 1 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 1 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointer918273645SE +/- 0.47, N = 3SE +/- 0.26, N = 339.3537.29MIN: 36.65 / MAX: 40.42MIN: 35.83 / MAX: 39.17

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 16 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 16 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointer918273645SE +/- 0.08, N = 3SE +/- 0.02, N = 338.9536.10MIN: 37.12 / MAX: 39.27MIN: 34.25 / MAX: 38.01

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 32 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 32 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointer918273645SE +/- 0.24, N = 3SE +/- 0.30, N = 1537.7137.16MIN: 35.52 / MAX: 38.25MIN: 34.12 / MAX: 39.48

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 64 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 64 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointer918273645SE +/- 0.30, N = 9SE +/- 0.31, N = 1537.8837.24MIN: 35.67 / MAX: 39.63MIN: 33.97 / MAX: 39.43

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 256 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 256 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointer918273645SE +/- 0.15, N = 3SE +/- 0.30, N = 1537.3636.60MIN: 35.47 / MAX: 37.85MIN: 33.07 / MAX: 39.53

PyTorch

Device: NVIDIA CUDA GPU - Batch Size: 512 - Model: Efficientnet_v2_l

OpenBenchmarking.orgbatches/sec, More Is BetterPyTorch 2.1Device: NVIDIA CUDA GPU - Batch Size: 512 - Model: Efficientnet_v2_lmanticmantic-no-omit-framepointer918273645SE +/- 0.03, N = 3SE +/- 0.33, N = 837.4337.22MIN: 35.81 / MAX: 38.02MIN: 34.99 / MAX: 39.08

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 16384 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 16384 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.00070.00140.00210.00280.0035SE +/- 0.000, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 30.0030.0030.003

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 16384 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 16384 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.0020.0040.0060.0080.01SE +/- 0.000, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 30.0090.0080.009

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 65536 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 65536 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.00360.00720.01080.01440.018SE +/- 0.000, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 150.0150.0150.016

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 65536 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 65536 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.00740.01480.02220.02960.037SE +/- 0.000, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 30.0320.0320.033

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 16384 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 16384 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.00070.00140.00210.00280.0035SE +/- 0.000, N = 3SE +/- 0.000, N = 15SE +/- 0.000, N = 120.0030.0020.003

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 16384 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 16384 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.0020.0040.0060.0080.01SE +/- 0.000, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 30.0090.0080.008

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 65536 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 65536 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.00340.00680.01020.01360.017SE +/- 0.000, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 70.0150.0150.015

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 65536 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 65536 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.00770.01540.02310.03080.0385SE +/- 0.000, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 30.0330.0330.034

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 262144 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 262144 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.01370.02740.04110.05480.0685SE +/- 0.001, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 30.0610.0580.060

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 262144 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 262144 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.02990.05980.08970.11960.1495SE +/- 0.001, N = 3SE +/- 0.000, N = 3SE +/- 0.002, N = 30.1310.1320.133

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 262144 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 262144 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.0140.0280.0420.0560.07SE +/- 0.001, N = 3SE +/- 0.000, N = 3SE +/- 0.000, N = 30.0620.0580.061

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 262144 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 262144 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.03060.06120.09180.12240.153SE +/- 0.000, N = 3SE +/- 0.001, N = 3SE +/- 0.001, N = 30.1310.1280.136

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 1048576 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 1048576 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.05920.11840.17760.23680.296SE +/- 0.002, N = 3SE +/- 0.000, N = 3SE +/- 0.002, N = 30.2630.2620.261

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 1048576 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 1048576 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.1420.2840.4260.5680.71SE +/- 0.001, N = 3SE +/- 0.000, N = 3SE +/- 0.006, N = 30.6190.6180.631

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 4194304 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 4194304 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.32310.64620.96931.29241.6155SE +/- 0.003, N = 3SE +/- 0.004, N = 3SE +/- 0.003, N = 31.4021.4051.436

PyHPC Benchmarks

Device: CPU - Backend: Numpy - Project Size: 4194304 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: CPU - Backend: Numpy - Project Size: 4194304 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.6121.2241.8362.4483.06SE +/- 0.010, N = 3SE +/- 0.002, N = 3SE +/- 0.010, N = 32.6702.6262.720

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 1048576 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 1048576 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.05920.11840.17760.23680.296SE +/- 0.002, N = 3SE +/- 0.001, N = 3SE +/- 0.002, N = 30.2630.2600.262

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 1048576 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 1048576 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.1420.2840.4260.5680.71SE +/- 0.002, N = 3SE +/- 0.007, N = 3SE +/- 0.004, N = 30.6310.6220.630

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 4194304 - Benchmark: Equation of State

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 4194304 - Benchmark: Equation of Statemanticmantic-no-omit-framepointernoble0.32540.65080.97621.30161.627SE +/- 0.004, N = 3SE +/- 0.001, N = 3SE +/- 0.006, N = 31.4221.4111.446

PyHPC Benchmarks

Device: GPU - Backend: Numpy - Project Size: 4194304 - Benchmark: Isoneutral Mixing

OpenBenchmarking.orgSeconds, Fewer Is BetterPyHPC Benchmarks 3.0Device: GPU - Backend: Numpy - Project Size: 4194304 - Benchmark: Isoneutral Mixingmanticmantic-no-omit-framepointernoble0.60031.20061.80092.40123.0015SE +/- 0.006, N = 3SE +/- 0.006, N = 3SE +/- 0.006, N = 32.6622.6202.668

PyPerformance

Benchmark: go

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: gomanticmantic-no-omit-framepointernoble306090120150SE +/- 0.00, N = 3SE +/- 0.33, N = 3SE +/- 0.00, N = 3129131121

PyPerformance

Benchmark: 2to3

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: 2to3manticmantic-no-omit-framepointer50100150200250SE +/- 0.00, N = 3SE +/- 0.33, N = 3221224

PyPerformance

Benchmark: chaos

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: chaosmanticmantic-no-omit-framepointer1428425670SE +/- 0.03, N = 3SE +/- 0.20, N = 362.863.6

PyPerformance

Benchmark: float

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: floatmanticmantic-no-omit-framepointer1530456075SE +/- 0.03, N = 3SE +/- 0.10, N = 367.466.9

PyPerformance

Benchmark: nbody

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: nbodymanticmantic-no-omit-framepointer20406080100SE +/- 0.06, N = 3SE +/- 0.07, N = 376.277.1

PyPerformance

Benchmark: pathlib

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: pathlibmanticmantic-no-omit-framepointer510152025SE +/- 0.00, N = 3SE +/- 0.00, N = 319.720.2

PyPerformance

Benchmark: raytrace

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: raytracemanticmantic-no-omit-framepointer60120180240300SE +/- 0.33, N = 3SE +/- 0.33, N = 3262274

PyPerformance

Benchmark: json_loads

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: json_loadsmanticmantic-no-omit-framepointernoble510152025SE +/- 0.06, N = 3SE +/- 0.03, N = 3SE +/- 0.03, N = 319.520.822.8

PyPerformance

Benchmark: crypto_pyaes

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: crypto_pyaesmanticmantic-no-omit-framepointer1530456075SE +/- 0.06, N = 3SE +/- 0.00, N = 365.166.6

PyPerformance

Benchmark: regex_compile

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: regex_compilemanticmantic-no-omit-framepointer306090120150SE +/- 0.00, N = 3SE +/- 0.33, N = 3116120

PyPerformance

Benchmark: python_startup

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: python_startupmanticmantic-no-omit-framepointernoble246810SE +/- 0.01, N = 3SE +/- 0.01, N = 3SE +/- 0.01, N = 37.617.648.76

PyPerformance

Benchmark: django_template

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: django_templatemanticmantic-no-omit-framepointer714212835SE +/- 0.03, N = 3SE +/- 0.06, N = 328.529.5

PyPerformance

Benchmark: pickle_pure_python

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyPerformance 1.0.0Benchmark: pickle_pure_pythonmanticmantic-no-omit-framepointer60120180240300SE +/- 0.33, N = 3SE +/- 0.58, N = 3259263

PyBench

Total For Average Test Times

OpenBenchmarking.orgMilliseconds, Fewer Is BetterPyBench 2018-02-16Total For Average Test Timesmanticmantic-no-omit-framepointernoble2004006008001000SE +/- 1.00, N = 3SE +/- 1.20, N = 3SE +/- 8.70, N = 4774790839


Phoronix Test Suite v10.8.5