3200u april

AMD Ryzen 3 3200U testing with a MOTILE PF4PU1F (N.1.03 BIOS) and AMD Radeon Vega 3 512MB on Ubuntu 20.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2204013-NE-3200UAPRI16
Jump To Table - Results

View

Do Not Show Noisy Results
Do Not Show Results With Incomplete Data
Do Not Show Results With Little Change/Spread
List Notable Results
Show Result Confidence Charts

Statistics

Show Overall Harmonic Mean(s)
Show Overall Geometric Mean
Show Wins / Losses Counts (Pie Chart)
Normalize Results
Remove Outliers Before Calculating Averages

Graph Settings

Force Line Graphs Where Applicable
Convert To Scalar Where Applicable
Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Highlight
Result
Hide
Result
Result
Identifier
View Logs
Performance Per
Dollar
Date
Run
  Test
  Duration
A
April 01 2022
  40 Minutes
B
April 01 2022
  2 Hours, 38 Minutes
C
April 01 2022
  2 Hours, 48 Minutes
Invert Hiding All Results Option
  2 Hours, 2 Minutes

Only show results where is faster than
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


3200u aprilOpenBenchmarking.orgPhoronix Test SuiteAMD Ryzen 3 3200U @ 2.60GHz (2 Cores / 4 Threads)MOTILE PF4PU1F (N.1.03 BIOS)AMD Raven/Raven23584MB128GB BIWIN SSDAMD Radeon Vega 3 512MB (1200/1200MHz)AMD Raven/Raven2/FenghuangRealtek RTL8111/8168/8411 + Intel Dual Band-AC 3168NGWUbuntu 20.045.15.0-051500-generic (x86_64)GNOME Shell 3.36.9X Server 1.20.134.6 Mesa 22.0.0-devel (git-9cb9101 2022-01-08 focal-oibaf-ppa) (LLVM 13.0.0 DRM 3.42)GCC 9.4.0ext41920x1080ProcessorMotherboardChipsetMemoryDiskGraphicsAudioNetworkOSKernelDesktopDisplay ServerOpenGLCompilerFile-SystemScreen Resolution3200u April BenchmarksSystem Logs- Transparent Huge Pages: madvise- --build=x86_64-linux-gnu --disable-vtable-verify --disable-werror --enable-checking=release --enable-clocale=gnu --enable-default-pie --enable-gnu-unique-object --enable-languages=c,ada,c++,go,brig,d,fortran,objc,obj-c++,gm2 --enable-libstdcxx-debug --enable-libstdcxx-time=yes --enable-multiarch --enable-multilib --enable-nls --enable-objc-gc=auto --enable-offload-targets=nvptx-none=/build/gcc-9-yTrUTS/gcc-9-9.4.0/debian/tmp-nvptx/usr,hsa --enable-plugin --enable-shared --enable-threads=posix --host=x86_64-linux-gnu --program-prefix=x86_64-linux-gnu- --target=x86_64-linux-gnu --with-abi=m64 --with-arch-32=i686 --with-default-libstdcxx-abi=new --with-gcc-major-version-only --with-multilib-list=m32,m64,mx32 --with-target-system-zlib=auto --with-tune=generic --without-cuda-driver -v - Scaling Governor: acpi-cpufreq ondemand (Boost: Enabled) - CPU Microcode: 0x8108102- OpenJDK Runtime Environment (build 11.0.14+9-Ubuntu-0ubuntu2.20.04)- Python 3.8.10- itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl and seccomp + spectre_v1: Mitigation of usercopy/swapgs barriers and __user pointer sanitization + spectre_v2: Mitigation of Full AMD retpoline IBPB: conditional STIBP: disabled RSB filling + srbds: Not affected + tsx_async_abort: Not affected

ABCResult OverviewPhoronix Test Suite100%110%121%131%fast-clispeedtest-cliJava JMHoneDNNperf-bench

3200u aprilonednn: Recurrent Neural Network Training - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Training - u8s8f32 - CPUonednn: Recurrent Neural Network Training - f32 - CPUjava-jmh: Throughputfast-cli: Internet Loaded Latency (Bufferbloat)fast-cli: Internet Latencyfast-cli: Internet Upload Speedfast-cli: Internet Download Speedonednn: Recurrent Neural Network Inference - bf16bf16bf16 - CPUonednn: Recurrent Neural Network Inference - u8s8f32 - CPUonednn: Recurrent Neural Network Inference - f32 - CPUspeedtest-cli: Internet Latencyspeedtest-cli: Internet Upload Speedspeedtest-cli: Internet Download Speedperf-bench: Memset 1MBperf-bench: Epoll Waitperf-bench: Futex Lock-Piperf-bench: Futex Hashperf-bench: Sched Pipeonednn: Deconvolution Batch shapes_1d - f32 - CPUonednn: Deconvolution Batch shapes_1d - u8s8f32 - CPUperf-bench: Memcpy 1MBonednn: IP Shapes 1D - f32 - CPUonednn: IP Shapes 1D - u8s8f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - f32 - CPUonednn: Matrix Multiply Batch Shapes Transformer - u8s8f32 - CPUonednn: IP Shapes 3D - f32 - CPUonednn: IP Shapes 3D - u8s8f32 - CPUperf-bench: Syscall Basiconednn: Convolution Batch Shapes Auto - u8s8f32 - CPUonednn: Convolution Batch Shapes Auto - f32 - CPUonednn: Deconvolution Batch shapes_3d - f32 - CPUonednn: Deconvolution Batch shapes_3d - u8s8f32 - CPUonednn: IP Shapes 1D - bf16bf16bf16 - CPUABC38516.638532.438761.82843596070.7889254.98020185.12022520075.542.5415.8970.1545.8408174952893725447994821898755.87337.059514.93741336.865728.545916.448713.343216.84516.689781245725968.548848.299164.396549.060239570.939394.239175.72953181899.576188254.87020641.720561.220542.529.9325.9271.7244.3749704939713747437385321396456.817039.054214.63208438.173228.774516.647813.332116.83436.751471317106669.817349.076565.647449.557740176.839940.839745.02961014078.74234254.85320988.020752.720661.944.0905.6849.9644.0991814937223755440581321136656.885738.742714.81864238.386528.629516.683113.376816.94446.739201325629269.755749.497465.957650.0337OpenBenchmarking.org

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: bf16bf16bf16 - Engine: CPUCBA9K18K27K36K45KSE +/- 53.42, N = 3SE +/- 251.58, N = 340176.839570.938516.6MIN: 39914.6MIN: 38855.2MIN: 38386.31. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: u8s8f32 - Engine: CPUCBA9K18K27K36K45KSE +/- 265.92, N = 3SE +/- 199.84, N = 339940.839394.238532.4MIN: 39404.3MIN: 39008.4MIN: 383871. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Training - Data Type: f32 - Engine: CPUCBA9K18K27K36K45KSE +/- 132.17, N = 3SE +/- 210.62, N = 339745.039175.738761.8MIN: 39411.1MIN: 38587.5MIN: 38608.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

Java JMH

This very basic test profile runs the stock benchmark of the Java JMH benchmark via Maven. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgOps/s, More Is BetterJava JMHThroughputABC600M1200M1800M2400M3000M2843596070.782953181899.582961014078.74

fast-cli

This test profile uses the open-source fast-cli client to benchmark your Internet connection's upload/download performance and latency against Netflix's fast.com service. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is Betterfast-cliInternet Loaded Latency (Bufferbloat)CBA50100150200250SE +/- 20.74, N = 15SE +/- 5.46, N = 1523418889

OpenBenchmarking.orgms, Fewer Is Betterfast-cliInternet LatencyCBA612182430SE +/- 0.63, N = 15SE +/- 0.54, N = 15252525

OpenBenchmarking.orgMbit/s, More Is Betterfast-cliInternet Upload SpeedBCA1.10252.2053.30754.415.5125SE +/- 0.13, N = 15SE +/- 0.14, N = 154.84.84.9

OpenBenchmarking.orgMbit/s, More Is Betterfast-cliInternet Download SpeedCBA20406080100SE +/- 1.87, N = 15SE +/- 1.56, N = 15537080

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: bf16bf16bf16 - Engine: CPUCBA4K8K12K16K20KSE +/- 25.94, N = 3SE +/- 45.51, N = 320988.020641.720185.1MIN: 20793.7MIN: 20454.6MIN: 20078.81. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: u8s8f32 - Engine: CPUCBA4K8K12K16K20KSE +/- 76.03, N = 3SE +/- 77.64, N = 320752.720561.220225.0MIN: 20485MIN: 20284.8MIN: 20127.61. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Recurrent Neural Network Inference - Data Type: f32 - Engine: CPUCBA4K8K12K16K20KSE +/- 32.94, N = 3SE +/- 9.59, N = 320661.920542.520075.5MIN: 20509.8MIN: 20405.1MIN: 19976.51. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

speedtest-cli

This test profile uses the open-source speedtest-cli client to benchmark your Internet connection's upload/download performance and latency against the Speedtest.net servers. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is Betterspeedtest-cli 2.1.3Internet LatencyCAB1020304050SE +/- 2.30, N = 14SE +/- 0.79, N = 1544.0942.5429.93

OpenBenchmarking.orgMbit/s, More Is Betterspeedtest-cli 2.1.3Internet Upload SpeedCAB1.3322.6643.9965.3286.66SE +/- 0.09, N = 14SE +/- 0.08, N = 155.685.895.92

OpenBenchmarking.orgMbit/s, More Is Betterspeedtest-cli 2.1.3Internet Download SpeedCAB1632486480SE +/- 1.56, N = 14SE +/- 1.89, N = 1549.9670.1571.72

perf-bench

This test profile is used for running Linux perf-bench, the benchmark support within the Linux kernel's perf tool. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/sec, More Is Betterperf-benchBenchmark: Memset 1MBCBA1020304050SE +/- 0.38, N = 12SE +/- 0.41, N = 1044.1044.3745.841. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lslang -lz -lnuma

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Epoll WaitCBA110K220K330K440K550KSE +/- 1579.52, N = 3SE +/- 1986.39, N = 34937224939714952891. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lslang -lz -lnuma

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Futex Lock-PiABC8001600240032004000SE +/- 11.02, N = 3SE +/- 4.04, N = 33725374737551. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lslang -lz -lnuma

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Futex HashBCA1000K2000K3000K4000K5000KSE +/- 18272.82, N = 3SE +/- 25564.63, N = 34373853440581344799481. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lslang -lz -lnuma

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Sched PipeCBA50K100K150K200K250KSE +/- 3146.65, N = 4SE +/- 3038.65, N = 32113662139642189871. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lslang -lz -lnuma

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: f32 - Engine: CPUCBA1326395265SE +/- 0.36, N = 3SE +/- 0.20, N = 356.8956.8255.87MIN: 52.24MIN: 51.77MIN: 51.911. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_1d - Data Type: u8s8f32 - Engine: CPUBCA918273645SE +/- 0.39, N = 3SE +/- 0.47, N = 339.0538.7437.06MIN: 32.94MIN: 32.14MIN: 32.441. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

perf-bench

This test profile is used for running Linux perf-bench, the benchmark support within the Linux kernel's perf tool. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgGB/sec, More Is Betterperf-benchBenchmark: Memcpy 1MBBCA48121620SE +/- 0.12, N = 3SE +/- 0.05, N = 314.6314.8214.941. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lslang -lz -lnuma

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: f32 - Engine: CPUCBA918273645SE +/- 0.58, N = 3SE +/- 0.61, N = 338.3938.1736.87MIN: 34.09MIN: 33.92MIN: 34.131. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 1D - Data Type: u8s8f32 - Engine: CPUBCA714212835SE +/- 0.06, N = 3SE +/- 0.08, N = 328.7728.6328.55MIN: 26.44MIN: 26.65MIN: 26.621. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: f32 - Engine: CPUCBA48121620SE +/- 0.01, N = 3SE +/- 0.02, N = 316.6816.6516.45MIN: 15.42MIN: 15.56MIN: 15.491. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Matrix Multiply Batch Shapes Transformer - Data Type: u8s8f32 - Engine: CPUCAB3691215SE +/- 0.01, N = 3SE +/- 0.01, N = 313.3813.3413.33MIN: 12.06MIN: 11.94MIN: 11.951. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: f32 - Engine: CPUCAB48121620SE +/- 0.05, N = 3SE +/- 0.03, N = 316.9416.8516.83MIN: 16.69MIN: 16.75MIN: 16.671. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: IP Shapes 3D - Data Type: u8s8f32 - Engine: CPUBCA246810SE +/- 0.02701, N = 3SE +/- 0.03112, N = 36.751476.739206.68978MIN: 6.08MIN: 6.12MIN: 6.211. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

perf-bench

This test profile is used for running Linux perf-bench, the benchmark support within the Linux kernel's perf tool. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgops/sec, More Is Betterperf-benchBenchmark: Syscall BasicABC3M6M9M12M15MSE +/- 84928.86, N = 3SE +/- 95015.37, N = 31245725913171066132562921. (CC) gcc options: -O6 -ggdb3 -funwind-tables -std=gnu99 -lunwind-x86_64 -lunwind -llzma -Xlinker -lpthread -lrt -lm -ldl -lelf -lcrypto -lslang -lz -lnuma

oneDNN

This is a test of the Intel oneDNN as an Intel-optimized library for Deep Neural Networks and making use of its built-in benchdnn functionality. The result is the total perf time reported. Intel oneDNN was formerly known as DNNL (Deep Neural Network Library) and MKL-DNN before being rebranded as part of Intel oneAPI. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: u8s8f32 - Engine: CPUBCA1632486480SE +/- 0.27, N = 3SE +/- 0.32, N = 369.8269.7668.55MIN: 68.71MIN: 68.52MIN: 68.191. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Convolution Batch Shapes Auto - Data Type: f32 - Engine: CPUCBA1122334455SE +/- 0.30, N = 3SE +/- 0.27, N = 349.5049.0848.30MIN: 47.05MIN: 47.39MIN: 46.761. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: f32 - Engine: CPUCBA1530456075SE +/- 0.07, N = 3SE +/- 0.05, N = 365.9665.6564.40MIN: 63.15MIN: 63.07MIN: 61.731. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

OpenBenchmarking.orgms, Fewer Is BetteroneDNN 2.6Harness: Deconvolution Batch shapes_3d - Data Type: u8s8f32 - Engine: CPUCBA1122334455SE +/- 0.25, N = 3SE +/- 0.05, N = 350.0349.5649.06MIN: 45.6MIN: 46.99MIN: 47.611. (CXX) g++ options: -O3 -march=native -fopenmp -msse4.1 -fPIC -std=c++11 -pie -lpthread -ldl

Harness: Convolution Batch Shapes Auto - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: Matrix Multiply Batch Shapes Transformer - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: Deconvolution Batch shapes_3d - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: Deconvolution Batch shapes_1d - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: IP Shapes 3D - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

Harness: IP Shapes 1D - Data Type: bf16bf16bf16 - Engine: CPU

A: The test run did not produce a result.

B: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.

C: The test run did not produce a result. The test run did not produce a result. The test run did not produce a result.