tflite

ARMv8 Cortex-A78E testing with a NVIDIA Jetson AGX Orin Developer Kit (36.3.0-gcid-36191598 BIOS) and Orin on Ubuntu 22.04 via the Phoronix Test Suite.

Compare your own system(s) to this result file with the Phoronix Test Suite by running the command: phoronix-test-suite benchmark 2408262-NE-TFLITE58672
Jump To Table - Results

Statistics

Remove Outliers Before Calculating Averages

Graph Settings

Prefer Vertical Bar Graphs

Multi-Way Comparison

Condense Multi-Option Tests Into Single Result Graphs

Table

Show Detailed System Result Table

Run Management

Result
Identifier
Performance Per
Dollar
Date
Run
  Test
  Duration
all of them
August 26
  19 Minutes
Only show results matching title/arguments (delimit multiple options with a comma):
Do not show results matching title/arguments (delimit multiple options with a comma):


tfliteOpenBenchmarking.orgPhoronix Test SuiteARMv8 Cortex-A78E @ 2.20GHz (12 Cores)NVIDIA Jetson AGX Orin Developer Kit (36.3.0-gcid-36191598 BIOS)30GB1000GB Samsung SSD 960 EVO 1TB + 64GB G1M15MOrinRealtek RTL8822CE 802.11ac PCIeUbuntu 22.045.15.136-tegra (aarch64)GNOME Shell 42.9X Server 1.21.1.4NVIDIA1.3.251GCC 11.4.0 + CUDA 12.2ext46582x1234ProcessorMotherboardMemoryDiskGraphicsNetworkOSKernelDesktopDisplay ServerDisplay DriverVulkanCompilerFile-SystemScreen ResolutionTflite BenchmarksSystem Logs- Transparent Huge Pages: always- Scaling Governor: tegra194 performance- gather_data_sampling: Not affected + itlb_multihit: Not affected + l1tf: Not affected + mds: Not affected + meltdown: Not affected + mmio_stale_data: Not affected + retbleed: Not affected + spec_rstack_overflow: Not affected + spec_store_bypass: Mitigation of SSB disabled via prctl + spectre_v1: Mitigation of __user pointer sanitization + spectre_v2: Mitigation of CSV2 but not BHB + srbds: Not affected + tsx_async_abort: Not affected

tflitetensorflow-lite: Inception ResNet V2tensorflow-lite: Mobilenet Quanttensorflow-lite: Mobilenet Floattensorflow-lite: NASNet Mobiletensorflow-lite: Inception V4tensorflow-lite: SqueezeNetall of them78285.92522.194698.1821324.484257.56212.21OpenBenchmarking.org

TensorFlow Lite

This is a benchmark of the TensorFlow Lite implementation focused on TensorFlow machine learning for mobile, IoT, edge, and other cases. The current Linux support is limited to running on CPUs. This test profile is measuring the average inference time. Learn more via the OpenBenchmarking.org test page.

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Inception ResNet V2all of them20K40K60K80K100KSE +/- 274.21, N = 378285.9

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Mobilenet Quantall of them5001000150020002500SE +/- 4.68, N = 32522.19

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Mobilenet Floatall of them10002000300040005000SE +/- 16.55, N = 34698.18

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: NASNet Mobileall of them5K10K15K20K25KSE +/- 292.35, N = 321324.4

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: Inception V4all of them20K40K60K80K100KSE +/- 131.77, N = 384257.5

OpenBenchmarking.orgMicroseconds, Fewer Is BetterTensorFlow Lite 2022-05-18Model: SqueezeNetall of them13002600390052006500SE +/- 6.38, N = 36212.21