Platform |
Submitter |
System |
Device |
Devices per node |
Framework | Performance |
Platform |
Submitter |
System |
Device |
Devices per node |
Framework | Performance |
rigel-nvidia_original-gpu-tensorrt-vdefault-default_config | CTuning | One Stop Systems Rigel Edge Supercomputer | NVIDIA HGX A100 SXM | 4 | Nvidia inference implementation with CM API, TensorRT v8.6.1.6 | 173501 |
R750xa_A100_PCIE_80GBx4_TRT | Dell | Dell PowerEdge R750xa (4x A100-PCIe-80GB, TensorRT) | NVIDIA A100-PCIe-80GB | 4 | TensorRT 9.0.0, CUDA 12.2 | 158285 |
r282_z93_q5 | Qualcomm | GIGABYTE R282-Z93 (5x QAIC100 Pro) | QUALCOMM Cloud AI 100 PCIe/HHHL Pro | 5 | QUALCOMM Cloud AI SDK v1.9.1 | 125950 |
R760xa_L40x4_edge_TRT | Dell | Dell PowerEdge R760xa (4x L40, TensorRT) | NVIDIA L40 | 4 | TensorRT 9.0.0, CUDA 12.2 | 115291 |
r282_z93_q5e | Qualcomm | GIGABYTE R282-Z93 (5x QAIC100 Pro, EE) | QUALCOMM Cloud AI 100 PCIe/HHHL Pro | 5 | QUALCOMM Cloud AI SDK v1.9.1 | 106349 |
se450_q4_std | Lenovo | Lenovo ThinkEdge SE450 Edge Server (4x QAIC100 Standard) | QUALCOMM Cloud AI 100 PCIe/HHHL Standard | 4 | QUALCOMM Cloud AI SDK v1.9.1 | 80092.7 |
e920d_q4_std | HPE | HPE ProLiant e920d (4x QAIC100 Standard) | QUALCOMM Cloud AI 100 PCIe/HHHL Standard | 4 | QUALCOMM Cloud AI SDK v1.9.1 | 79948.9 |
NX_3155G_G8_A100_PCIe_80GBx2_TRT | Nutanix | NX_3155G_G8_A100_PCIe_80GBx2 | NVIDIA A100-PCIe-80GB | 2 | TensorRT 8.6.0, CUDA 12.0 | 74652.6 |
Lenovo_L40_PCIe_48GBx2_TRT | Lenovo | Lenovo ThinkEdge SE450 Edge Server (2x NVIDIA L40 PCIe 48GB) | NVIDIA L40-PCIE-48GB | 2 | TensorRT 9.0.0, CUDA 12.2 | 52623.4 |
amd_zen4_workstation-nvidia_original-gpu-tensorrt-vdefault-default_config | CTuning | PCSPECIALIST AMD AM5 PC with Nvidia RTX 4090 | NVIDIA GeForce RTX 4090 (Ada Lovelace) | 1 | Nvidia inference implementation with CM API, TensorRT v8.6.1.6 | 45772.1 |
xr4520c_q2_std | Dell | Dell PowerEdge XR4520c (2x QAIC100 Standard) | QUALCOMM Cloud AI 100 PCIe/HHHL Standard | 2 | QUALCOMM Cloud AI SDK v1.9.1 | 40362.5 |
amd_ryzen_workstation_triton-nvidia_original-gpu-tensorrt-vdefault-using_triton | CTuning | PCSPECIALIST AMD AM5 PC (Nvidia Triton server) | NVIDIA GeForce RTX 4090 (Ada Lovelace) | 1 | Nvidia inference implementation with CM API, TensorRT v8.6.1.6 | 37842.2 |
R5300G6_L40x1_TRT | H3C | H3C UniServer R5300 G6 (1x L40, TensorRT) | NVIDIA L40 | 1 | TensorRT 9.0.0, CUDA 12.2 | 27970.4 |
NF5468M6_A40x1_TRT | IEI | NF5468M6 (1x A40, TensorRT) | NVIDIA A40 | 1 | TensorRT 9.0.0, CUDA 12.2 | 19391.2 |
L4x1_TRT | NVIDIA | ASROCKRACK 1U1G-MILAN (1x L4, TensorRT) | NVIDIA L4 | 1 | TensorRT 9.0.0, CUDA 12.2 | 12881.7 |
XR7620_L4x1_TRT | Dell | Dell PowerEdge XR7620 (1x L4, TensorRT) | NVIDIA L4 | 1 | TensorRT 9.0.0, CUDA 12.2 | 12839.4 |
XR4520c_L4x1_TRT | Dell | Dell PowerEdge XR4520c (1x L4, TensorRT) | NVIDIA L4 | 1 | TensorRT 9.0.0, CUDA 12.2 | 12836.1 |
gcp_g2.standard.4-nvidia_original-gpu-tensorrt-vdefault-default_config | CTuning | Google Cloud Platform (g2.standard.4) | NVIDIA L4 | 1 | Nvidia inference implementation with CM API, TensorRT v8.6.1.6 | 12557.8 |
XR5610_L4x1_TRT_MaxQ | Dell | Dell PowerEdge XR5610 (1x L4, MaxQ, TensorRT) | NVIDIA L4 | 1 | TensorRT 9.0.0, CUDA 12.2, | 12432.5 |
KR580S1-nvidia_original-gpu-tensorrt-vdefault-default_config | TTA | KR580S1 | NVIDIA T4 | 2 | tensorrt 8.6.1, | 10639.8 |
gloria_highend | Qualcomm | Foxconn Gloria (Highend) | QUALCOMM Cloud AI 100 DM.2 | 1 | QUALCOMM Cloud AI SDK v1.9.1 | 9808.25 |
Orin_TRT | NVIDIA | NVIDIA Jetson AGX Orin Developer Kit 64G (TensorRT) | NVIDIA Jetson AGX Orin 64G | 1 | Jetpack 5.1.1, TensorRT 8.5.2, CUDA 11.4 | 6423.63 |
nvidia_orin_64k_default_jetpack-nvidia_original-gpu-tensorrt-vdefault-default_config | CTuning | Siliconhighway NVIDIA Jetson AGX Orin (64K page size) | NVIDIA Jetson AGX Orin 32G | 1 | Jetpack 5.1.1 with TensorRT 8.5.2, cuDNN 8.6.0, CUDA 11.4.19 | 6341.65 |
nvidia_orin_maxn_pagesize.4k-nvidia_original-gpu-tensorrt-vdefault-default_config | CTuning | Siliconhighway NVIDIA Jetson AGX Orin (4K page size) | NVIDIA Jetson AGX Orin 32G | 1 | Jetpack 5.1.1 with TensorRT 8.5.2, cuDNN 8.6.0, CUDA 11.4.19 | 6341.65 |
aws_g4dn.xlarge-nvidia_original-gpu-tensorrt-vdefault-default_config | CTuning | AWS cloud instance g4dn.xlarge | NVIDIA T4 | 1 | Nvidia inference implementation with CM API, TensorRT v8.6.1.6 | 6049.73 |
aws_t4-nvidia_original-gpu-tensorrt-vdefault-default_config | CTuning | AWS cloud instance g4dn.xlarge | NVIDIA T4 | 1 | Nvidia inference implementation with CM API, TensorRT v8.6.1.6 | 4601.81 |
gcp_t4-nvidia_original-gpu-tensorrt-vdefault-default_config | CTuning | Google Cloud Platform (n1-standard-4) | Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz | 2 | Nvidia inference implementation with CM API, TensorRT v8.6.1.6 | 4575.76 |
amd_ryzen_workstation-cpp-gpu-onnxruntime-vdefault-default_config | CTuning | PCSPECIALIST AMD AM5 PC | NVIDIA GeForce RTX 4090 (Ada Lovelace) | 1 | cTuning/MLCommons C++ Modular Inference Library (MILT), Onnxruntime v1.15.1 | 4007.25 |
amd_ryzen_workstation-reference-gpu-onnxruntime-v1.15.1-default_config | CTuning | PCSPECIALIST AMD AM5 PC with Nvidia RTX 4090 | NVIDIA GeForce RTX 4090 (Ada Lovelace) | 1 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 3857.76 |
Orin_TRT_MaxQ | NVIDIA | NVIDIA Jetson AGX Orin Developer Kit 64G (MaxQ, TensorRT) | NVIDIA Jetson AGX Orin 64G | 1 | Jetpack 5.1.1, TensorRT 8.5.2, CUDA 11.4 | 3526.29 |
orin-nvidia_original-gpu-tensorrt-vdefault-default_config | CTuning | Siliconhighway NVIDIA Jetson AGX Orin (non MaxN image) | NVIDIA Jetson AGX Orin 32G | 1 | Nvidia inference implementation with CM API, Jetpack 5.1.1 with TensorRT 8.5.2, cuDNN 8.6.0, CUDA 11.4.19 | 3129.69 |
davinci_hhhl | SiMa | MLSoC HHHL | DaVinci | 1 | SiMa SDK Version 0.7 | 2952.58 |
Orin_NX_TRT | NVIDIA | NVIDIA Orin NX 16G (TensorRT) | NVIDIA Orin NX 16G | 1 | Jetpack 5.1.1, TensorRT 8.5.2, CUDA 11.4 | 2640.51 |
Orin_NX_TRT | ConnectTechInc | NVIDIA Orin NX 16G (TensorRT) + CTI Hadron Carrier (NGX012) | NVIDIA Orin NX 16G | 1 | Jetpack 5.1.1, TensorRT 8.5.2, CUDA 11.4 | 2565.64 |
amd_ryzen_workstation-reference-gpu-tf-v2.13.0-default_config | CTuning | PCSPECIALIST AMD AM5 PC with Nvidia RTX 4090 | NVIDIA GeForce RTX 4090 (Ada Lovelace) | 1 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 | 2359.98 |
amd_ryzen_workstation-reference-gpu-tf-v2.12.0-default_config | CTuning | PCSPECIALIST AMD AM5 PC with Nvidia RTX 4090 | NVIDIA GeForce RTX 4090 (Ada Lovelace) | 1 | MLCommons reference implementation with CM API, Tensorflow v2.12.0 | 2131.38 |
Orin_NX_TRT_MaxQ | NVIDIA | NVIDIA Orin NX 16G (MaxQ, TensorRT) | NVIDIA Orin NX 16G | 1 | Jetpack 5.1.1, TensorRT 8.5.2, CUDA 11.4 | 1681.87 |
aws_g4dn.xlarge-reference-gpu-onnxruntime-v1.15.1-default_config | CTuning | AWS cloud instance g4dn.xlarge | NVIDIA T4 | 1 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 (with gpu support) | 528.194 |
aws_g4dn.xlarge-reference-gpu-tf-v2.13.0-default_config | CTuning | AWS cloud instance g4dn.xlarge | NVIDIA T4 | 1 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 | 342.669 |
macbook_pro_m1_2-reference-cpu-tf-v2.13.0-default_config | CTuning | Apple MacBook Pro M1 | undefined | 10 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 with tensorflow-metal | 307.693 |
51535c5ae107-reference-gpu-onnxruntime-v1.15.1-default_config | CTuning | Collabora AMD ThreadRipper | AMD Ryzen Threadripper 1920X 12-Core Processor | 12 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 305.092 |
7920t-kilt-onnxruntime_gpu | Krai | Dell Precision 7920 Tower (1x NVIDIA RTX A5000 GPU) | NVIDIA RTX A5000 GPU | 1 | KRAI Inference Library Technology (KILT) with ONNX Runtime GPU support | 272.245 |
amd_ryzen_workstation-reference-gpu-onnxruntime-v1.14.0-default_config | CTuning | PCSPECIALIST AMD AM5 PC with Nvidia RTX 4090 | NVIDIA GeForce RTX 4090 (Ada Lovelace) | 1 | MLCommons reference implementation with CM API, Onnxruntime v1.14.0 (with GPU support) | 255.537 |
amd_ryzen_workstation-reference-cpu-onnxruntime-v1.15.1-default_config | CTuning | PCSPECIALIST AMD AM5 PC | AMD Ryzen 9 7950X 16-Core Processor | 16 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 253.917 |
amd_zen4_workstation-reference-cpu-onnxruntime-v1.14.0-default_config | CTuning | PCSPECIALIST AMD AM5 PC | AMD Ryzen 9 7950X 16-Core Processor | 16 | MLCommons reference implementation with CM API, Onnxruntime v1.14.0 | 252.609 |
amd_zen4_workstation-reference-cpu-tf-v2.12.0-default_config | CTuning | PCSPECIALIST AMD AM5 PC | AMD Ryzen 9 7950X 16-Core Processor | 16 | MLCommons reference implementation with CM API, Tensorflow v2.12.0 | 252.406 |
amd_ryzen_workstation-reference-cpu-tf-v2.13.0-default_config | CTuning | PCSPECIALIST AMD AM5 PC | AMD Ryzen 9 7950X 16-Core Processor | 16 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 | 252.083 |
amd_ryzen_workstation-reference-cpu-tf-v2.12.0-default_config | CTuning | PCSPECIALIST AMD AM5 PC | AMD Ryzen 9 7950X 16-Core Processor | 16 | MLCommons reference implementation with CM API, Tensorflow v2.12.0 | 248.944 |
amd_ryzen_workstation-cpp-cpu-onnxruntime-vdefault-default_config | CTuning | PCSPECIALIST AMD AM5 PC | AMD Ryzen 9 7950X 16-Core Processor | 16 | cTuning/MLCommons C++ Modular Inference Library (MILT), Onnxruntime v1.15.1 | 234.449 |
7920t-kilt-onnxruntime_cpu | Krai | Dell Precision 7920 Tower | Intel(R) Xeon(R) Gold 6240R CPU @ 2.40GHz | 24 | KRAI Inference Library Technology (KILT) with ONNX Runtime CPU support | 134.187 |
eb6-kilt-snpe_aip | Krai | Thundercomm TurboX EB6 Edge AI Box | QUALCOMM NPU 230 AIP | 1 | KRAI Inference Library Technology (KILT) with Snapdragon AIP support | 128.469 |
eb6-kilt-snpe_dsp | Krai | Thundercomm TurboX EB6 Edge AI Box | QUALCOMM Hexagon 698 DSP | 1 | KRAI Inference Library Technology (KILT) with Snapdragon DSP support | 105.985 |
amd_zen4_workstation-reference-cpu-ncnn-vdefault-default_config | CTuning | PCSPECIALIST AMD AM5 PC | AMD Ryzen 9 7950X 16-Core Processor | 16 | MLCommons reference implementation with CM API, NCNN vdefault | 92.4211 |
macbook_pro_m1_2-reference-cpu-onnxruntime-v1.15.1-default_config | CTuning | Apple MacBook Pro M1 | Apple M1 Pro | 10 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 69.2985 |
gcp.c3_standard_8-reference-cpu-tf-v2.13.0-default_config | CTuning | Google Cloud Platform (c3.standard.8) | Intel(R) Xeon(R) Platinum 8481C CPU @ 2.70GHz | 4 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 | 66.8676 |
gcp.c3_standard_8-cpp-cpu-onnxruntime-vdefault-default_config | CTuning | Google Cloud Platform (c3.standard.8) | Intel(R) Xeon(R) Platinum 8481C CPU @ 2.70GHz | 4 | cTuning/MLCommons C++ Modular Inference Library (MILT), Onnxruntime v1.15.1 | 65.7066 |
gcp.c3_standard_8-reference-cpu-onnxruntime-v1.15.1-default_config | CTuning | Google Cloud Platform (c3.standard.8) | Intel(R) Xeon(R) Platinum 8481C CPU @ 2.70GHz | 4 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 64.5665 |
nvidia_orin-reference-cpu-tf-v2.13.0-default_config | CTuning | Siliconhighway NVIDIA Jetson AGX Orin | ARMv8 Processor rev 1 (v8l) | 12 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 | 41.6048 |
nvidia_orin-reference-gpu-tf-v2.13.0-default_config | CTuning | Siliconhighway NVIDIA Jetson AGX Orin | ARMv8 Processor rev 1 (v8l) | 12 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 | 41.1616 |
nvidia_orin-reference-cpu-onnxruntime-v1.15.1-default_config | CTuning | Siliconhighway NVIDIA Jetson AGX Orin | ARMv8 Processor rev 1 (v8l) | 12 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 32.3828 |
aws_g4dn.xlarge-reference-cpu-tf-v2.13.0-default_config | CTuning | AWS cloud instance g4dn.xlarge | Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz | 2 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 | 31.8977 |
nvidia_orin_maxq-reference-cpu-onnxruntime-v1.15.1-default_config | CTuning | Siliconhighway NVIDIA Jetson AGX Orin (non MaxN image) | ARMv8 Processor rev 1 (v8l) | 12 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 28.9732 |
aws_g4dn.xlarge-reference-cpu-onnxruntime-v1.15.1-default_config | CTuning | AWS cloud instance g4dn.xlarge | Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz | 2 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 27.9115 |
aws_inf2.xlarge-reference-cpu-onnxruntime-v1.15.1-default_config | CTuning | AWS cloud instance inf2.xlarge | AMD EPYC 7R13 Processor | 2 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 24.3065 |
eb6-kilt-snpe_gpu | Krai | Thundercomm TurboX EB6 Edge AI Box | QUALCOMM Adreno 650 GPU | 1 | KRAI Inference Library Technology (KILT) with Snapdragon GPU support | 21.0084 |
amd_ryzen_workstation-tflite_cpp-cpu-tflite-vmaster-default_config | CTuning | PCSPECIALIST AMD AM5 PC | AMD Ryzen 9 7950X 16-Core Processor | 16 | cTuning/MLCommons TFLite inference implementation, tflite vmaster | 20.0564 |
amd_ryzen_workstation-tflite_cpp-cpu-tflite-vmaster-with_live_preprocessing | CTuning | PCSPECIALIST AMD AM5 PC | AMD Ryzen 9 7950X 16-Core Processor | 16 | cTuning/MLCommons TFLite inference implementation, tflite vmaster (with live preprocessing) | 19.4557 |
51535c5ae107-reference-cpu-ncnn-vdefault-default_config | CTuning | Collabora AMD ThreadRipper | AMD Ryzen Threadripper 1920X 12-Core Processor | 12 | MLCommons reference implementation with CM API, NCNN vdefault | 18.4072 |
aws_t4-reference-cpu-tf-v2.13.0-default_config | CTuning | AWS cloud instance g4dn.xlarge | AMD EPYC 7R32 | 2 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 | 17.4295 |
gcp.c3_standard_8-tflite_cpp-cpu-tflite-vmaster-default_config | CTuning | Google Cloud Platform (c3.standard.8) | Intel(R) Xeon(R) Platinum 8481C CPU @ 2.70GHz | 4 | cTuning/MLCommons TFLite inference implementation, tflite vmaster | 10.8284 |
gcp.c3_standard_8-tflite_cpp-cpu-tflite-vmaster-with_live_preprocessing | CTuning | Google Cloud Platform (c3.standard.8) | Intel(R) Xeon(R) Platinum 8481C CPU @ 2.70GHz | 4 | cTuning/MLCommons TFLite inference implementation, tflite vmaster | 10.7615 |
aws_g4dn.xlarge-tflite_cpp-cpu-tflite-vmaster-default_config | CTuning | AWS cloud instance g4dn.xlarge | Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz | 2 | cTuning/MLCommons TFLite inference implementation, tflite vmaster | 8.54644 |
gcp_t2a.standard.2-reference-cpu-tf-v2.13.0-default_config | CTuning | Google Cloud Platform (t2a-standard-2) | Neoverse-N1 | 2 | MLCommons reference implementation with CM API, Tensorflow v2.13.0 | 8.20758 |
gcp_t2a.standard.2-reference-cpu-onnxruntime-v1.15.1-default_config | CTuning | Google Cloud Platform (t2a-standard-2) | Neoverse-N1 | 2 | MLCommons reference implementation with CM API, Onnxruntime v1.15.1 | 6.85995 |
eb6-kilt-snpe_cpu | Krai | Thundercomm TurboX EB6 Edge AI Box | QUALCOMM Snapdragon 865 (QRB5165) | 8 | KRAI Inference Library Technology (KILT) with Snapdragon CPU support | 6.3871 |
nvidia_orin-tflite_cpp-cpu-tflite-vmaster-default_config | CTuning | Siliconhighway NVIDIA Jetson AGX Orin | ARMv8 Processor rev 1 (v8l) | 12 | cTuning/MLCommons TFLite inference implementation, tflite vmaster | 3.99455 |
firefly_rk3399-reference-cpu-ncnn-vdefault-default_config | CTuning | Collabora firefly_rk3399 | Cortex-A53 | 6 | MLCommons reference implementation with CM API, NCNN vdefault | 2.74346 |
firefly_rk3399-tflite_cpp-cpu-tflite-vmaster-default_config | CTuning | Collabora firefly_rk3399 | Cortex-A53 | 6 | cTuning/MLCommons TFLite inference implementation, tflite vmaster | 0.873957 |