Jetson Benchmarks

Jetson is used to deploy a wide range of popular DNN models and ML frameworks to the edge with high performance inferencing, for tasks like real-time classification and object detection, pose estimation, semantic segmentation, and natural language processing (NLP).

MLPerf Inference Benchmarks

The tables below show inferencing benchmarks from the NVIDIA Jetson submissions to the MLPerf Inference Edge category.

Jetson AGX Orin Jetson Orin NX MLPerf v3.0 Results

Model NVIDIA Jetson AGX Orin (TensorRT) NVIDIA Orin MaxQ (TensorRT) NVIDIA Jetson Orin NX
Single Stream (Samples/s)Offline (Samples/s)Multi Stream (Samples/s)Offline (Samples/s)System Power(W)Offline (Samples/s)
Image Classification
ResNet-50
15386438.1036863525.9123.062517.99
Object Detection
Retinanet
51.5792.4060.0034.622.436.14
Medical Imaging
3D-Unet
.26.51N/A3.2828.64.19
Speech-to-text
RNN-T
9.8221170.23N/A1447225.64405.27
Natural Language Processing
BERT
144.36544.24N/A3685.3625.91163.57
  • Steps to reproduce these results can be found at v3.0 Results | MLCommons
  • These results were achieved with the NVIDIA Jetson AGX Orin Developer Kit running a preview of TensorRT 8.5.0, and CUDA 11.4
  • Note different configurations were used for single stream, offline and multistream. Reference the MLCommons page for more details


Jetson AGX Xavier and Jetson Xavier NX MLPerf v1.1 Results

Model Jetson Xavier NX (TensorRT) Jetson AGX Xavier 32GB (TensorRT)
Image Classification
ResNet-50
1245.102039.11
Object Detection
SSD-small
1786.912833.59
Object Detection
SSD-Large
36.9755.16
Speech to Text
RNN-T
259.67416.13
Natural Language
Processing
BERT-Large
61.3496.73


NVIDIA Pretrained Model Benchmarks

NVIDIA pretrained models from NGC start you off with highly accurate and optimized models and model architectures for various use cases. Pretrained models are production-ready. You can further customize these models by training with your own real or synthetic data, using the NVIDIA TAO (Train-Adapt-Optimize) workflow to quickly build an accurate and ready to deploy model.The table below shows inferencing benchmarks for some of our pretrained models running on Jetson modules.

Jetson Pretrained Model Benchmarks


Jetson Orin Results

ModelJetson Orin Nano 4GBJetson Orin Nano 8GBJetson Orin NX 8GBJetson Orin NX 16GBJetson AGX Orin 32GBJetson AGX Orin 64GB
PeopleNet (v2.5 unpruned)57117192240409685
Action Recognition 2D22037244048311581517
Action Recognition 3D1326323971108
LPR Net5529741314142728004213
Dashcam Net20040068987714822139
Bodypose Net69137169203360563
ModelJetson NanoJetson TX2 NXJetson Xavier NXJetson AGX Xavier
PeopleNet (v2.5 unpruned)25120195
Action Recognition 2D3288245472
Action Recognition 3D132132
LPR Net47867141236
Dashcam Net1126424667
Bodypose Net37104172

  • Jetson Orin & Jetson Xavier Benchmarks were run using Jetpack 5.1.1
  • Each Jetson module was run with maximum performance (MAXN for JAO64, JAO32, ONX16, ONX8; and 15W mode for JON8, and 10W mode for JON4)
  • For Jetson Nano and Jetson TX2 NX, these benchmarks were run using Jetpack 4.6.1
  • Each Jetson module was run with maximum performance (MAXN)
  • Reproduce these results by downloading these models from our NGC catalog

Jetson Family Benchmarks

ModelJetson Orin Nano 4GBJetson Orin Nano 8GBJetson Orin NX 8GBJetson Orin NX 16GBJetson AGX Orin 32GBJetson AGX Orin 64GB
Inveption_V41823615937691337.81702.6
VGG191743614425329371471
Super_resolution102203280386610882
UNET-sgmentation76148183217387584
Pose Estimation28054666580014242048
Yolov3-tiny3717311156144026113179
Resnet5062111581725218337174834
SSD-Mobilnet109421562893345764157671
SSD_Resnet34_1200x120018345272120163
Yolov5m69131162193342519
Yolov5s1583013794497851135
  • These Benchmarks were run using Jetpack 5.1.1
  • Each Jetson module was run with maximum performance (Max Frequencies in MAXN for JAO64, JAO32, ONX16, ONX8; and 15W mode for JON8, and 10W mode for JON4)
  • Steps to reproduce these results can be found here