| layout | post | ||||
|---|---|---|---|---|---|
| date | 2025-08-10 | ||||
| lastchange | v010 + GPUs 2024-12-21 :nvidia.md | ||||
| url | https://bomonike.github.io/nvidia | ||||
| file | nvidia | ||||
| title | NVIDIAJetson AI | ||||
| excerpt | How to get, install, and use NVIDIA's Jetson micro servers for AI at edge. | ||||
| tags |
|
||||
| comments | true | ||||
| created | 2024-12-21 |
{{ page.excerpt }}
{% include l18n.html %}
{% include _toc.html %}
From https://developer.nvidia.com/embedded/develop/software
https://github.com/NVIDIA/GenerativeAIExamples
NVIDIA has a program for training and certifying university educators and certifying Jetson AGX Orin developers.
-
WARNING: Certiport only works on Chrome and Microsoft Edge browsers (not on Safari & Firefox).
-
A "Safe Exam Browser" must be installed.
-
PROTIP: Restart your computer to enable pop-ups requesting configuration of microphone and video permissions.
-
Authorize screen recording by the app.
-
Launch the system compatibility check.
-
Reschedule 24 hours in advance for an examination date within two months of the date the examination fee was paid for that examination.
-
Results are pass/fail on Credly) to accept your digital badge.
Http failure response for https://market-production.azurewebsites.net/api/Organization/StoreExams/0894ab5e-cc68-4f2b-a8bc-d41a576d345d/slots?startDate=2/22/2025,%2012:00:00%20AM&endDate=2/22/2025,%2011:59:59%20PM&timeZone=America/Denver&slotDurationInMinutes=60: 401 OK
https://www.udemy.com/course/nvidia-certified-associate-generative-ai-llms-nca-genl/?couponCode=KEEPLEARNING $19.99 "NVIDIA-Certified Associate - Generative AI LLMs (NCA-GENL)" on Udemy
-
NVIDIA offers a 50 (40-60) question exam in one-hour with no breaks taken online, each at $135 for each retake. It's good for a 2-year validity period.
NVIDIA-Certified Associate: Generative AI and LLMs (NCA-GENL) validates skills in the use of generative AI and large language models:
30% Core Machine Learning and AI Knowledge
24% Software Development
22% Experimentation
14% Data Analysis and Visualization
10% Trustworthy AIStudy materials:
-
$FREE Generative AI Explained by Bryan Catanzaro, VP, Applied Deep Learning Research
-
$30 Introduction to Transformer-Based Natural Language Processing.
-
8-hour $500 Building Transformer-Based Natural Language Processing Applications.
-
1-hour $FREE Augment Your LLM Using Retrieval-Augmented Generation.
-
8-hour $90 Generative AI With Diffusion Models (to generate images from text)
-
8-hour $500 Efficient Large Language Model (LLM) Customization.
-
NVIDIA-Certified Associate: Generative AI Multimodal (NCA-GENM)
25% Experimentation
20% Core Machine Learning and AI Knowledge
15% Multimodel Data *
15% Software Development
10% Data Analysis and Visualization
10% Performance Optimization *
5% Trustworthy AINotice the two topics added (marked by *).
In addition to the resources for the GENL exam:
-
$500 Rapid Application Development with Large Lanaguage Models
-
High-Resolution Image Synthesis via Two-Stage Generative Models (on-demand video, 35 minutes)
-
Building Lifelike Digital Avatars With NVIDIA ACE Microservices (blog, 15 minutes)
-
The Future of Generative AI for Content Creation (on-demand video, 35 minutes)
-
NVIDIA-Certified Associate: AI Infrastructure and Operations (NCA-AIIO) validates fundamental skills in AI infrastructure and operations learned from Study Guide
15% Troubleshoot and Optimize
- Identify and troubleshoot hardware faults (e.g., GPU, fan, network card)
- Identify faulty cards, GPUs, power supplies
- Replace faulty cards, GPUs, power supplies
- Optimize AMD and Intel servers for performance
- Optimize storage
17% Systems and Network
- Configure routing tables on InfiniBand and NVIDIA Spectrum-X™
- Install and configure NVIDIA NVLink™ Switch
- Set up network fabric ports for the hosts
- Identify network topologies for data centers
33% Systems and Servers
- Install GPU-based servers
- Install physical GPUs
- Install NVIDIA® Bluefield® DPU-based servers
- Identify cable types and transceivers
- Validate hardware operation for workloads
- Validate hardware installation
- Validate power and cooling
- Establish storage requirements in a cluster design
35% Physical Layer Management
-
Install, update, and remove NVIDIA GPU drivers
-
Install the NVIDIA Container Toolkit
-
Demonstrate how to use NVIDIA GPUs with Docker
-
Install NGC command line interface on hosts
-
Configure and manage Bluefield
-
Configure MIG (AI and HPC)
-
Deploy the Bluefield OS image to Arm
-
Manage cloud-native stack
-
7-hour $150 AI Infrastructure Operations Fundamentals with exam coupon. This covers compute platforms, networking, and storage solutions. The course also addresses AI operations, focusing on infrastructure management and cluster orchestration.
-
7-sessions 4-hours each $3500 hands-on AI Infrastructure Professional Public Training explores configuration, management and troubleshooting of AI Infrastructure.
-
NVIDIA-Certified Professional: AI Infrastructure (NCP-AII), for $400 answer 50 questions in 90-minutes to validates the ability to deploy, manage, and maintain AI infrastructure by NVIDIA.
-
NVIDIA-Certified Professional: AI Operations (NCP-AIO) has 2-3 year preprequisite. For $400, answer 50 questions in 90-minutes to validate your ability to monitor, troubleshoot, and optimize AI infrastructure by NVIDIA.
36% Administration
-
Administer Fleet Command
-
Administer Slurm cluster
-
Understand data center architecture for AI workloads
-
Administer Base Command Manager (BCM) and cluster provisioning
-
Administer Run.ai (potentially part of ACM)
-
Configure MIG (for AI and HPC) 16% Workload Management
-
Administer Kubernetes cluster
-
Use system management tools to troubleshoot issues 26% Installation and Deployment
-
Install and configure BCM
-
Install and initialize Kubernetes on NVIDIA hosts using BCM
-
Deploy containers from NGC
-
Deploy cloud VMI containers
-
Understand storage requirements for AI data centers
-
Deploy DOCA services on DPU Arm 20% Troubleshooting and Optimization
-
Troubleshoot docker
-
Troubleshoot the fabric manager service for NVIDIA NVlink™/NVswitch™ systems
-
Troubleshoot BCM
-
Troubleshoot Magnum IO components
-
Troubleshoot storage performance
-
7-hour $150 AI Infrastructure & Operations Fundamentals includes exam certificate. covers essential components of AI infrastructure, including compute platforms, networking, and storage solutions. The course also addresses AI operations, focusing on infrastructure management and cluster orchestration.
-
$3,000 for six 4-hour session AI Operations Professional Public Training for hands-on experience with NVIDIA's DCGM, InfiniBand networking, NVIDIA BlueField™ DPUs, and GPU virtualization, while learning to leverage tools for infrastructure provisioning, workload scheduling, and cluster orchestration.
-
-
NVIDIA-Certified Professional: InfiniBand (NCP-IB). For $220 answer 40 questions in 90-minutes to validate skills in AI networking by NVIDIA. Correctly answer 40 questions in 90-minutes online, for $220, with a 2-year validity period for those who installs, configures, manages, troubleshoots, or monitors InfiniBand fabrics.
https://developer.nvidia.com/embedded/jetson-modules
All Jetson https://developer.nvidia.com/buy-jetson?product=all&location=US
Developer Kits:
https://developer.nvidia.com/buy-jetson?product=all&location=US
The NVIDIA Jetson Orin™ Nano Super Developer Kit is a compact, yet powerful computer that redefines generative AI for small edge devices. At just USD $249, it provides developers, students, and makers with the most affordable and accessible platform, backed by the support of NVIDIA AI software and a broad AI software ecosystem. Learn more
-
Jetson Orin Nano (8GB LPDDR5 RAM) 1024-core ARM CPU with 32 Tensor Cores
- $249 USD/$369 Euros from https://www.amazon.com/dp/B0BZJTQ5YP?th=1 has Athlon 6-core ARM CPU
- https://www.arrow.com/en/products/945-13766-0000-000/nvidia
- Datasheet https://static6.arrow.com/aropdfconversion/e5f9455a906908a5cd69a1b5f187a8e543689c1d/jetson-orin-datasheet-nano-developer-kit-3575392-r24.pdf
- https://www.sparkfun.com/products/22098
- https://www.seeedstudio.com/NVIDIAr-Jetson-Orintm-Nano-Developer-Kit-p-5617.html
-
Jetson AGX Orin (64GB) 2048-core ARM CPU with 64 Tensor Cores
- 64 GB eMMC (+ NVMeSSD)
- $1,999 USD https://www.seeedstudio.com/NVIDIArJetson-AGX-Orintm-64GB-Developer-Kit-p-5641.html https://www.youtube.com/watch?v=eFgsOeHMAW4
-
Jetson AGX Orin (32GB)
- 64 GB eMMC (+ NVMeSSD)
BTW: AGX is "not an acronym persay, but it loosely means Autonomous machines accelerator technology."
Others:
- Jetson Orin NX (16GB)
- reComputer J1020 v2 - Edge AI Computer with NVIDIA® Jetson Nano 4GB (SKU 110061441)
Previous :
- NVIDIA® Jetson AGX Xavier was the first generation of Jetson AGX platform. Released in 2019 and is now EOL.
- 40-pin expansion header
- x16 PCIe Slot support x8 PCIe Gen4
- Micro SD card slot
- two MIPI CSI connectors supporting camera modules with up to 4-lanes, allowing higher resolution and frame rate than before.
NVIDIA has a different SDK for different hardware
-
https://developer.nvidia.com/embedded/learn/jetson-agx-orin-devkit-user-guide/index.html Jetson AGX Orin Developer Kit User Guide
-
https://developer.nvidia.com/embedded/learn/jetson-orin-nano-devkit-user-guide/index.html Jetson Orin Nano Developer Kit User Guide
-
https://docs.nvidia.com/jetson/archives/r36.4/DeveloperGuide/index.html Linux Developer Guide
https://www.nvidia.com/en-au/glossary/
-
BSP
-
CUDA is NVIDIA's proprietary software for parallel computing on GPUs. Its competitor is Intel's DPC++ (Data Parallel C++).
CUDA 12.6
-
CuOPT
-
CSP & 3P Service
-
GPU = Graphics Processing Unit
- Turing GPUs (e.g., T4, Quadro RTX series)
- Ampere GPUs (e.g., RTX30 series, A30/40/100)
- Ada Lovelace GPUs (e.g., RTX40 series)
- Hopper GPUs (e.g., H100/H200)
-
HPC = High Performance Computing
-
JetPack 6.1 Cannonical Ubuntu Linux OS Kernel - https://developer.nvidia.com/embedded/jetpack supports the Jetson Orin Nano Super Developer Kit, featuring [MAXN mode] which boosts AI compute performance for the Jetson Orin Nano Developer Kit. https://developer.nvidia.com/blog/nvidia-jetson-orin-nano-developer-kit-gets-a-super-boost/
- Supports CUDA-X accelerated libraries and GPU APIs
- https://developer.nvidia.com/blog/nvidia-jetson-orin-nano-developer-kit-gets-a-super-boost/
-
Jetson is NVIDIA's proprietary GPU computing platform
- https://docs.nvidia.com/jetson/ = Jetson Software Documentation
-
Jetson Linux 36.4 provides the Linux Kernel 5.15, UEFI based bootloader, Ubuntu 22.04 based root file system, NVIDIA drivers, necessary firmwares, toolchain and more.
-
Jetson Platform Services (available soon.) is a collection of pre-built and cloud-native software services and reference workflows to accelerate AI applications on Jetson. These services are modular, API-driven and can be quickly configured to build Generative AI and other edge applications. There are 15+ services from AI services to system services. The services include:
-
NVIDIA SDK Manager
-
NVIDIA Jetpack SDK - https://docs.nvidia.com/jetson/archives/jetpack-archived/jetpack-61/install-setup/index.html#upgradable-compute-stack https://docs.nvidia.com/jetson/archives/jetpack-archived/jetpack-61/install-setup/index.html#package-management-tool
-
Jetson AI Stack
-
Isaac Framework for building high performance robotic applications https://developer.nvidia.com/isaac
-
Metropolis application framework to build, deploy and scale Vision AI application https://www.nvidia.com/en-us/autonomous-machines/intelligent-video-analytics-platform/
-
Holoscan for building high performance computing applications (HPC) with real time insights and sensor processing capabilities from edge to cloud. https://www.nvidia.com/en-us/clara/holoscan/
-
NVIDIA Omniverse™ Replicator for Synthetic Data Generation (SDG)
-
NVIDIA TAO Toolkit from data preparation to training to optimization, fine-tuning pretrained AI models from the NVIDIA NGC™ catalog of pre-trained models.
- TF2
- ODISE 1.1
-
OpenACC
-
OpenVLA (Vision-Language-Action) Model https://openvla.github.io
-
OpenUSD
-
TAO Toolkit https://developer.nvidia.com/tao-toolkit
-
Triton inference server
- CUDA 12.6
- TensorRT 10.3
- cuDNN 9.3
- VPI 3.2 VPI (Vision Programing Interface) is a unified API for computer vision and machine learning applications.
- DLA 3.1
- PVA
- ISP
- DLFW 24.0
- PDF: Jetson ORIN NX
- RIVA
- LoRA (Low-Rank Adaptation) to fine tune task-specific LLM models.
NVIDIA NIM, part of NVIDIA AI Enterprise, is a set of intuitive inference microservices designed to accelerate generative AI deployment in enterprises. NIM microservices provide interactive APIs to run inference on AI models.
- "Introduction to NVIDIA NIM™ Microservices" 2-hour Inference Microservices video course by Kevin Lee.
- VIDEO: by Mariya at Python Simplified
Each NIM is packaged as a Docker container image on a per model or model family basis.
NIM supports a wide range of AI models—including NVIDIA AI foundation, community, and custom—NIM ensures seamless, scalable AI inferencing, on-premises or in the cloud, all while leveraging industry-standard APIs.
NIM uses NVIDIA TensorRT-LLM to optimize the models, with specialized accelerated profiles optimally selected for:
- NVIDIA H100 Tensor Core GPUs,
- NVIDIA A100 Tensor Core GPUs,
- NVIDIA A10 Tensor Core GPUs,
- NVIDIA L40S GPUs.
- OpenUSD https://www.openusd.org 3D scenes USD scene description data files on Stages, Hydra rendering architecture, Prims (primatives hierarchy of objects from geometry, to materials, to lights and other organizational elements.), and Attributes https://learn.nvidia.com/courses/course-detail?course_id=course-v1:DLI+S-OV-17+V1
-
https://learn.nvidia.com/courses/course-detail?course_id=course-v1:DLI+S-OV-19+V1
-
rendering backend, such as OpenGL or DirectX. OpenUSD is an open-source USD library for creating and working with USD scenes.
-
HdStorm is included in OpenUSD
-
Each file format can be created through Python bindings in the OpenUSD library. When creating a new stage we can pass in a string to represent a file name that ends in .usdc, .usd, .usda, or .usdz. File Formats (USD, USDC, USDA and USDZ) are used for storing and exchanging various types of 3D scene data, including meshes, cameras, lights, and shaders.
-
A USD (.usd) file can have either ASCII or binary format. This switch can be done at any point without breaking references for debugging.
Separate heavier data from more light weight data. When doing so, consider using .usdc and .usda explicitly to avoid obfuscation and create large .usda files unintentionally.
-
USDA (.usda) is a native file format used by OpenUSD to store and exchange 3D scene data. Its format is ASCII text and therefore "Human Readable" and editable. This makes USDA optimal for small files, such as a stage that is referencing external content.
-
USDC (.usdc) - the Crate Binary Format -- is a compressed binary file format designed to minimize load time and provide a more efficient representation of the scene data compared to the human-readable ASCII format (USDA). USDC is extremely efficient for numerically-heavy data, like geometry. Various compression techniques reduce the file size and improve loading performance. It also employs memory mapping for faster file access and loading times.
-
USDZ (.usdz) is an atomic, uncompressed, zipped archive for delivery of all necessary assets ( a mesh with its texture) together in a single file. It’s generally intended as read-only and is optimal for XR experiences. We would not use USDZ if we are still making edits to the asset.
-
We may choose to use some other 3D format backed by an SdfFileFormatPlugin when we prefer to keep our source data as is and still leverage all of OpenUSD for scene manipulation and rendering.
https://www.nvidia.com/gtc/pricing/?nvid=nv-int-unbr-171401 Exhibits March 18–21 | Workshops March 16–20 | San Jose, CA & Virtual
https://forums.developer.nvidia.com/c/agx-autonomous-machines/jetson-embedded-systems/70 NVIDIA Community
TwitterX
https://www.youtube.com/@NVIDIADeveloper YouTube
- https://www.youtube.com/watch?v=mgUrthfw3ys
- https://www.youtube.com/watch?v=QHBr8hekCzg Dave's Garage
https://developer.nvidia.com/embedded/learn/get-started-jetson-orin-nano-devkit The NVIDIA® Jetson Orin Nano™ Developer Kit empowers the development of AI-powered robots, smart drones, and intelligent cameras built on the Jetson Orin series.
https://learn.nvidia.com/en-us/training/self-paced-courses
https://learn.nvidia.com/courses/course-detail?course_id=course-v1:DLI+C-RX-02+V2
https://www.nvidia.com/en-us/training/ DLI (DEEP LEARNING Institute)
https://www.jetson-ai-lab.com/tutorial-intro.html
https://www.jetson-ai-lab.com/ros.html
The 22GB for nano_llm:humble container image ros2_nanollm package provides ROS2 nodes for running optimized LLM's and VLM's locally inside a container. These are built on NanoLLM and ROS2 Humble for deploying generative AI models onboard your robot with Jetson.
-
Download Jetson Orin Nano Super Developer Kit https://developer.nvidia.com/downloads/embedded/L4T/r36_Release_v4.0/jp61-rev1-orin-nano-sd-card-image.zip
-
Download JETSON ORIN NANO DEVELOPER KIT SD card image from https://developer.nvidia.com/embedded/jetpack
https://docs.nvidia.com/jetson/archives/r36.4/DeveloperGuide/SD/Security/FirmwareTPM.html Firmware-based Trusted Platform Module (fTPM) on the Orin platform. Refer to the security page for all security features.
sudo apt dist-upgrade sudo apt-install nvidia-jetpack
https://www.youtube.com/watch?v=N_OOfkEWcOk Within https://github.com/NVIDIA/GenerativeAIExamples https://github.com/NVIDIA/GenerativeAIExamples/tree/main/community/5_mins_rag_no_gpu Run using Streamlit: