Run NVIDIA Nemotron 3 Nano Omni locally in a single command
Canonical
on 28 April 2026
Today, NVIDIA introduced the NVIDIA Nemotron™ 3 Nano Omni, a highly-efficient multimodal model designed to understand and reason across video, audio, images, and language.
Canonical is enabling immediate access to Nemotrom 3 Nano Omni through inference snaps: pre-packaged AI inference runtimes distributed as snap packages for consistent deployment across systems. Developers and enterprises can deploy the model seamlessly across NVIDIA-enabled environments with a single command:
sudo snap install nemotron-3-nano-omni
With the Nemotron 3 Nano Omni delivered through inference snaps, deployment shifts from a complex integration task to a repeatable, standardized operation. A single install produces a consistent, production-grade runtime that scales across environments without rework, enabling teams to focus on building and operating agentic applications rather than managing infrastructure.
What is Nemotron 3 Nano Omni?
NVIDIA Nemotron 3 Nano Omni is an open multimodal foundation model that unifies reasoning across text, images, video, audio, and documents within a single architecture. It features a 256K token context window and a hybrid mixture of experts (MoE) architecture optimized for high multimodal throughput and accuracy.
Nemotron 3 Nano Omni powers perception sub-agents, giving agentic systems “eyes and ears” while maintaining a unified multimodal context across steps. There is no need to stitch together separate vision, speech, and language models, reducing latency and orchestration complexity.
Easy deployment with inference snaps
Inference snaps bundle models, dependencies, and execution environments so inference workloads can run reproducibly on edge devices, workstations or servers – without manual setup.
Canonical inference snaps provide a direct path from local installation to a production deployment. To get started, simply run:
sudo snap install nemotron-3-nano-omni
This installs a fully packaged, production-ready inference stack, including the model, runtime, and optimizations; no manual configuration required.
Inference snaps provide:
- Zero-friction deployment: no dependency resolution, no environment drift, no custom build pipelines
- Consistent runtime across environments: identical behavior on cloud, on-premises, and edge systems
- Secure, confined execution: strict isolation with automatic updates and verified distribution
- Optimized performance out of the box: pre-tuned for supported hardware
- Simplified operations: standardized packaging reduces maintenance, patching, and upgrade complexity
For enterprises, this translates into significantly faster deployment times, saving weeks of integration and validation work, and enabling scalable AI infrastructure with a single portable artifact.
Learn more at: https://github.com/canonical/inference-snaps
About Canonical
Canonical, the publisher of Ubuntu, provides open source security, support, and services. Our portfolio covers critical systems, from the smallest devices to the largest clouds, from the kernel to containers, from databases to AI. With customers that include top tech brands, emerging startups, governments and home users, Canonical delivers trusted open source for everyone.
Learn more at https://canonical.com/
Learn more
Find out more about Canonical’s collaboration with NVIDIA.
Enterprise AI, simplified
AI doesn’t have to be difficult. Accelerate innovation with an end-to-end stack that delivers all the open source tooling you need for the entire AI/ML lifecycle.
Newsletter signup
Related posts
Canonical welcomes NVIDIA’s donation of the GPU DRA driver to CNCF
At KubeCon Europe in Amsterdam, NVIDIA announced that it will donate the GPU Dynamic Resource Allocation (DRA) Driver to the Cloud Native Computing Foundation...
Meet Canonical at NVIDIA GTC 2026
Previewing at the event: NVIDIA CUDA support in Ubuntu 26.04 LTS, NVIDIA Vera Rubin NVL72 architecture support in Ubuntu 26.04 LTS, Canonical’s official...
Canonical announces Ubuntu support for the NVIDIA Rubin platform
Official Ubuntu support for the NVIDIA Rubin platform, including the NVIDIA Vera Rubin NVL72 rack-scale systems, announced at CES 2026 CES 2026, Las Vegas. –...