On-Prem

Systems

Nvidia upgrades tiny Jetson Orin Nano dev kits for the holidays

'Super' edition promises 67 TOPS and 102GB/s of memory bandwidth for your GenAI projects


Nvidia is bringing the AI hype home for the holidays with the launch of a tiny new dev board called the Jetson Orin Nano Super.

Nvidia cheekily bills the board as the "most affordable generative AI supercomputer," though that might be stretching the term quite a bit.

The dev kit consists of a system on module, similar to a Raspberry Pi Compute Module, that sits atop a reference carrier board for I/O and power. And, just like a Raspberry Pi, Nvidia's diminutive dev kit is aimed at developers and hobbyists looking to experiment with generative AI.

Similar to a Raspberry Pi Compute Module, Nvidia's Jetson Orin Nano Super development kit consists of a compute module containing the SoC that attaches to a carrier board for I/O and power. - Click to enlarge

Under the hood, the Orin Nano features six Arm Cortex-A78AE cores along with an Nvidia GPU based on its older Ampere architecture with 1024 CUDA cores and 32 tensor cores.

The hardware design appears to be identical to the original Jetson Orin Nano. However, Nvidia says the board now ships alongside upgraded software that unlocks additional performance and comes at a lower price of $249, compared to the original $499 price tag.

In terms of performance, the Jetson Orin Nano Super packs 67 TOPS at INT8, which is faster than the NPUs in any of Intel, AMD, or Qualcomm's latest AI PCs backed by 8GB of LPDDR5 memory capable of 102GB/s of memory bandwidth. According to Nvidia, these specs reflect a 70 percent uplift in performance and 50 percent more memory bandwidth than its predecessor.

The bandwidth boost is particularly important for those looking to play with the kind of large language models (LLMs) that power modern AI chatbots at home. At 102GB/s, we estimate the dev kit should be able to generate words at around 18-20 tokens a second when running a 4-bit quantized version of Meta's 8-billion-parameter Llama 3.1 model.

If you're curious about how TOPS, memory capacity, and bandwidth relate to model performance, you can check out our guide here.

For I/O, the dev kit's carrier board features the usual fare of connectivity for an SBC, including gigabit Ethernet, DisplayPort, four USB 3.2 Gen 2 type-A ports, USB-C, dual M.2 slots with M and E keys, along with a variety of expansion headers.

In terms of software support, you might think those Arm cores might be problematic; however, that really isn't the case. Nvidia has supported GPUs on Arm processors for years, with its most sophisticated designs - the GH200 and GB200 - utilizing its custom Arm-based Grace CPU. This means you can expect broad support for the GPU giant's software suite including Nvidia Isaac, Metropolis, and Holoscan, to name a few.

Along with a variety of open AI models available via Nvidia's Jetson AI lab, the dev kit also supports up to four cameras for robotics or vision processing workloads.

Alongside the new dev kit, Nvidia is also rolling out a software update to its older Jetson Orin NX and Nano system on modules, which it says should boost GenAI performance by 1.7x, so if you already picked up an Orin Nano, you shouldn't be missing out on too much. ®

Send us news
14 Comments

Take a closer look at Nvidia's buy of Run.ai, European Commission told

Campaign groups, non-profit orgs urge action to prevent GPU maker tightening grip on AI industry

AI's rising tide lifts all chips as AMD Instinct, cloudy silicon vie for a slice of Nvidia's pie

Analyst estimates show growing apetite for alternative infrastructure

Just how deep is Nvidia's CUDA moat really?

Not as impenetrable as you might think, but still more than Intel or AMD would like

Jury spares Qualcomm's AI PC ambitions, but Arm eyes a retrial

The victory may be short lived as the chip designer gears up for second round

US bipartisan group publishes laundry list of AI policy requests

Chair Jay Obernolte urges Congress to act – whether it will is another matter

Google Gemini 2.0 Flash comes out with real-time conversation, image analysis

Chocolate Factory's latest multimodal model aims to power more trusted AI agents

Cheat codes for LLM performance: An introduction to speculative decoding

Sometimes two models really are faster than one

Infosec experts divided on AI's potential to assist red teams

Yes, LLMs can do the heavy lifting. But good luck getting one to give evidence

Million GPU clusters, gigawatts of power – the scale of AI defies logic

It's not just one hyperbolic billionaire – the entire industry is chasing the AI dragon

Are you better value for money than AI?

Tech vendors start saying the quiet part out loud – do enterprises really need all that headcount?

Apple called on to ditch AI headline summaries after BBC debacle

'Facts can't be decided by a roll of the dice'

American cops are using AI to draft police reports, and the ACLU isn't happy

Do we really need to explain why this is a problem?