NVIDIA kicked off their line of GPU-accelerated single board computers back in 2014 with the Jetson TK1, a $200 USD development system for those looking to get involved with the burgeoning world of so-called “edge computing”. It was designed to put high performance computing in a small and energy efficient enough package that it could be integrated directly into products, rather than connecting to a data center half-way across the world.
The TK1 was an impressive piece of hardware, but not something the hacker and maker community was necessarily interested in. For one thing, it was fairly expensive. But perhaps more importantly, it was clearly geared more towards industry types than consumers. We did see the occasional project using the TK1 and the subsequent TX1 and TX2 boards, but they were few and far between.
Then came the Jetson Nano. Its 128 core Maxwell CPU still packed plenty of power and was fully compatible with NVIDIA’s CUDA architecture, but its smaller size and $99 price tag made it far more attractive for hobbyists. According to the company’s own figures, the number of active Jetson developers has more than tripled since the Nano’s introduction in March of 2019. With the platform accessible to a larger and more diverse group of users, new and innovative applications for machine learning started pouring in.
Cutting the price of the entry level Jetson hardware in half was clearly a step in the right direction, but NVIDIA wanted to bring even more developers into the fray. So why not see if lightning can strike twice? Today they’ve officially announced that the new Jetson Nano 2GB will go on sale later this month for just $59. Let’s take a close look at this new iteration of the Nano to see what’s changed (and what hasn’t) from last year’s model.
To be clear the new Jetson Nano 2GB is not a new device, it’s essentially just a cost optimized version of the hardware that was released back in 2019. It’s still the same size, draws the same amount of power, and has the exact same Maxwell GPU. In broad terms, it’s a drop-in replacement for the more expensive Nano. In fact, it’s so similar that you might not even be able to tell the difference between the two models at first. Especially since the biggest change isn’t visible: as the name implies, the new model only has two gigabytes of RAM compared to four in the original Nano.
The board has lost a few ports as part of the effort to get it down to half the original price, however. The Nano 2GB drops the DisplayPort for HDMI (the previous version had both), deletes the second CSI camera connector, does away with the M.2 slot, and reduces the number of USB ports from four to three. Losing a USB port probably isn’t a deal breaker for most applications, but if you need high-speed data, it’s worth noting that only one of them is 3.0. Overall, it seems clear that NVIDIA took a close look at the sort of devices that folks were connecting to their Nano and adjusted the type and number of ports accordingly.
Of course, the 40 pin header on the side remains unchanged so the new board should remain pin-compatible with anything you’ve already built. The Gigabit Ethernet port is still there, but unfortunately wireless still didn’t make the cut this time around. So if you need WiFi for your project, count on one of those USB ports being permanently taken up with a dongle.
It’s not just slimmed down, but updated as well. The 2GB removes the old school DC barrel jack and replaces it with a USB-C port. On the original Nano you could run it off of the micro USB port for most tasks, but it was recommended to use a laptop style power supply if you were going to be pushing the hardware. Now you can just use a 15 watt USB-C power supply and be covered in all situations.
Since the hardware is nearly identical between the two versions of the Nano, there’s really no point running any new benchmarks on it. If your software worked on the $99 Nano, it will run just as well on the $59 one. Or at least, that’s the idea. In reality, having only half the available RAM might be a problem for some applications.
NVIDIA sent me a review unit so as a simple test, I ran the detectnet.py script that makes up part of NVIDIA’s AI training course on the live video from a Logitech C270 camera. While the Nano maintained a respectable 22 to 24 frames per second, the system ran out of RAM almost immediately and had to dip into swap to keep up. Naturally this is pretty problematic on an SD card, and certainly not something you’d want to do for any extended period of time unless you happen to own SanDisk stock.
To help combat this, NVIDIA recommends disabling the GUI on the Nano 2GB and running headless if you’re planning on doing any computationally intensive tasks. That should save you 200 to 300 MB of memory, but obviously isn’t going to work in all situations. It’s also a bit counter-intuitive considering the default Ubuntu 18.04 system image boots directly into a graphical environment. It’ll be interesting to see if some lightweight operating system choices are offered down the line to help address this issue.
It’s probably not fair to call the Jetson Nano 2GB a direct competitor to the Raspberry Pi, but clearly NVIDIA wants to close the gap. While the lack of built-in WiFi and Bluetooth will likely give many makers pause, there’s no question that the Nano will run circles around the Pi 4 if you’re looking to experiment with things like computer vision. At $99 that might not have mattered for budget-conscious hardware hackers, but now that the Nano is essentially the same price as the mid-range Pi 4, it’s going to be a harder decision to make.
Especially since NVIDIA is using the release of the new board to help kick off the Jetson AI Certification Program. This free self-paced source is comprised of tutorials and video walkthroughs that cover everything from the fundamentals of training up to practical applications like collision avoidance and object following. To complete the Jetson AI Specialist course and be granted the certification, applicants will have to submit an open source project to NVIDIA’s Community Projects forum for review and approval.
If you wanted to get your feet wet with AI and machine learning, picking up a Jetson Nano at $99 was a great choice. Now that there’s a $59 version that includes access to a training and certification program, there’s barely even a choice left to make. Which in the end, is exactly what NVIDIA wants.
I admit, the initial price of the Jetson kept me from even considering the purchase of one.
Time will tell if one is in my future.
“According to the company’s own figures, the number of active Jetson developers has more than tripped since the Nano’s”
Would really like to see a simple cluster board for the jetson nano sometime soon. would love to have a jetson nano cluster running k8s + all those compute cores readily accessible in my closet for any software project needs.
Broadcom sells PCIe switches that support host-to-host communication and sharing of peripherals. Given that the module has a 4 lane PCIe gen 2 port, it would be cool to see a bunch of these in a cluster tied together with that fabric locally and talking with other cluster boards over 10GE.
This is why I’ll never consider nvidia for anything, no matter how cheap they make it.
I mean, if you demand fully open source you’re out of the current SBC market until RISC-V starts hitting shelves. Raspberry Pi, Beaglebone, Udoo, Pine64, Orenge Pi, every SBC has closed-source portions.
Honestly out of the loop here: what’s closed nowadays on the Tegra line of parts? I remember the rage over Tegra2 years ago having everything under NDA and nVidia in a constant war with hackers over the bootloader on tablets.
Has that changed? What’s still a blob or precompiled library? Are docs public or under lock and key?
Are there any better options? RPi has improved since the early days, and scalar performance is now better than the Jetson Nano, but its GPGPU is still nowhere near anything nVidia sells and isn’t a platform that is scalable to mass production.
is there a more open alternative for embedded which actually works without endless headaches and has real documentation? Nvidia is pretty good on providing documentation and learning resources. I’d say that more than offsets whatever closedness there is in CUDA. It’s enough to have to deal with compatibility issues within tensorflow versions itself than to also have to spend time on getting it to work with whatever GPU you have.
Nvidia if you’re reading this keep doing what you’re doing. The machine vision community doesn’t need more script kiddies.
Please make a sample project in visual studio though, no one likes eclipse. A friend of mine got a project up and running in visual studio with visualgdb and it works quite nicely. Give people options, I don’t know a single dev that is stoked about eclipse. Please don’t force it on people.
Sounds like they’re putting their lack of ARM royalty payments and their soon to be incoming ARM royalty payments to good use. Or are at least betting their purchase won’t get blocked so they can subsidize this board from funds received from other vendors.
By using our website and services, you expressly agree to the placement of our performance, functionality and advertising cookies. Learn more
Nvidia Jetson, Computer, Software development kit, Artificial intelligence, Raspberry Pi
World news – THAT – NVIDIA Announces $59 Jetson Nano 2GB, A Single Board Computer With Makers In Mind