Visit our Content Hub!
Access free downloadable content curated by our editors.

Can Nvidia’s AI Platform Enable OEMs to Build Next-Generation Packaging Machines?

Nvidia has released a low-cost embedded AI computer optimized for robotics and vision, backed by a powerful software eco-system. Can OEMs use it to build the next generation of AI-powered packaging and processing machines?

Can this palm-sized embedded AI computer revolutionize how machine builders incorporate generative AI-based vision and robotics into their equipment?
Can this palm-sized embedded AI computer revolutionize how machine builders incorporate generative AI-based vision and robotics into their equipment?

For those tuned into the AI arms race among the Big Three, Open AI’s ChatGPT, Google’s Gemini and Anthropic’s Claude, it’s been a non-stop stream of one-upmanship as their language models and underlying platforms evolve at a dizzying pace. However, one development that was easy to overlook was Nvidia’s December 2024 release of its newest AI computer, the Jetson Orin Nano Super Developer Kit, essentially an AI embedded computer. (Nvidia has an excellent blog post that provides a good overview of the device, along with a video of Nvidia’s founder and CEO Jensen Huang talking about the latest technology.)

What does this have to do with packaging and processing machine designers? Potentially, plenty. So much so, that for this column, we are going to depart from our usual editorial standard of writing about one supplier’s offerings. (We always reserve the right for detailed product or vendor coverage for significant developments, as we have done recently for Rockwell Automation, Siemens and Schneider Electric’s incorporation of generative AI into their platforms.)

Don’t let the new (lower) price of only $249, the term “developer kit”, and the palm-of-your-hand size fool you. This is a full-fledged AI computer designed from the ground up to give generative AI capabilities to real-world devices, including industrial equipment such as packaging and processing machines. In fact, the computer is specifically designed around not only enabling applications with embedded large language models, but also embedded vision language models and robotics models.  This can be used as an embedded computer inside a traditional controls cabinet (DIN-rail mountable with a separately purchased DIN rail adapter).

Whereas traditional robotics favors pre-programmed rules and simple sensor inputs, this device is said to enable robots to use more sophisticated, AI-based methods for perceiving and understanding the physical environment. The kit supports what Nvidia calls visual simultaneous localization and mapping (SLAM) which allows a robot to compute its location and movement from images by tracking visual features around its environment. Further, advanced image detection capabilities enables a more nuanced understanding of the robot's surroundings.

There’s no question that these robotic capabilities are geared towards mobile robots navigating physical space, and that traditional fixed packaging lines have different needs. Nonetheless, at a minimum these advanced capabilities bear investigating for use in packaging machinery. In the best case scenario, these advanced AI capabilities suggest a more intelligent, adaptive and context-aware robot. But significant adaptation may be required, as fixed robotics differ drastically in hardware configuration and integration. That said, packaging-specific use cases might focus more on real-time vision inspection, high-speed product, package or component tracking, or advanced pattern recognition—less about navigation and more about throughput and reliable detection of defects on a moving line. 

This device can run not only large language models, but also pre-trained "vision language models", which is basically a way to pair objects inside of a visual field with textual annotations or captioning, enabling visual question answering, image-text matching, visual reasoning, and more. The Jetson Orin Nano Super supports no less than six off-the-shelf VLMs out of the box, allowing machine designers to select the most suitable model based on their application's requirements. The kit is designed to integrate VLMs into robotics and vision AI applications, leveraging optimized libraries like NanoLLM for high-performance inference of VLMs on the Jetson platform.

There are some caveats with vision models. Although the unit supports half a dozen off-the-shelf vision language models, it’s not as simple as plug and play. Each model typically requires what could be a significant degree of tuning or adaptation. Off-the-shelf models may work well for broad object detection tasks, but specialized packaging applications usually demand customized datasets, domain-specific labels, and iterative validation to ensure reliable performance on high-speed lines or visually challenging materials, containers or components.

Finally, Nvidia is highlighting the kit's potential for "agentic AI", suggesting more autonomous, goal-oriented behavior, making decisions based on high-level objectives rather than following pre-defined rules. (However, given the transparency, traceability and regulatory requirements specific to packaging and processing, it’s more realistic that the AI will assist operators and technicians or handle specialized vision tasks rather than fully taking over packaging line operations any time soon.)

Not just hardware, but software

Though Nvidia itself is often thought of as “just” a chip manufacturer, that belies its true footprint when it comes to AI more generally, and vision and robotics specifically. In the context of the Jetson Orin Nano, the company includes a variety of mature frameworks (many of which have been iterating since the device's first release in 2019) to support AI-powered vision applications, including:

DeepStream: This framework is specifically designed for vision AI applications, making it ideal for developing smart cameras and intelligent video analytics systems.