Edge AI acceleration with Raspberry Pi

Machine Learning Workshop – Raspberry Pi AI Kit

© Lead Image © robsnowstock, 123RF.com

© Lead Image © robsnowstock, 123RF.com

Article from Issue 288/2024
Author(s):

Raspberry Pi enters the artificial intelligence accelerator fray with a low-cost solution.

Public clouds are doubtlessly the most convenient and cost-effective tool for individuals to use to access the latest neural processing unit (NPU) artificial intelligence (AI) accelerating hardware. Due to its high cost and lack of availability, it is hard for a freelance developer to justify acquiring one of the now famous (and infamous) NVIDIA A100 [1] Ampere cards with tensor processing units, with costs running into the tens of thousands of dollars and the cards destined for inevitable obsolescence as the newer Hopper (NVIDIA H100) [2] units arrive (as I am writing, I can see A100s advertised for $1.29/hour by a cloud vendor). The economic choice is easy: Develop the code offline and only then provision a souped-up cloud instance for the actual training or inference task.

Low-cost development strategies comprise Google's Colab [3], a hosted Jupyter Notebook environment providing free access to GPU and TPU (Google's own AI chip) resources for research and learning, or purchasing older accelerators being sold secondhand at heavily discounted prices (with varying degrees of residual usefulness). At the Dragon Propulsion Laboratory, we are partial to another strategy: edge AI accelerators. Designed to enhance devices at the edge of the network, this class of chips is both low in power demands and affordable cost-wise, as well as worth understanding in its own right (Figure 1).

First Look

The Raspberry Pi AI Kit just became available, giving me the opportunity to present a first look at what promises to be a very interesting new entry in the edge accelerator class. The AI Kit consists of the standard Raspberry Pi M.2 HAT+ daughterboard and a Hailo-8L AI accelerator. Connections to the host Pi 5 board run through the GPIO pins and a dedicated ribbon connector, resulting in a single-lane PCIe 3.0 connection with 8GBbps of bandwidth (Figure 2). Rated at 13 Tera Operations per Second (TOPS) of inference performance, the accelerator should offload the Pi's CPU entirely of neural network execution, leaving it free to respond to its results, as hinted by object recognition, segmentation, and tracking demos showcased in the launch announcement [4].

[...]

Use Express-Checkout link below to read the full article (PDF).

Buy this article as PDF

Express-Checkout as PDF
Price $2.95
(incl. VAT)

Buy Linux Magazine

SINGLE ISSUES
 
SUBSCRIPTIONS
 
TABLET & SMARTPHONE APPS
Get it on Google Play

US / Canada

Get it on Google Play

UK / Australia

Related content

  • Raspberry Pi Pet Camera

    A Raspberry Pi, a Pi-compatible camera, and a mesh VPN are all you need to watch your pets from afar.

  • DIY Lenticular Camera

    You can take lenticular images with a homemade camera to re-create the "wiggle" pictures of your childhood.

  • Wildlife Photo Trap

    Armed with no more than a Raspberry Pi photo trap, you can discover who pays a visit to your garden at night.

  • TensorFlow AI on the Pi

    You don't need a powerful computer system to use AI. We show what it takes to benefit from AI on the Raspberry Pi and what tasks the small computer can handle.

  • NVidia Driver Install

    A terminal-based solution helps ease the frustration of installing NVIDIA drivers.

comments powered by Disqus
Subscribe to our Linux Newsletters
Find Linux and Open Source Jobs
Subscribe to our ADMIN Newsletters

Support Our Work

Linux Magazine content is made possible with support from readers like you. Please consider contributing when you’ve found an article to be beneficial.

Learn More

News