Edge computing has naturally been a hot topic at CES with companies highlighting a myriad of use cases where the pre-trained edge device runs inference locally to produce the desired output, never once interacting with the cloud. The complexity of these nodes has grown to not only include multimodal support with the fusion and collaboration between sensors for context-aware devices but also multiple cores to ratchet up the compute power.
Naturally, any hardware acceleration has become desirable with embedded engineers craving solutions that ease the design and development burden. The solutions vary where many veer towards developing applications with servers in the cloud that are then virtualized or containerized to run at the edge. Ultimately, there is no one-size-fits-all solution for any edge compute application.
It is clear that support for some kind of hardware acceleration has become paramount for success in breaking into the intelligent embedded edge. Company approaches to the problem run the full gamut from hardware accelerated MCUs with abundant software support and reference code, to an embedded NPU.
Table 1 highlights this with a list of a few companies and their hardware acceleration support.
Company | Hardware acceleration | Implemented in | Throughput | Software |
NXP | eIQ Neutron NPU | select MCX, i.MX RT crossover MCUs, and i.MX applications processors | 32 Ops/cycle to over 10,000 Ops/cycle | eIQ Toolkit, eIQ Time Series Studio |
STMicroelectronics | Neural-ART Accelerator NPU | STM32N6 | up to 600 GOPS | ST Edge AI Suite |
Renesas | DRP-AI | RZ/V2MA, RZ/V2L, RZ/V2M | – | DRP-AI Translator, DRP-AI TVM |
Silicon Labs | Matrix Vector Processor, AI/ML co-processor | BG24 and MG24 | – | MVP Math Library API, partnership with Edge Impulse |
TI | NPU | TMS320F28P55x, F29H85x, C2000 and more | Up to 1200 MOPS (on 4bWx8bD) Up to 600 MOPS (on 8bWx8bD) | Model Composer GUI or Tiny ML Modelmaker |
Synaptics | NPU | Astra (SL1640, SL1680) | 1.6 to 7.9 TOPS | Open software with complete GitHub project |
Infineon | Arm Ethos-U55 micro-NPU processor | PSOC Edge MCU series, E81, E83 and E84 | – | ModusToolbox |
Microchip | AI-accelerated MCU, MPU, DSC, or FPGA | 8-, 16- and 32-bit MCUs, MPUs, dsPIC33 DSCs, and FPGAs | – | MPLAB Machine Learning Development Suite, VectorBlox Accelerator Software Development (for FPGAs) |
Qualcomm | Hexagon NPU | Oryon CPU, Adreno GPU | 45 TOPS | Qualcomm Hexagon SDK |
Table 1: Various company’s approaches for hardware acceleration.
Synaptics, for instance, has their Astra platform that is beginning to incorporate Google’s multi-level intermediate representation (MLIR) framework. “The core itself is supposed to take in models and operate in a general-purpose sense. It’s sort of like an open RISC-V core based system but we’re adding an engine alongside it, so the compiler decides whether it goes to the engine or whether it works in a general-purpose sense.” said Vikram Gupta, senior VP and general manager of IoT processors and chief product officer, “We made a conscious choice that we wanted to go with open frameworks. So,whether it’s a Pytorch model or a TFLite model, it doesn’t matter. You can compile it to the MLIR representation, and then from there go to the back end of the engine.” One of their CES demos can be seen in Figure 1.
Figure 1: A smart camera solution showing the Grinn SoM that uses the Astra SL1680 and software from Arcturus to provide both identification and tracking. New faces are assigned an ID and an associated confidence interval that will adjust according to the distance from the camera itself.
TI showcased its TMS320F28P55x C2000 real-time controller (RTC) MCU series with an integrated NPU with an arc fault detection solution for solar inverter applications. The system performs power conversion while at the same time doing real-time arc fault detection using AI. The solution follows the standard process of obtaining data, labeling, and training the arc fault models that are then deployed onto the C2000 device (Figure 2).
Figure 2: TI’s solar arc fault detection edge AI solution
One of Microchip’s edge demos detected true touches in the presence water using its mTouch algorithm in combination with their PIC16LF1559 MCU (Figure 3). Another solution highlighted was in partnership with Edge Impulse and used the FOMO ML architecture to perform object detection in a truck loading bay. Other companies, such as Nordic Semiconductor, have also partnered with Edge Impulse to ease the process of labeling, training, and deploying AI to their hardware. The company has also eased the process of leveraging NVIDIA TAO models to adapt well-established AI models to a specific end-application on any Edge-Impulse-supported target hardware.
Figure 3: Some of Microchip’s edge AI solutions at CES 2025. Truck loading bay augmented by AI in partnership with Edge Impulse (left) and a custom-tailored Microchip solution using their mTouch algorithm to differentiate between touch and water (right).
Aalyia Shaukat, associate editor at EDN, has worked in the design publishing industry for six years. She holds a Bachelor’s degree in electrical engineering from Rochester Institute of Technology, and has published works in major EE journals as well as trade publications.
Related Content
- CES 2025: A Chat with Siemens EDA CEO Mike Ellow
- Software-defined vehicle (SDV): A technology to watch in 2025
- ADI’s efforts for a wirelessly upgraded software-defined vehicle
- CES 2025: Moving toward software-defined vehicles
googletag.cmd.push(function() { googletag.display(‘div-gpt-ad-native’); });
–>
The post CES 2025: Approaches towards hardware acceleration appeared first on EDN.