Neural Processing Unit (NPU): Accelerating neural network inference

dc.contributor.authorEmery, Stéphane
dc.date.accessioned2026-02-26T15:49:33Z
dc.date.available2026-02-26T15:49:33Z
dc.date.issued2026
dc.description.abstractArtificial intelligence (AI) algorithms, such as neural networks, are fundamental to many advanced computer vision and signal processing capabilities, often exceeding human-level performance. These algorithms enable the automation of repetitive tasks by processing input data through millions of computations, achieving high accuracy and reliable performance. However, the significant memory and power requirements for processing these networks pose challenges for deployment in miniaturized, wearable, and other energy-constrained applications. Local processing of AI/ML algorithms within the end node, known as edge processing, can enhance the performance of many applications. However, unless a large battery size or frequent recharging can be afforded, embedded AI/ML processing is restricted to low complexity tasks or necessitates offloading to cloud processing. This offloading incurs the costs of energy-intensive radio communications, increased latency, and additional privacy concerns. Optimized and energy-efficient AI/ML chips address these challenges by accelerating neural network computations within a constrained power budget, thereby enhancing edge processing capabilities. CSEM’s next-generation neural processing unit (NPU) has been designed to address these challenges, enabling neural network edge processing in power- and energy-constrained embedded systems. The NPU is a standalone AI/ML accelerator IP that delivers state-of-the-art ML acceleration performance, optimized for embedded edge processing. With nearly 200x higher measured throughput (910GOP/s) and significantly increased efficiency (3.5TOPS/W) than its predecessor, CSEM’s latest AI/ML system-on-chip showcases the NPU’s cutting-edge performance for low-power AI chips.
dc.identifier.citationNeural Processing Unit: Accelerating neural network inference, ASICs for the Edge, EMERY Stéphane
dc.identifier.urihttps://hdl.handle.net/20.500.12839/1998
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 International*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/*
dc.titleNeural Processing Unit (NPU): Accelerating neural network inference
dc.typeTechnical Factsheet
dc.type.csemdivisionsBU-M
dc.type.csemresearchareasASICs for the Edge
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
NPU-Neural-Processing-Unit-white-paper.pdf
Size:
869.42 KB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.82 KB
Format:
Item-specific license agreed upon to submission
Description: