The Hexagon NPU (Neural Processing Unit) is a dedicated hardware module designed by Qualcomm to meet the computational demands of artificial intelligence (AI). It is specifically built for low power and high performance AI inference tasks, playing a crucial role in generative AI applications on the edge (such as smartphones, PCs, and other devices). The following sections will provide a detailed analysis of its architecture, performance, application scenarios, and advantages.
1. Evolution of Hexagon NPU Architecture
The architecture of the Hexagon NPU has undergone multiple iterations and upgrades, evolving from the initial Hexagon DSP (Digital Signal Processor) to a dedicated processor that supports AI inference. Its core architecture integrates scalar, vector, and tensor accelerators, supporting various integer and floating-point precision operations, including INT4, INT8, INT16, and FP16. This architectural design enables it to efficiently handle complex AI models such as Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), and Transformers.
In the third-generation Snapdragon 8 mobile platform, the Hexagon NPU has been further upgraded, introducing Micro Tile Inferencing and support for local 4-bit integer (INT4) operations, significantly enhancing the performance and energy efficiency of generative AI. Additionally, the Hexagon NPU is equipped with dedicated power delivery tracks that can dynamically adjust power supply based on workload, achieving a balance between performance and power consumption.

2. Performance
The Hexagon NPU demonstrates outstanding performance, especially when handling large models (such as generative AI), outperforming traditional CPUs and GPUs in both performance and energy efficiency. For instance, the Hexagon NPU in the third-generation Snapdragon 8 has achieved a 98% performance improvement compared to its predecessor, while reducing power consumption by 40%. This leap in performance is primarily attributed to its new microarchitecture design, which includes higher clock speeds, greater bandwidth, and more efficient memory management.
In practical applications, the Hexagon NPU can achieve rapid AI inference with extremely low power consumption. For example, the Hexagon NPU in the Xiaomi 14 smartphone can complete inference of the MobileNet-V2 model in 0.6 milliseconds, which is 23 times faster than a mobile CPU and 3.2 times faster than a mobile GPU. This performance advantage makes the Hexagon NPU an ideal choice for edge AI applications.
3. Application Scenarios
The Hexagon NPU is widely used in various AI scenarios, including but not limited to:
- Generative AI The Hexagon NPU can support models with up to 10 billion parameters, achieving industry-leading speeds for both the generation of the first token and the rate of token generation per second. This makes it excel in generative AI tasks such as natural language processing, image generation, and audio synthesis.
- Image and Video Processing The Hexagon NPU supports AI imaging and AI video tasks, efficiently handling operations such as image recognition, enhancement, and compression.
- Speech and Audio Processing Originally designed for audio signal processing, the Hexagon NPU can efficiently execute tasks such as speech recognition and speech synthesis.
- Sensor Data Processing The Hexagon NPU also supports real-time data processing for sensing hubs, enhancing the perception capabilities of devices.
4. Differentiated Advantages
The core advantage of the Hexagon NPU lies in its system-level solutions, custom design, and rapid innovation. Qualcomm’s custom design of the NPU and control over the Instruction Set Architecture (ISA) allow for quick design evolution and expansion to address bottleneck issues and optimize performance. This customized design enables the Hexagon NPU to flexibly adjust across different AI tasks, achieving optimal performance.
Moreover, the heterogeneous computing architecture of the Hexagon NPU allows it to work in conjunction with other processors such as the Adreno GPU, Kryo, or Oryon CPU, forming an efficient AI computing platform. This heterogeneous computing architecture not only enhances overall performance but also optimizes memory bandwidth efficiency and reduces power consumption.
5. Future Prospects
As the complexity and parameter scale of generative AI models continue to increase, the evolution of the Hexagon NPU becomes particularly important. Qualcomm is continuously optimizing its NPU architecture to support more complex AI models and a wider range of application scenarios. For example, the latest version of the Hexagon NPU can now support Transformer networks, making it excel in handling large-scale language models. In the future, the Hexagon NPU is expected to enable broader AI applications across more edge devices, promoting the proliferation and development of AI endpoints.