The field of edge AI is moving towards increased efficiency and scalability with the development of hardware-aware neural networks and novel number formats. Researchers are focusing on designing models and architectures that can optimize performance, power consumption, and area usage. Notable advancements include the proposal of a variable-point number format for efficient multiplication of high-dynamic-range numbers and the development of hardware-aware neural architecture search frameworks for early exiting networks on edge accelerators. Furthermore, the creation of NPU-native vision-language models and the design of near-memory architectures for event-based computer vision are pushing the boundaries of edge AI applications. These innovations have the potential to enable real-time and low-latency processing on resource-constrained devices, making them suitable for deployment in various edge environments. Particularly noteworthy papers include the proposal of the HDAP framework for hardware-aware DNN compression and the development of the AutoNeural architecture for NPU-native vision-language models.
Advancements in Hardware-Aware Neural Networks and Edge AI
Sources
From RISC-V Cores to Neuromorphic Arrays: A Tutorial on Building Scalable Digital Neuromorphic Processors
Ternary-Input Binary-Weight CNN Accelerator Design for Miniature Object Classification System with Query-Driven Spatial DVS
Variable Point: A Number Format for Area- and Energy-Efficient Multiplication of High-Dynamic-Range Numbers
Adversarial Robustness of Traffic Classification under Resource Constraints: Input Structure Matters