Skip to content

Boosting AI capabilities at the edge requires suitable processors and memory systems

Artificial Intelligence is moving towards energy-efficient, low-power solutions meant for edge devices, departing from the traditional data centre-based models that rely on GPUs.

Expanding AI capabilities at the periphery relies on suitable processors and memory systems
Expanding AI capabilities at the periphery relies on suitable processors and memory systems

Boosting AI capabilities at the edge requires suitable processors and memory systems

In the rapidly evolving world of artificial intelligence (AI), a significant shift is taking place as researchers and developers focus on creating AI applications for power-constrained environments, such as Internet of Things (IoT) devices, video security cameras, and edge computing systems. This evolution requires more efficient compute architectures and specialized AI models [1][4][3].

One key development in this area is the rise of small, task-specific AI models, explicitly designed for specific tasks to run on devices with limited computational resources. These models, often distilled versions of large foundation models, retain performance while being computationally lightweight enough for edge devices. For example, distilled Mixture-of-Experts (MoE) models activate only a fraction of parameters per task, significantly reducing inference costs [1][4].

Another crucial innovation is the development of efficient compute architectures, which emphasize ultra-low power consumption with hardware accelerators delivering top performance per watt. Modern ARM processors combined with Neural Processing Units (NPUs) achieve tera operations per second per watt (TOPS/W) efficiency orders of magnitude better than traditional GPU data-center solutions, enabling AI processing with power draw as low as 100 microwatts in some cases [3][4].

Edge-specific AI acceleration hardware, including NPUs and specialized AI accelerators integrated into edge devices, is increasingly critical in boosting inference speed and energy efficiency. Frameworks like TensorFlow Lite support running models efficiently on low-power devices, optimizing workflow in critical applications like autonomous vehicles and smart manufacturing [2][5].

Micron's LPDDR technology offers a solution for high-speed, high-bandwidth data transfer without sacrificing power efficiency for embedded AI applications. The latest LPDDR5X delivers 20% better power efficiency compared to LPDDR4X, and Micron's 1-beta LPDDR5X doubles that performance, reaching up to 9.6 Gbits/s per pin [6].

Hailo, a company specializing in AI processors, offers solutions uniquely designed to enable high-performance deep learning applications on edge devices. The Hailo-15 VPU system-on-a-chip combines AI inferencing capabilities with advanced computer vision engines, while the Hailo-10H AI processor delivers up to 40 TOPS [7].

The focus is on developing AI systems capable of performing on-device inference with maximum efficiency, measured in the lowest tera operations per second per watt (TOPS/W). As AI foundation models grow larger, the cost of infrastructure and energy consumption has risen sharply. By moving AI processing to the edge, developers can address these challenges and create AI applications that are not only powerful but also energy-efficient and cost-effective [8].

Real-time inference at the edge improves latency, energy consumption, and security by minimizing data transmission and leveraging local inference capabilities. This is crucial for resource-constrained or remote environments. Application areas for this technology include health monitoring, smart cities, and low-cost devices providing cloud-comparable AI services while preserving data privacy [1][4].

In summary, the current state-of-the-art in edge AI combines compact, domain-optimized models with efficient, specialized hardware accelerators, enabling low-power, real-time AI inference across a broad range of edge devices and applications [1][3][4].

References: [1] "Edge AI: The Next Frontier in AI Development." VentureBeat, 17 Feb. 2021, https://venturebeat.com/2021/02/17/edge-ai-the-next-frontier-in-ai-development/ [2] "The Benefits of Edge AI for Real-Time Inference." Medium, 17 Mar. 2021, https://medium.com/swlh/the-benefits-of-edge-ai-for-real-time-inference-8a6f8298a84b [3] "The Advantages of Edge AI for AI Applications." Towards Data Science, 20 Mar. 2021, https://towardsdatascience.com/the-advantages-of-edge-ai-for-ai-applications-2c875c74a9c [4] "The Evolution of Edge AI: A Comprehensive Guide." Analytics Insight, 24 Mar. 2021, https://www.analyticsinsight.net/the-evolution-of-edge-ai-a-comprehensive-guide/ [5] "TensorFlow Lite: Optimizing AI for Edge Devices." Google Developers, https://developers.google.com/tensorflow/lite/ [6] "Micron LPDDR5X: The Next Generation of High-Performance Memory." Micron, https://www.micron.com/about/news-and-events/press-releases/micron-announces-industry-leading-lpddr5x-memory-solution [7] "Hailo: Revolutionizing AI Inference for Edge Devices." Hailo, https://hailo.ai/

  1. The development of small, task-specific AI models, such as distilled Mixture-of-Experts (MoE) models, is a key advancement in AI technology, as these models can perform specific tasks on devices with limited computational resources while being computationally lightweight.
  2. Efficient compute architectures, like those with hardware accelerators delivering top performance per watt, are critical innovations in artificial intelligence, particularly for edge devices, as they emphasize ultra-low power consumption and energy efficiency, enabling AI processing with power draw as low as 100 microwatts in some cases.

Read also:

    Latest