Entdecken Sie, wie Edge AI die Echtzeit-, sichere und effiziente KI-Verarbeitung auf Geräten ermöglicht und Branchen wie das Gesundheitswesen und autonome Fahrzeuge transformiert.
Edge AI refers to the deployment of artificial intelligence (AI) algorithms and models directly on local hardware devices—such as smartphones, IoT sensors, drones, and connected vehicles—rather than relying on centralized cloud computing centers. This decentralized approach allows data to be processed at the source of its creation, significantly reducing the latency involved in sending information back and forth to remote servers. By executing machine learning (ML) tasks locally, devices can make instantaneous decisions, operate reliably without internet connectivity, and enhance data privacy by keeping sensitive information on the device itself.
The core of Edge AI involves running an inference engine on an embedded system. Because edge devices typically have limited battery life and computational power compared to cloud servers, the AI models must be highly efficient. Developers often employ techniques like model quantization or model pruning to compress large neural networks without sacrificing significant accuracy.
Specialized hardware accelerators are frequently used to handle these workloads efficiently. Examples include the NVIDIA Jetson platform for robotics and the Google Coral Edge TPU for low-power inference. Software frameworks also play a vital role; tools like TensorRT and TFLite optimize models specifically for these constrained environments, ensuring fast real-time inference.
While the terms are often used interchangeably, it is helpful to distinguish between them:
Edge AI is transforming industries by enabling autonomous decision-making in critical scenarios:
Deploying a model to the edge typically involves training a model in a high-compute environment and then exporting it to a format compatible with edge devices, such as ONNX or OpenVINO. The Ultralytics Platform simplifies this workflow, allowing users to train and automatically export models for various edge targets.
The following example demonstrates how to export a lightweight YOLO26 model—specifically designed for efficiency—to a format suitable for mobile and edge deployment.
from ultralytics import YOLO
# Load the YOLO26 Nano model, which is optimized for speed on edge devices
model = YOLO("yolo26n.pt")
# Export the model to TFLite format for deployment on Android or Raspberry Pi
# This creates a 'yolo26n.tflite' file ready for edge inference
model.export(format="tflite")
Advanced edge deployments often utilize containerization technologies like Docker to package applications, ensuring they run consistently across different device architectures, from Raspberry Pi units to industrial gateways.