Edge AI

Edge AI runs inference on-device to reduce latency and preserve privacy, emphasizing optimized models and efficient runtimes. Techniques include model compression, on-device accelerators, and careful resource management for limited-memory and low-power environments.