A curated list of efficient on-device AI systems, including practical inference engines, benchmarks, and state-of-the-art research papers for mobile and edge devices.
This repository bridges the gap between Systems Research (academic papers) and Practical Deployment (engineering frameworks), focusing on optimizing ML models (e.g., LLM/VLMs, ViTs, etc.) on resource-constrained hardware.
- 🚀 Inference Engines
- 📝 Research Papers
Frameworks and runtimes designed for deploying models on edge devices.
- LiteRT (formerly TensorFlow Lite) - Google's framework for on-device inference.
- ExecuTorch - PyTorch’s end-to-end solution for enabling on-device AI.
- ONNX Runtime - Cross-platform inference engine for ONNX models.
- MNN - Lightweight deep learning framework by Alibaba.
- NCNN - High-performance NN inference framework by Tencent.
- llama.cpp - LLM inference in C/C++ with minimal dependencies.
- MLC LLM - Universal solution for deploying LLMs on any hardware (based on TVM).
- mllm - A fast and lightweight LLM inference engine for mobile and edge devices.
- OmniInfer - High-performance, on-device VLM inference with hybrid NPU acceleration.
- RunAnywhere - Open-source SDK for running LLMs and multimodal models on-device across iOS, Android, and cross-platform apps.
- Qualcomm QNN - Qualcomm AI Stack for Snapdragon NPUs/DSPs.
- Apple Core ML - Framework to integrate ML models into iOS/macOS apps.
- FluidAudio - Local audio AI SDK for Apple platforms with ASR, speaker diarization, VAD, and TTS optimized for Apple Neural Engine.
- NVIDIA TensorRT - SDK for high-performance deep learning inference on NVIDIA GPUs (including Jetson).
- Intel OpenVINO - Toolkit for optimizing and deploying AI inference on Intel hardware (CPU/GPU/NPU).
- MediaTek NeuroPilot - AI ecosystem and SDK for MediaTek NPUs.
Note: Some of the works are designed for inference acceleration on cloud/server infrastructure, which has much higher computational resources, but I also include them here if they can be potentially generalized to on-device inference use cases.
- [MLSys 2025] MAS-Attention: Memory-Aware Stream Processing for Attention Acceleration on Resource-Constrained Edge Devices
- [MLSys 2025] TurboAttention: Efficient attention approximation for High Throughputs LLMs
- [ASPLOS 2023] FLAT: An Optimized Dataflow for Mitigating Attention Bottlenecks
- [NeurIPS 2022] FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
- [SenSys 2026] LLM as a System Service on Mobile Devices
- [EuroSys 2026] Scaling LLM Test-Time Compute with Mobile NPU on Smartphones
- [SOSP 2025] Characterizing Mobile SoC for Accelerating Heterogeneous LLM Inference
- [ASPLOS 2025] Neuralink: Fast on-Device LLM Inference with Neuron Co-Activation Linking
- [ASPLOS 2025] Fast On-device LLM Inference with NPUs
- [arXiv 2024] PowerInfer-2: Fast Large Language Model Inference on a Smartphone
- [MLSys 2025] TileLink: Generating Efficient Compute-Communication Overlapping Kernels using Tile-Centric Primitives
- [ASPLOS 2024] SmartMem: Layout Transformation Elimination and Adaptation for Efficient DNN Execution on Mobile
- [ASPLOS 2024] SoD2: Statically Optimizing Dynamic Deep Neural Network Execution
- [MICRO 2023] Improving Data Reuse in NPU On-chip Memory with Interleaved Gradient Order for DNN Training
- [MICRO 2022] GCD2: A Globally Optimizing Compiler for Mapping DNNs to Mobile DSPs
- [PLDI 2021] DNNFusion: Accelerating Deep Neural Networks Execution with Advanced Operator Fusion
- [MLSys 2024] AWQ: Activation-aware Weight Quantization for On-Device LLM Compression and Acceleration
- [ISCA 2023] OliVe: Accelerating Large Language Models via Hardware-friendly Outlier-Victim Pair Quantization
- [MobiSys 2025] ARIA: Optimizing Vision Foundation Model Inference on Heterogeneous Mobile Processors for Augmented Reality
- [PPoPP 2024] Shared Memory-contention-aware Concurrent DNN Execution for Diversely Heterogeneous SoCs
- [MobiSys 2024] Pantheon: Preemptible Multi-DNN Inference on Mobile Edge GPUs
- [MobiCom 2024] Perceptual-Centric Image Super-Resolution using Heterogeneous Processors on Mobile Devices
- [Sensys 2023] Miriam: Exploiting Elastic Kernels for Real-time Multi-DNN Inference on Edge GPU
- [MobiSys 2023] NN-Stretch: Automatic Neural Network Branching for Parallel Inference on Heterogeneous Multi-Processors
- [ATC 2023] Decentralized Application-Level Adaptive Scheduling for Multi-Instance DNNs on Open Mobile Devices
- [IPSN 2023] PointSplit: Towards On-device 3D Object Detection with Heterogeneous Low-power Accelerators
- [SenSys 2022] BlastNet: Exploiting Duo-Blocks for Cross-Processor Real-Time DNN Inference
- [MobiSys 2022] Band: Coordinated Multi-DNN Inference on Heterogeneous Mobile Processors
- [MobiSys 2022] CoDL: efficient CPU-GPU co-execution for deep learning inference on mobile devices
- [ICCAD 2025] Mitigating Resource Contention for Responsive On-device Machine Learning Inferences
- [RTSS 2024] FLEX: Adaptive Task Batch Scheduling with Elastic Fusion in Multi-Modal Multi-View Machine Perception
- [MobiCom 2024] Panopticus: Omnidirectional 3D Object Detection on Resource-constrained Edge Devices
- [MobiSys 2023] OmniLive: Super-Resolution Enhanced 360° Video Live Streaming for Mobile Devices
- [MobiSys 2023] HarvNet: Resource-Optimized Operation of Multi-Exit Deep Neural Networks on Energy Harvesting Devices
- [MobiCom 2022] NeuLens: Spatial-based Dynamic Acceleration of Convolutional Neural Networks on Edge
- [MobiCom 2021] Flexible high-resolution object detection on edge devices with tunable latency
- [ASPLOS 2025] Nazar: Monitoring and Adapting ML Models on Mobile Devices
- [SenSys 2024] AdaShadow: Responsive Test-time Model Adaptation in Non-stationary Mobile Environments
- [SenSys 2023] EdgeFM: Leveraging Foundation Model for Open-set Learning on the Edge
- [MobiCom 2023] Cost-effective On-device Continual Learning over Memory Hierarchy with Miro
- [MobiCom 2023] AdaptiveNet: Post-deployment Neural Architecture Adaptation for Diverse Edge Environments
- [MobiSys 2023] ElasticTrainer: Speeding Up On-Device Training with Runtime Elastic Tensor Selection
- [SenSys 2023] On-NAS: On-Device Neural Architecture Search on Memory-Constrained Intelligent Embedded Systems
- [MobiCom 2022] Mandheling: mixed-precision on-device DNN training with DSP offloading
- [MobiSys 2022] Memory-efficient DNN training on mobile devices