Tag: ONNX

Edge AI with ONNX Runtime: Running Models On-Device

Posted on 6 min read

Last year, I deployed an AI model to a mobile device. The first attempt failed—the model was too large, inference was too slow, and battery drain was unacceptable. After optimizing 15+ models for edge deployment using ONNX Runtime, I’ve learned what works. Here’s the complete guide to running AI models on-device with ONNX Runtime. Figure… Continue reading