Run LLMs privately on iPhone, iPad, and Mac with Apple Silicon MLX optimization
Locally AI is a free, privacy-first application that lets you run large language models directly on your Apple devices without any internet connection or cloud processing. Built specifically for the Apple ecosystem, it leverages Apple's MLX machine learning framework to deliver optimized inference on Apple Silicon chips, achieving performance that rivals GPT-4 and GPT-4o-mini on capable devices like iPad Pro and Mac. The app supports a wide range of open-source models including Meta Llama 3.2 and 3.1, Google Gemma 2, 3, and 3n, Qwen 2.5, 3, and 3.5 with vision capabilities, DeepSeek R1, IBM Granite, Hugging Face SmolLM, Liquid Foundation Models, and Deep Cogito reasoning models. Both language and vision models are supported, enabling text generation and image analysis entirely on-device. Locally AI integrates deeply with the Apple ecosystem through Siri voice activation, Control Center and Lock Screen quick access, and Apple Shortcuts automation for building custom AI workflows. Real-time voice conversations are processed entirely on-device, ensuring complete privacy. The app requires no account creation, no login, and collects zero user data. With a 4.8-star rating from over 660 App Store reviews, Locally AI has earned praise for its elegant interface, strong Apple Silicon performance, and genuine commitment to user privacy. It requires iOS 18.0 or later for iPhone and iPad, and macOS 26.0 for Mac. The app is completely free with no in-app purchases or subscription fees, making advanced local AI accessible to anyone with a compatible Apple device.