What if you could use powerful AI models directly on your smartphone—without sending a single byte of data to the cloud?
That’s exactly what Google AI Edge Gallery delivers. This open-source application lets you run large language models (LLMs) and multimodal AI models locally on Android and iOS devices. No remote servers, no hidden data collection, no constant internet connection required.
Even better, AI Edge Gallery is released under the Apache 2.0 license, which means the project is truly open and future-proof. Even if Google were to step away one day, the code would remain available for the community to maintain, fork, and improve.
And clearly, people care about this approach. The app surpassed 500,000 downloads in just two months after its GitHub release, proving that on-device AI and privacy-first tools are no longer niche—they’re becoming mainstream.

What Is Google AI Edge Gallery?
AI Edge Gallery is a mobile-first AI runtime and playground designed to showcase what modern smartphones can do with on-device machine learning.
Instead of relying on cloud APIs like ChatGPT or Gemini, all inference happens entirely on your phone. Conversations, image analysis, audio transcription—everything stays local.
This makes it especially appealing to:
- Privacy-conscious users
- Professionals handling sensitive documents
- Developers experimenting with mobile AI
- Anyone who wants AI access even when offline
Key Features: What Can You Do with AI Edge Gallery?
AI Edge Gallery covers a surprisingly wide range of everyday and advanced use cases.






🔹 AI Chat (Offline LLM Chat)
Chat with a local LLM just like you would with ChatGPT—but without an internet connection.
Perfect for:
- Brainstorming ideas
- Writing emails or notes
- Asking technical questions
- Using AI while traveling or offline
All conversations remain stored only on your device.
🔹 Ask Image (On-Device Image Analysis)
Take a photo or upload an image and ask the AI to explain what it sees.
Use cases include:
- Identifying plants or objects
- Understanding diagrams or charts
- Extracting information from invoices or receipts
Since everything runs locally, your photos never leave your phone.
🔹 Audio Scribe (Speech-to-Text + Translation)
Record audio—meetings, interviews, or voice notes—and instantly convert it into text.
Recent updates added:
- On-device translation into other languages
- Improved transcription accuracy
This makes AI Edge Gallery a strong offline alternative to cloud-based transcription tools.
🔹 Prompt Lab (For Developers & Power Users)
Prompt Lab is a testing and benchmarking environment for experimenting with different models.
It provides real-time performance metrics such as:
- Time to first token
- Decoding speed
- Latency and responsiveness
If you care about optimization and model efficiency, this feature alone is worth exploring.
🔹 Tiny Garden (Experimental Offline AI Game)
Tiny Garden is a fun experimental mini-game where you interact using natural language to:
- Plant seeds
- Water crops
- Harvest flowers
It’s mostly a playful demo, but it clearly illustrates how language-driven interactions can work fully offline.
🔹 Mobile Actions (Experimental Device Control)
For advanced users, Mobile Actions allows you to:
- Fine-tune a model using open-source recipes
- Load it into the app
- Control certain phone functions offline
This feature is still experimental, but it hints at a future where local AI assistants can interact deeply with your device—without cloud access.
Supported AI Models
AI Edge Gallery supports a growing ecosystem of on-device models, including both Google and third-party options.
Google Models
- Gemma 3 (1B & 4B parameters)
- Gemma 3n, optimized for lower-end devices and now supporting audio

Third-Party Models
- Qwen 2.5
- Phi-4 Mini (Microsoft)
- DeepSeek-R1 (for more advanced reasoning tasks)
Specialized Models
- TranslateGemma – supports translation across 55 languages
- FunctionGemma – designed for function calling and structured outputs
All models run via LiteRT, Google’s lightweight on-device inference runtime.
Thanks to community support, Hugging Face already hosts many models converted to LiteRT format. Advanced users can even load custom models using the .litertlm format.
How to Install AI Edge Gallery on Android
Installing on Android is straightforward:
- Search for “AI Edge Gallery” on the Google Play Store
- Or download the APK directly from GitHub releases
Requirements:
- Android 12 or later
- Minimum 4 GB RAM (8 GB recommended for larger models)
On first launch, the app will prompt you to download AI models.
Model sizes range from 500 MB to 4 GB, depending on complexity. Once downloaded, they are stored locally and work entirely offline.
What About iOS and macOS?
iOS
- Available via TestFlight (beta)
- Limited to 10,000 testers
- Requires at least 6 GB of RAM
The app is functional but still missing a few features. Google is targeting an official App Store release in early 2026.
macOS
There is currently no native macOS version.
If you really want to test it on a Mac, you can use:
- Android Studio’s built-in emulator
- BlueStacks (BlueStacks Air is optimized for Apple Silicon)
That said, macOS users who want local LLMs should seriously consider Ollama or LM Studio, both of which offer native support and a better desktop experience.
Why AI Edge Gallery Actually Matters
Privacy First
Your data never touches external servers.
No cloud uploads, no third-party processing, no hidden telemetry.
This is especially valuable if you:
- Work with confidential documents
- Analyze personal photos or audio
- Simply don’t trust cloud AI services
Works Completely Offline
AI Edge Gallery doesn’t care if you’re:
- On a plane
- In the subway
- Traveling in low-coverage areas
No network latency. No “server overloaded” messages. Your AI is always available.
Open Source and Future-Proof
Because the project is open source, the community can:
- Add new models
- Improve performance
- Fix bugs
- Fork the project if needed
Even if Google loses interest (which wouldn’t be unprecedented), the software won’t disappear.
Final Thoughts
Google AI Edge Gallery is a clear signal of where AI is heading: powerful, private, and local.
It won’t replace cloud-based AI platforms overnight, but for privacy-focused users, developers, and offline use cases, it’s already one of the most compelling on-device AI solutions available today.
If you want to explore advanced usage, custom models, or LiteRT APIs, the project’s GitHub wiki offers extensive documentation and technical guides.
Local AI is no longer a future concept—it’s already in your pocket.
And if you'd like to go a step further in supporting us, you can treat us to a virtual coffee ☕️. Thank you for your support ❤️!
We do not support or promote any form of piracy, copyright infringement, or illegal use of software, video content, or digital resources.
Any mention of third-party sites, tools, or platforms is purely for informational purposes. It is the responsibility of each reader to comply with the laws in their country, as well as the terms of use of the services mentioned.
We strongly encourage the use of legal, open-source, or official solutions in a responsible manner.


Comments