An open-source mobile app enables text generation, image creation, vision analysis, and speech transcription entirely offline using quantized LLMs and diffusion models.
{{IMAGE:1}}
The Off Grid mobile application represents a significant advancement in fully offline AI capabilities for smartphones. Unlike conventional AI apps requiring cloud connectivity, this open-source solution processes all data locally on mobile hardware, addressing privacy concerns while maintaining functionality in disconnected environments.
Developed by Ali Chherawalla, Off Grid integrates multiple AI subsystems into a cohesive interface:
Core Functionalities
Text Generation: Supports quantized versions of Qwen 3, Llama 3.2, Gemma 3, and Phi-4 models through GGUF format compatibility. The implementation achieves 15-30 tokens per second on flagship devices using efficient quantization techniques. Users can import custom GGUF models, though performance varies significantly with model size and quantization levels.
{{IMAGE:3}} Caption: Text generation interface showing real-time streaming responses
Image Generation: Incorporates on-device Stable Diffusion variants including Absolute Reality and DreamShaper. NPU acceleration on Qualcomm Snapdragon chips reduces generation time to 5-10 seconds, while Apple's Core ML optimizes iOS performance. The trade-off remains visual quality versus generation speed, with simpler models producing faster results.
{{IMAGE:4}} Caption: Image generation workflow with prompt enhancement and preview
Vision Capabilities: Implements multimodal models like SmolVLM and Qwen3-VL for real-time scene analysis. Users can point their camera at objects, documents, or receipts for contextual understanding (~7s processing on flagship devices). The system performs OCR and document analysis through native PDF text extraction libraries.
{{IMAGE:5}} Caption: Vision AI analyzing real-world objects through smartphone camera
Supporting Features:
- Voice Transcription: Utilizes Whisper.cpp for offline speech-to-text
- Document Processing: Handles PDFs, CSVs, and code files through system-level text extraction
- Prompt Enhancement: Automatically refines basic prompts into detailed diffusion instructions
Performance Considerations Benchmarks reveal hardware-dependent limitations:
| Task | Flagship Devices | Mid-range Devices |
|---|---|---|
| Text gen | 15-30 tok/s | 5-15 tok/s |
| Image gen (NPU) | 5-10s | N/A |
| Image gen (CPU) | ~15s | ~30s |
| Vision | ~7s | ~15s |
These metrics, verified on Snapdragon 8 Gen 2/3 and Apple A17 Pro platforms, highlight the computational constraints of mobile AI. Larger models or higher-quality outputs substantially increase processing time and energy consumption.
Technical Architecture The React Native application integrates specialized native modules:
- llama.rn: For GGUF model execution
- local-dream: Optimized Stable Diffusion implementation
- MNN inference: Cross-platform neural network acceleration
The technical documentation details the layered architecture separating UI components from native inference modules. This design allows platform-specific optimizations while maintaining a unified codebase.
Installation & Development Android users can install pre-built APKs from GitHub Releases. Building from source requires:
- Node.js 20+
- Android SDK 34 (for Android)
- Xcode 15+ (for iOS)
The repository provides comprehensive instructions for environment setup and compilation.
Practical Limitations While impressive for mobile deployment, several constraints persist:
- Model selection restricted to quantized versions with accuracy trade-offs
- No cloud fallback option for complex tasks
- Storage requirements (GGUF models average 3-7GB)
- Thermal throttling during extended sessions
As an open-source project (GitHub repository), Off Grid demonstrates the current boundaries of mobile AI. It delivers genuine offline functionality but requires hardware compromises atypical of cloud-based alternatives. For privacy-conscious users willing to accept these trade-offs, it represents a notable step toward truly personal AI assistants.
Comments
Please log in or register to join the discussion