Open Source · Edge AI · Mobile First

AI That Runs Anywhere.

A lightweight, open-source AI assistant that runs locally on phones, Raspberry Pi, and low-RAM devices — no cloud, no subscriptions, no beefy GPU required.

⭐ View on GitHub 💬 Join Discord
~512MB
Minimum RAM
100%
Offline / Local
MIT
License
Android
+ Linux
Platforms
// Memory Footprint

Other tools need a server room.
We need a pocket.

GPT-4 (cloud) Cloud-only · $$$/month
NOT LOCAL
Llama 3 70B (full) ~140GB RAM required
140 GB+
Competing edge tools ~4–8 GB RAM
4–8 GB
NanoMind ✓ 512 MB – 1 GB
512 MB
📱 Android 8+
🍓 Raspberry Pi 4
💻 Old Laptops
🐧 Linux ARM
🪟 Windows (low-spec)
📟 IoT Boards
// Head to Head

Why NanoMind wins
on the edge.

Feature
Big LLM Tools
⚡ NanoMind
Min RAM Required
8 GB – 140 GB
512 MB
Works on Phone
No
Yes (Android)
100% Offline
Mostly cloud
Fully local
Open Source
Varies
MIT License
Privacy / No Telemetry
Data sent to cloud
Zero data leaves device
Install Complexity
High (Docker, GPU drivers)
Single binary / APK
Model Swappable
Limited
Any GGUF model
// Zero Config. One Command.

From download to
running AI in 60 seconds.

$ curl -fsSL https://nanomind.sh/install | sh
→ Detecting device... Raspberry Pi 4 (4GB RAM)
→ Fetching optimal quantized model (Q4_K_M)...
→ Model size: 1.1 GB | Download: ████████░░ 78%
✓ NanoMind ready. Memory used: 891 MB / 4096 MB
 
$ nanomind chat
NanoMind 0.1.0 — running locally on your device
You: Explain quantum computing in simple terms
AI: Quantum computing uses quantum bits (qubits) that can exist...
// Architecture

Built for constrained
environments from day one.

🧠

4-bit Quantization

Models compressed to Q4_K_M format via llama.cpp — 75% size reduction with minimal quality loss.

CPU-Only Inference

No GPU required. Optimized SIMD/NEON instructions squeeze maximum speed out of any ARM or x86 chip.

📱

Native Android Runtime

JNI bridge to llama.cpp. No Termux. No hacks. A real APK that works out of the box.

🔒

Zero Telemetry

No analytics, no crash reports sent out. Your data, conversations, and models stay on your device.

🔌

REST API Built-In

OpenAI-compatible local API. Plug NanoMind into any app or script already using the OpenAI SDK.

📦

Plug Any GGUF Model

Not locked to one model. Use Mistral, Phi-3, Gemma, or any community-quantized model from HuggingFace.

// Roadmap

From v0.1 to
global standard.

01

Android APK + CLI

Ship a working Android app and Linux CLI binary. Basic chat, model download, offline inference.

02

Model Hub

One-tap model download optimized per device class. No manual GGUF hunting on HuggingFace.

03

Plugin System

Web search, RAG on local files, voice input. Community-built extensions via an open plugin API.

04

iOS + Windows

Expand to iOS TestFlight and Windows ARM. Every major platform covered.

Built in public.
Owned by everyone.

NanoMind is 100% open source under the MIT license. Fork it, improve it, redistribute it. This belongs to the community.

View Source on GitHub