Run AI on any device. No PC. No subscription. No struggle.
llamdrop is a free, open-source tool that lets anyone run a local AI model on whatever device they own โ an Android phone, an old laptop, a Raspberry Pi, a budget PC, even a gaming console running Linux.
It reads your hardware automatically, finds AI models that will actually work on your specs, downloads the right one, and runs it. You don't need to know what quantization means. You don't need to read any documentation. You just run it.
llamdrop will always be completely free. It cannot be sold. Ever. That's not a promise โ it's written into the license (GPL v3).
This project was born from a real experience โ spending hours trying to run local AI on a phone with no PC, no budget, and no guidance. Dozens of crashes, incompatible models, RAM errors that made no sense.
llamdrop is for anyone on low-end or budget hardware who keeps getting left out:
- ๐ฑ Phone users โ Android via Termux, no PC needed
- ๐ป Old laptop owners โ that 2012 laptop collecting dust can run AI
- ๐ Raspberry Pi / SBC users โ Pi 4, Pi 5, Orange Pi, etc.
- ๐ฎ Console / embedded Linux users โ if it runs Linux, llamdrop runs on it
- ๐ธ People who can't afford ChatGPT, Claude, or Gemini subscriptions
- ๐ Users in regions where $20/month is not a small amount
- ๐งโ๐ Students and self-learners wanting to experiment with AI for free
- ๐ง Developers and tinkerers who want to test local AI on constrained hardware
If you've ever given up trying to run local AI because it was too complicated, crashed too many times, or cost too much โ this is for you.
- ๐ Auto device detection โ reads your RAM, CPU, OS without asking you anything
- ๐ Smart model browser โ two modes:
- โ Verified catalog โ curated models confirmed working on low-end devices
- ๐ Live HuggingFace search โ search any model, with live RAM estimates
- โฌ๏ธ Smart downloader โ auto-picks the right quantization for your RAM, shows progress, resumes if interrupted
- ๐ Auto-tuned launcher โ sets threads, context size, batch size for your exact device
- ๐ฌ Stable chat โ automatic context trimming prevents out-of-memory crashes
- ๐พ Session save/load โ resume conversations where you left off
โ ๏ธ RAM monitor โ live warning if memory gets dangerous during chat
llamdrop uses a two-layer model system:
A community-maintained list of models confirmed to work on low-RAM devices. Every entry has been tested, has known RAM requirements, and is safe to download. This is what most users should use.
Search any model on HuggingFace directly from llamdrop. The tool estimates RAM requirements from file size and quantization type. Clearly marked as unverified โ for experienced users who want to explore beyond the catalog.
Current verified model tiers:
| Tier | Available RAM | Example Models |
|---|---|---|
| 1 โ Ultra low | 1.5 โ 3 GB | Qwen2.5-0.5B, SmolLM2-360M, Qwen2.5-1.5B Q2 |
| 2 โ Standard | 3 โ 5 GB | Qwen2.5-1.5B Q4, Phi-3-mini, Gemma-2-2B, Llama-3.2-1B |
| 3 โ Better hardware | 5 โ 7 GB | Llama-3.2-3B, Qwen2.5-3B, Phi-3.5-mini |
All verified models are free, open-source, and downloadable without login or account.
llamdrop runs on any device that can run Python 3 in a Linux terminal.
| Platform | Status | Notes |
|---|---|---|
| Android via Termux | ๐ฏ Primary test platform | Built and tested here first |
| Linux laptop / desktop | โ Fully supported | Any distro, x86_64 or ARM64 |
| Raspberry Pi 4 / 5 | โ Fully supported | ARM64 |
| Old Windows PC (WSL) | โ Should work | Via Windows Subsystem for Linux |
| Chromebook (Linux mode) | ๐ Should work | ARM64 or x86_64 |
| Orange Pi / SBC | ๐ Should work | ARM64 Linux |
| iOS | โ Not supported | No proper terminal environment |
curl -sL https://raw.githubusercontent.com/ypatole035-ai/llamdrop/main/install.sh | bash
โ ๏ธ llamdrop v0.1 is under active development. The installer is not functional yet. Star and Watch this repo to get notified the moment it's ready.
llamdrop/
โโโ llamdrop.py # Main entry point
โโโ install.sh # One-line installer
โโโ models.json # Verified model catalog
โโโ modules/
โ โโโ device.py # Hardware detection (RAM, CPU, OS)
โ โโโ browser.py # Model browser โ verified + HF live search
โ โโโ downloader.py # Smart download with resume + quantization picker
โ โโโ launcher.py # llama.cpp wrapper with auto-tuned flags
โ โโโ chat.py # Chat loop with context trimming + RAM monitor
โโโ docs/
โโโ CONTRIBUTING.md # How to contribute
โโโ DEVICES.md # Community device compatibility list
- Device detection (RAM, CPU, OS, storage)
- Verified model browser with tier system
- Smart downloader with quantization auto-selection
- llama.cpp auto-installer
- Auto-tuned launcher
- Basic chat with context trimming
- Live HuggingFace model search with RAM estimates
- Session save and resume
- RAM live monitor during chat
- Vulkan GPU acceleration for supported devices
- Multi-language UI (Hindi, Spanish, Arabic, Portuguese)
- Better error messages in plain language
- Web-based model catalog (GitHub Pages)
- Community device profile submissions
- Automated model testing before catalog addition
You don't need to be a developer to contribute:
- ๐ฒ Test a model on your device โ open a PR to update
models.json - ๐ Translate the UI into your language
- ๐ Write a setup guide for your specific device
- ๐ Report a crash via GitHub Issues
- โญ Star this repo โ it helps others find it when they need it most
See CONTRIBUTING.md for full details.
GNU General Public License v3.0 โ see LICENSE
In plain language:
- โ Free to use forever
- โ Free to modify and share
- โ Cannot be sold
- โ Cannot be made closed-source
- โ Cannot be put behind a paywall
llamdrop will always be free. That is non-negotiable.
This project started because one developer spent hours trying to run local AI on an Oppo F19 Pro+ with no PC and no budget. Dozens of crashes. Models that were incompatible. RAM errors with no explanation. When it finally worked โ with a tiny 1.5B model running in Termux โ the thought was: nobody should have to go through all of that just to get started.
llamdrop is the tool that should have existed already.
Built by @ypatole035-ai and contributors. If llamdrop helped you, star the repo and share it with someone who needs it.