On-device shell command generator for macOS Tahoe. Uses Apple's 3B model with dynamic few-shot retrieval from 21k tldr examples.
-
Updated
Apr 21, 2026 - Python
On-device shell command generator for macOS Tahoe. Uses Apple's 3B model with dynamic few-shot retrieval from 21k tldr examples.
Agentic Android Open Source Project (AAOSP) — Android fork with native LLM system service, MCP-aware apps, and an agent-driven launcher. On-device Qwen 2.5 via llama.cpp. Apps declare tools in their manifest. The OS runs the model.
Android 16 fork. AI as a platform primitive. Twelve capabilities, one shared runtime, every app. OEM-pluggable. Apache 2.0.
KnoLo Core is a local-first knowledge base engine built for small language models (LLMs). It packages your documents into a compact .knolo file and enables fully deterministic querying — no embeddings, no vector databases, no cloud services required. Designed for on-device and edge LLM deployments.
High-performance Android SDK for on-device LLM inference (GGUF). Privacy-focused, offline-first, and powered by llama.cpp with a clean Kotlin Coroutines API.
Self-hosted lifelong AI companion — hatches once, imprints on you, grows alongside you for life. Daily rituals, persistent memory, per-user on-device QLoRA, Apache 2.0.
Documentation for MobileTransformers - a lightweight, modular framework based on ONNX Runtime for running and adapting large language models (LLMs) directly on mobile and edge devices. It supports on-device fine-tuning (PEFT), efficient inference, quantization, weight merging, and direct inference from merged models.
📱 手机端 AI 操作系统全景知识库 — 334+ 篇深度页面,覆盖端侧大模型、AI Agent、芯片适配、推理优化 | 自动更新
Add a description, image, and links to the on-device-llm topic page so that developers can more easily learn about it.
To associate your repository with the on-device-llm topic, visit your repo's landing page and select "manage topics."