LLM·Dex
On-deviceLocal LLMsEdge

On-Device LLMs Are Finally Here

Phi-4, Llama-4-8B, Qwen-2.5-7B running on a MacBook. What works, what doesn't, what's next for local inference.

By LLMDex Editorial

Three years ago, "running an LLM on your laptop" meant a heavily quantized 7B model that responded slower than dial-up and produced answers that felt like a worse version of GPT-3.5. Today, the same hardware runs Phi-4, Llama-4-8B, or Qwen-2.5-7B at 30+ tokens per second with quality that rivals 2023's frontier closed models. On-device LLMs aren't a research curiosity anymore. They're a real deployment target.

This article walks through what works in 2026, what doesn't, and how to set up a local LLM workflow that's actually useful day-to-day.

What's possible on consumer hardware

The single biggest shift since 2023 is that small models got dramatically better. The 7-8B parameter class is now where models like Phi-4, Llama-4-8B, and Qwen-2.5-7B live, and they all handily exceed GPT-3.5's quality on most tasks. Running them on consumer hardware:

  • Apple Silicon (M2 / M3 / M4), 16GB unified memory comfortably runs 7-8B models at 4-bit quantization at 30-50 tokens/sec. 32GB+ runs 13-14B at similar speeds.
  • Consumer NVIDIA (4090, 5090), 24GB VRAM runs 13-14B at FP8 or 32B at INT4. Speeds 60-100 tokens/sec.
  • Workstation NVIDIA (A100, H100), 70B-class models comfortably. Speeds 30-50 tokens/sec on a single H100.
  • Phones, 1-3B parameter models (Llama-3.2-3B, SmolLM2) at 4-bit. Useful for narrow tasks; quality is materially below 7B class.

For a typical engineer's MacBook (M3, 16-32GB), the practical sweet spot is a 4-bit quantized 8B model like Phi-4 or Llama-4-8B. Speeds feel like a fast cloud API. Quality is more than enough for most coding-completion, summarization, and general chat workloads.

What the tooling looks like

The on-device LLM ecosystem in 2026 is mature. Three tools cover 95% of users:

Ollama

The most popular on-device LLM runtime. Trivial install, simple CLI, REST API on localhost:11434. Pull a model (ollama pull phi-4), point your application at the API. Works on macOS, Linux, Windows.

For most developers, Ollama is the right starting point. It's not the fastest serving stack, but the developer experience is best-in-class.

LM Studio

A graphical app for browsing, downloading, and chatting with local models. More polished UX than Ollama for casual use. Same underlying llama.cpp engine. Integrates with most chat UIs that speak the OpenAI API.

llama.cpp directly

For speed-sensitive use cases (real-time voice, low-latency completion), running llama.cpp's llama-server directly with CUDA / Metal optimizations is the lowest-overhead path. More setup, faster inference.

Real workflows that work in 2026

Three on-device workflows that are genuinely useful:

1. Code completion

The Continue.dev or Cline VS Code extensions can be configured to use a local model for inline completion. With Phi-4 or Qwen-2.5-Coder-7B running on Ollama, you get GitHub-Copilot-equivalent autocomplete latency and quality without sending your code to a third party.

For IP-sensitive teams (legal, financial, defense), this is the only acceptable AI-coding setup. It's also a real cost saver, you're not paying per-token, and the throughput on a single machine is enough for several engineers.

2. Personal RAG

Run a local LLM, point it at your personal notes (Obsidian, plain text, PDFs), and you have a private knowledge assistant. Tools like LlamaIndex, LangChain, and Haystack all support local-model backends. Privex, Open WebUI, and AnythingLLM are user-facing apps that bundle the whole stack.

Quality is below cloud-frontier RAG (you're using an 8B model, not Gemini 3 Pro), but for personal knowledge bases the gap is acceptable and the privacy benefit is real.

3. Realtime voice assistants

Cartesia / Whisper STT + Phi-4 + ElevenLabs / Cartesia TTS, running locally, gives you a voice agent that responds in under 800ms with no internet dependency. The setup is non-trivial but documented in several open-source projects (Pipecat, Mac Whisper).

This is the workflow that's most exciting to us. On-device voice eliminates the latency floor that cloud setups can't avoid.

Where on-device falls short

Three failure modes:

1. Hard reasoning

Math, science, hard agent tasks require frontier-class quality that 7-8B models don't reach. Cloud GPT-5.5, Claude Opus, or DeepSeek-R1 win unambiguously here.

2. Long context

Most on-device models top out at 128K context. Loading and running long-context inference is RAM-bound; consumer hardware doesn't have the memory bandwidth. For long-doc workloads, cloud models with 200K-1M windows are necessary.

3. Multimodal

Vision-capable on-device models exist (Llama-3.2-90B-Vision, Qwen2-VL, Pixtral-12B), but the quality gap to closed-frontier vision is meaningful. For serious vision work, cloud is still the right answer.

The hardware question

If you're shopping for hardware specifically to run local LLMs, three rules:

  1. Apple Silicon is the value pick. Unified memory + Metal acceleration + low power = best laptop experience. M3 Max with 64GB or M4 Max with 96GB are the sweet spots.
  2. Consumer NVIDIA wins on raw speed if you have a desktop. RTX 4090 / 5090 with 24GB VRAM beats Apple Silicon on tokens/sec for 13-14B models.
  3. Don't buy a workstation just for LLMs. Cloud inference is cheap enough that an H100 in your closet doesn't pay back unless you're running 24/7 production workloads.

Picking a local model in 2026

Three opinions:

  • Phi-4 is the best quality-per-byte at the 14B size class. Microsoft's curated synthetic-data approach really did work.
  • Llama-4-8B is the best ecosystem support (broadest tooling, largest community). Slightly behind Phi-4 on raw quality.
  • Qwen-2.5-7B is the best multilingual / code-leaning option. Apache 2.0 licensed.
  • Qwen-2.5-Coder-32B if you have the hardware. Best open-weight code model.
  • Mistral-Nemo-12B for European-language workloads.

Browse the full Best Local LLMs ranking for context.

What's coming

Three trends to watch through 2026:

  • Mixture-of-experts in the small-model class. 12B MoE models with 3B active parameters could deliver dramatically better quality on the same hardware. Several labs have hinted at this.
  • Speculative decoding becoming default. A draft model + frontier verifier locally can give you cloud-class speeds with on-device privacy. The tooling is finally usable.
  • Hardware accelerators in laptops. Apple's Neural Engine and AMD/Intel NPU efforts are still immature for LLM inference but improving. Expect dedicated NPUs to matter more by end-2026.

The practical recommendation

If you're an engineer who hasn't tried on-device LLMs in 2026, you should. Install Ollama, pull phi-4, point your editor at it. Within an hour you'll have a setup that meaningfully changes your workflow, and the privacy properties are a bonus.

For teams shipping IP-sensitive products, on-device is no longer the constrained option. It's a real production target with mature tooling and acceptable quality for most workloads. The cloud frontier still wins on hardest tasks, but the gap on routine work is genuinely small.

Further reading

Keep reading

Friday digest

Intelligence, distilled weekly.

One short email every Friday, new model launches, leaderboard moves, and pricing drops. Curated by hand. Free, no spam.