Skip to main content
Ollama ExplorerBeta
LanguageadvancedTools

lfm2

Llama

LFM2 is a family of hybrid models designed for on-device deployment. LFM2-24B-A2B is the largest model in the family, scaling the architecture to 24 billion parameters while keeping inference efficient.

1.6M pullsUpdated Feb 24, 20266 tags32K context

Quick start

ollama run lfm2

Available sizes

TagSizeQuantizationContextMin RAM
lfm2:latest14GBq4_k_m32K context17.5 GB

Run with

Claude Code
ollama launch claude --model lfm2
Codex
ollama launch codex --model lfm2
OpenCode
ollama launch opencode --model lfm2
OpenClaw
ollama launch openclaw --model lfm2

Strengths & Limitations

Strengths

  • On-device deployment
  • Efficient inference
  • Large parameter size (24B)

Related models