LFM2.5-1.2B-Thinking is a lightweight reasoning-focused model optimized for agentic tasks, data extraction, and RAG—while still running comfortably on edge devices. It supports long context (up to 32K tokens) and is designed to provide higher-quality “thinking” responses in a small 1.2B model.
Modalities
In / Out Price
$0 / $0per 1M
Context
33K
Weekly Rank
#241on OpenRouter