LFM2 represents an advanced series of on-device foundation models designed to provide a remarkably swift generative-AI experience across a diverse array of devices. By utilizing a novel hybrid architecture, it achieves decoding and pre-filling speeds that are up to twice as fast as those of similar models, while also enhancing training efficiency by as much as three times compared to its predecessor. These models offer a perfect equilibrium of quality, latency, and memory utilization suitable for embedded system deployment, facilitating real-time, on-device AI functionality in smartphones, laptops, vehicles, wearables, and various other platforms, which results in millisecond inference, device durability, and complete data sovereignty. LFM2 is offered in three configurations featuring 0.35 billion, 0.7 billion, and 1.2 billion parameters, showcasing benchmark results that surpass similarly scaled models in areas including knowledge recall, mathematics, multilingual instruction adherence, and conversational dialogue assessments. With these capabilities, LFM2 not only enhances user experience but also sets a new standard for on-device AI performance.