We're launching a series of small but mighty language models: 🏎️ Super fast - runs on laptops, phones, you name it! 📏 3 sizes: 130M, 350M, and 1.5B parameters 🥇 Outperforms same size models from Meta, Microsoft, and Qwen 🔓 Fully open-source: datasets, training code, models
𝐊𝐞𝐲 𝐟𝐞𝐚𝐭𝐮𝐫𝐞𝐬 - Trained on FineWeb-Edu and Cosmopedia v2 (largest synthetic pre-training dataset) - No cloud needed - run locally for privacy and energy efficiency - Everything is public, from data curation to training steps
𝐏𝐨𝐭𝐞𝐧𝐭𝐢𝐚𝐥 𝐮𝐬𝐞 𝐜𝐚𝐬𝐞𝐬 - On-device autocomplete - Local request parsing - Custom fine-tuning for specific needs without the need for expensive GPUs