view article Article We’re open-sourcing our text-to-image model and the process behind it 25 days ago • 73
KTO: Model Alignment as Prospect Theoretic Optimization Paper • 2402.01306 • Published Feb 2, 2024 • 20
💧LFM2-8B-A1B-MoE Collection Best in Class MoE, better than Qwen3. Optimised for Smaller devices sub 16 GB (M1/2/3/4) Apple Silicon. • 7 items • Updated Oct 9 • 4
ServiceNow-Apriel Collection Apriel-1.5-15b-Thinker is a multimodal reasoning model in ServiceNow’s Apriel SLM series which achieves competitive performance against models 10 time • 6 items • Updated Oct 5 • 1
MedGemma Collection Collection of Gemma 3 variants for performance on medical text and image comprehension to accelerate building healthcare-based AI applications. • 14 items • Updated Jul 10 • 5
Qwen3-Coder-MoE Collection 💻 Significant Performance: among open models on Agentic Coding, Agentic Browser-Use, and other foundational coding tasks, achieving ~Claude Sonnet. • 6 items • Updated Oct 4 • 1
Open LLM Leaderboard best models ❤️🔥 Collection A daily uploaded list of models with best evaluations on the LLM leaderboard: • 65 items • Updated Mar 20 • 653
MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases Paper • 2402.14905 • Published Feb 22, 2024 • 134