view post Post 155 Mistral's new SOTA coding models Devstral 2 can now be Run locally! (25GB RAM) 🐱We fixed the chat template, so performance should be much better now!24B: unsloth/Devstral-Small-2-24B-Instruct-2512-GGUF123B: unsloth/Devstral-2-123B-Instruct-2512-GGUF🧡Step-by-step Guide: https://docs.unsloth.ai/models/devstral-2 See translation 🔥 1 1 ❤️ 1 1 🚀 1 1 🤗 1 1 + Reply
view post Post 3372 Mistral's new Ministral 3 models can now be Run & Fine-tuned locally! (16GB RAM)Ministral 3 have vision support and the best-in-class performance for their sizes.14B Instruct GGUF: unsloth/Ministral-3-14B-Instruct-2512-GGUF14B Reasoning GGUF: unsloth/Ministral-3-14B-Reasoning-2512-GGUF🐱 Step-by-step Guide: https://docs.unsloth.ai/new/ministral-3All GGUFs, BnB, FP8 etc. variants uploads: https://huggingface.co/collections/unsloth/ministral-3 See translation 3 replies · 🔥 17 17 🤗 7 7 ❤️ 5 5 🚀 3 3 + Reply
view post Post 8294 Qwen3-Next can now be Run locally! (30GB RAM)Instruct GGUF: unsloth/Qwen3-Next-80B-A3B-Instruct-GGUFThe models come in Thinking and Instruct versions and utilize a new architecture, allowing it to have ~10x faster inference than Qwen32B.💜 Step-by-step Guide: https://docs.unsloth.ai/models/qwen3-nextThinking GGUF: unsloth/Qwen3-Next-80B-A3B-Thinking-GGUF See translation 🔥 37 37 ❤️ 11 11 🚀 7 7 🤗 3 3 + Reply
view post Post 4236 You can now run Kimi K2 Thinking locally with our Dynamic 1-bit GGUFs: unsloth/Kimi-K2-Thinking-GGUFWe shrank the 1T model to 245GB (-62%) & retained ~85% of accuracy on Aider Polyglot. Run on >247GB RAM for fast inference.We also collaborated with the Moonshot AI Kimi team on a system prompt fix! 🥰Guide + fix details: https://docs.unsloth.ai/models/kimi-k2-thinking-how-to-run-locally See translation ❤️ 10 10 🚀 9 9 🔥 6 6 🤗 4 4 🤯 3 3 + Reply
Translation Errors Significantly Impact Low-Resource Languages in Cross-Lingual Learning Paper • 2402.02080 • Published Feb 3, 2024 • 2