view post Post 214 You can now run Kimi K2 Thinking locally with our Dynamic 1-bit GGUFs: unsloth/Kimi-K2-Thinking-GGUFWe shrank the 1T model to 245GB (-62%) & retained ~85% of accuracy on Aider Polyglot. Run on >247GB RAM for fast inference.We also collaborated with the Moonshot AI Kimi team on a system prompt fix! 🥰Guide + fix details: https://docs.unsloth.ai/models/kimi-k2-thinking-how-to-run-locally See translation ❤️ 3 3 🔥 2 2 🚀 2 2 🤗 2 2 🤯 1 1 + Reply
Granite Quantized Models Collection Quantized versions of IBM Granite models. Licensed under the Apache 2.0 license. • 42 items • Updated about 24 hours ago • 24