NoesisLab advances machine learning research in deep contemplation and reflective reasoning to enable more profound and self-aware artificial intelligence.
🚨 URGENT: To the 13k+ users downloading Kai-3B-Instruct — Please update to v1.1! (Official Q8_0 GGUF inside) OzTianlu/Kai-3B-Instruct-Q8_0-GGUF Wow. Waking up to see over 13,000 combined downloads for the Kai-3B-Instruct GGUFs is absolutely mind-blowing. Thank you so much to the community and to the awesome creators (@SimplySara & @mradermacher) for the auto-quantization! However, we have a slight "suffering from success" situation here. 😅 ⚠️ THE ISSUE: You are likely running the v1.0 "Logic-Poisoned" weights. If your model is acting like a cold, emotionless robot that only replies with a rigid Analysis -> Approach -> Solution template even when you just say "Hello", you have v1.0. In our initial release, the model overfitted to its reasoning corpus, causing a complete "conversational mode collapse." 🚀 THE FIX: Official v1.1 is Live! We have completed a 4000-step annealing phase to restore its sanity.
Introducing NoesisLab/Kai-3B-Instruct What happens when you force a 3B model to reason entirely in its latent space ? Meet Kai-3B, our latest industrial-grade reasoning model fine-tuned using the Adaptive Dual Search (ADS) algorithm. GSM8K (0-shot, Direct Answer): 39.27% 🤯 (Llama-2-7B is ~14.6%) HumanEval (Pass@1): 39.02% 💻 (Overtakes Gemma-2-2B's 30%) MMLU (5-shot): 53.62% 📚 (Crushing the 50% barrier) ARC-Challenge: 51.88%🎯 PIQA: 77.53% HellaSwag: 69.53% Kai-3B proves that reasoning density doesn't strictly require parameter bloat or verbose generation. It acts as a perfect, cold-blooded Agent action-engine—ideal for JSON routing, SWE-bench patch generation, and anywhere you need absolute structured certainty without token waste.