AI & ML interests

NoesisLab advances machine learning research in deep contemplation and reflective reasoning to enable more profound and self-aware artificial intelligence.

Recent Activity

OzTianlu  updated a model 41 minutes ago
NoesisLab/Kai-30B-Instruct
OzTianlu  published a model about 2 hours ago
NoesisLab/Kai-30B-Instruct
OzTianlu  updated a collection about 22 hours ago
Kai Models Series
View all activity

OzTianlu 
posted an update about 16 hours ago
view post
Post
117
🚨 URGENT: To the 13k+ users downloading Kai-3B-Instruct — Please update to v1.1! (Official Q8_0 GGUF inside)
OzTianlu/Kai-3B-Instruct-Q8_0-GGUF
Wow. Waking up to see over 13,000 combined downloads for the Kai-3B-Instruct GGUFs is absolutely mind-blowing. Thank you so much to the community and to the awesome creators ( @SimplySara & @mradermacher ) for the auto-quantization!
However, we have a slight "suffering from success" situation here. 😅
⚠️ THE ISSUE: You are likely running the v1.0 "Logic-Poisoned" weights.
If your model is acting like a cold, emotionless robot that only replies with a rigid Analysis -> Approach -> Solution template even when you just say "Hello", you have v1.0. In our initial release, the model overfitted to its reasoning corpus, causing a complete "conversational mode collapse."
🚀 THE FIX: Official v1.1 is Live!
We have completed a 4000-step annealing phase to restore its sanity.
  • 2 replies
·
OzTianlu 
posted an update 2 days ago
view post
Post
1659
Scaling UP in Kai! 🌊
NoesisLab/Kai-3B-Instruct

Introducing NoesisLab/Kai-3B-Instruct What happens when you force a 3B model to reason entirely in its latent space ?
Meet Kai-3B, our latest industrial-grade reasoning model fine-tuned using the Adaptive Dual Search (ADS) algorithm.
GSM8K (0-shot, Direct Answer): 39.27% 🤯 (Llama-2-7B is ~14.6%)
HumanEval (Pass@1): 39.02% 💻 (Overtakes Gemma-2-2B's 30%)
MMLU (5-shot): 53.62% 📚 (Crushing the 50% barrier)
ARC-Challenge: 51.88%🎯
PIQA: 77.53%
HellaSwag: 69.53%
Kai-3B proves that reasoning density doesn't strictly require parameter bloat or verbose generation. It acts as a perfect, cold-blooded Agent action-engine—ideal for JSON routing, SWE-bench patch generation, and anywhere you need absolute structured certainty without token waste.
  • 2 replies
·