Sparse Auto-Encoders (SAEs) for Mechanistic Interpretability Collection A compilation of sparse auto-encoders trained on large language models. • 37 items • Updated Dec 16, 2025 • 25
Rope to Nope and Back Again: A New Hybrid Attention Strategy Paper • 2501.18795 • Published Jan 30, 2025 • 13
MHLA: Restoring Expressivity of Linear Attention via Token-Level Multi-Head Paper • 2601.07832 • Published Jan 12 • 52
view article Article MedEmbed: Fine-Tuned Embedding Models for Medical / Clinical IR Oct 20, 2024 • 53
view article Article Accelerate ND-Parallel: A guide to Efficient Multi-GPU Training +3 Aug 8, 2025 • 93
ThisUsernameAlreadyExistsAlreadyExists/gemma-3n-billberes Image-Text-to-Text • 6B • Updated Sep 5, 2025
ThisUsernameAlreadyExistsAlreadyExists/gemma-3n-billberes Image-Text-to-Text • 6B • Updated Sep 5, 2025