openbmb/MiniCPM-o-4_5-awq
Any-to-Any
•
Updated
•
40
•
4
Large Language Models
Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts
InfLLM-V2: Dense-Sparse Switchable Attention for Seamless Short-to-Long Adaptation