OsirisCortex-v7-Censo-MLX

Cortex v7 (Censored + Vision) — Osiris's main reasoning + vision brain. VLM (Vision-Language Model) that can process images, screenshots, and video frames. Runs natively on Apple Silicon via MLX Metal.

Architecture

  • Base Model: Qwen3.5-9B VLM (9B params, vision + language)
  • Architecture: Qwen3_5ForConditionalGeneration (multimodal)
  • Format: MLX mxfp4 quantized (Apple Silicon native)
  • Size: ~5.3 GB (includes vision encoder)
  • Vision: Full image understanding, OCR, screenshot analysis, video frames
  • Note: This is the censored (base) version. For uncensored, see OsirisCortex-v7-MLX.

Usage

from mlx_vlm import load, generate

model, processor = load("osirisbrain/OsirisCortex-v7-Censo-MLX")
output = generate(model, processor, "Describe this image", ["path/to/image.jpg"])

Credits

Converted by RepublicOfKorokke. Original model: Qwen/Qwen3.5-9B by Alibaba.

Downloads last month
12
Safetensors
Model size
2B params
Tensor type
U8
·
U32
·
BF16
·
F32
·
MLX
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for osirisbrain/OsirisCortex-v7-Censo-MLX

Finetuned
Qwen/Qwen3.5-9B
Quantized
(101)
this model