DJ-AI-Radio-MLX / README.md
felixmanojh's picture
Improve model card with professional structure (inspired by Chess Tutor)
b415370 verified
metadata
language:
  - en
  - ta
license: apache-2.0
tags:
  - text-generation
  - mlx
  - qwen2.5
  - tanglish
  - radio-dj
  - code-mixing
  - tamil
library_name: mlx-lm
pipeline_tag: text-generation
base_model: Qwen/Qwen2.5-0.5B-Instruct

๐ŸŽต Tamil AI DJ Radio - MLX Optimized

Blazing-fast Qwen 2.5-0.5B for Apple Silicon (M1/M2/M3/M4)

MLX-optimized model for generating energetic Tanglish (Tamil-English) radio DJ commentary. Fused LoRA weights for maximum performance on Apple Silicon.


๐ŸŽฏ Model Overview

  • Base Model: Qwen/Qwen2.5-0.5B-Instruct (4-bit quantized)
  • Model Type: MLX Fused (LoRA weights merged)
  • Training Data: 5,027 Tanglish DJ commentary examples
  • Best Checkpoint: Iteration 2900 (validation loss: 1.856)
  • Model Size: 276MB
  • Framework: MLX (Apple Silicon optimized)

โšก Performance

Speed (M1 Mac)

  • Loading: ~2 seconds
  • Inference: ~3 seconds for 150 tokens
  • Memory: <2GB RAM usage
  • Latency: ~20ms per token

Why MLX?

  • ๐Ÿš€ 3-5x faster than Transformers on Mac
  • ๐Ÿ’พ Lower memory usage with unified memory
  • ๐Ÿ”‹ Better power efficiency on Apple Silicon
  • ๐ŸŽฏ Native Metal acceleration

๐Ÿš€ Quick Start

Installation

pip install mlx mlx-lm

Simple Usage

from mlx_lm import load, generate

# Load MLX-optimized model
model, tokenizer = load("felixmanojh/DJ-AI-Radio-MLX")

# Generate DJ commentary
messages = [
    {"role": "system", "content": "You are a Tamil AI radio DJ who speaks energetic Tanglish."},
    {"role": "user", "content": "Hype up a party track"}
]

prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
response = generate(model, tokenizer, prompt=prompt, max_tokens=150, verbose=False)
print(response)

Example Output:

Party mode activate! Friday night ah Saturday night mode activate!
Club vibes high-energy vibes! Dance floor crowded! Everyone jumping!
Party starter! Energy maximum! Music energizing!

๐Ÿ’ก Why Use MLX Version?

โœ… Advantages

Feature MLX Transformers
Speed (M1 Mac) ~3s ~10s
Memory <2GB ~4GB
Loading 2s 10s
Platform macOS only Cross-platform
Power Usage Low Higher

Use This Version For:

  • ๐ŸŽ Mac development (M1/M2/M3/M4)
  • ๐Ÿš€ Local inference with maximum speed
  • ๐Ÿ’ป Demos and prototypes on Apple Silicon
  • ๐Ÿ”ฌ Research with fast iteration cycles
  • ๐Ÿ“ฑ macOS apps with native performance

๐Ÿ“Š Training Details

Parameter Value
LoRA Rank 8 (fused into base)
LoRA Alpha 16
Training Iterations 6,000 (best @ 2900)
Validation Loss 1.856
Training Data 5,027 examples, 68 themes
Framework MLX (Apple)

๐Ÿ“š Supported Vibes

The model generates commentary for various moods:

High Energy:

  • ๐ŸŽ‰ Party anthems
  • ๐Ÿ’ช Workout motivation
  • ๐ŸŽฎ Gaming streams

Chill Vibes:

  • ๐ŸŒŠ Beach relaxation
  • ๐Ÿ“š Study focus
  • ๐ŸŒง๏ธ Rain moods

Themed:

  • ๐ŸŽฌ Movie/Cinema vibes
  • ๐Ÿœ Street food sessions
  • ๐Ÿš— Road trips
  • ๐ŸŽ† Festival celebrations

๐ŸŽ“ Intended Use

โœ… Recommended

  • Local development on Mac
  • Fast prototyping and demos
  • macOS applications
  • Real-time commentary generation
  • Educational demonstrations

โŒ Not Recommended

  • Linux/Windows deployment (use Merged model)
  • Production servers (use Transformers version)
  • Non-Mac platforms

๐Ÿ”„ Model Variants

๐Ÿ“Œ Choose Your Format:

Model Format Size Platform Speed
DJ-AI-Radio-LoRA LoRA adapter 17MB Any Medium
DJ-AI-Radio Merged (HF) 276MB Any Medium
DJ-AI-Radio-MLX (this) Fused (MLX) 276MB Mac only Fast

๐ŸŒ Live Demo

Try the complete AI radio station:

๐Ÿ”— Tamil AI DJ Radio Space

Features:

  • LLM-generated commentary (this model's merged version)
  • Voice cloning with Coqui XTTS
  • AI-generated music playback

๐Ÿ“ Citation

@software{tamil_ai_dj_radio_mlx_2025,
  author = {Felix Manojh},
  title = {Tamil AI DJ Radio - MLX Optimized},
  year = {2025},
  publisher = {HuggingFace},
  url = {https://huggingface.co/felixmanojh/DJ-AI-Radio-MLX},
  note = {MLX-optimized Qwen 2.5-0.5B for Tanglish DJ commentary}
}

๐Ÿ“„ License

Apache 2.0 (inherits from Qwen 2.5)


๐Ÿ™ Acknowledgments

  • Base Model: Qwen Team (Alibaba Cloud)
  • MLX Framework: Apple ML Research
  • Training Data: Claude API (Anthropic)
  • Inspiration: Tamil music culture

Built with โค๏ธ for the Tamil-speaking community