YAML Metadata Warning: empty or missing yaml metadata in repo card
Check out the documentation for more information.
Good story telling models that can fit in an RTX 3060 12GB. Updated March 2026.
Models
Current favorite: Qwen 3.5 with abliteration:
At this point a roleplay finetune does more harm than good. The 27b model is very knowledgable, and the abliteration methods work perfectly.
zerofata/MS3.2-PaintedFantasy-v2-24B - unfortunately v3/v4/visage are cooked
Contender: LatitudeGames/Hearthfire-24B
Creators
- Whitelist: Sao, nbeerbower, zerofata
- Blacklist: DavidAU, SicariusSicariiStuff, Allura
- Greylst: The Drummer
Remarks
- Roleplay and storywriting are distinct tasks! Some models excel at one and fail at the other.
- Dont waste time on sampler settings; use recommended and optimize the prompt
- Don't "overparameterize" by writing too long a prompt
- Don't underestimate the original instruct models
Links
- llama.cpp and llama-cpp-python - preferred LLM software
- /r/localllama
- /lmg/
- LMSys Chatbot Arena Leaderboard
- Uncensored General Intelligence Leaderboard
- /r/SillyTavernAI
- NothingiisReal discord
- NeverSleep discord
- SillyTavern discord
- BeaverAI discord
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support