Trouper-12B

A character roleplay model trained on the custom "Actors" dataset, fine-tuned from Mistral-Nemo-Base-12B. This model was made to expand on the things I learned from TinyRP, and to overcome certain limitations I found from it; also on an entirely new dataset made just for this model.

This model writes more naturally, less like "AI"; even more so than the 24B model I'm also releasing. I suppose this is because the 12B model saw less synthethic data, and is thus less likely to use phrases typical in AI writing & prose.

-> If you're looking for the larger model in this series: Prima-24B

Looking for feedback, so please do share if you got any!

Key Features

  • Clean prose: Minimal AI slop patterns, natural speech
  • Character depth: Handles emotional progression and vulnerability well
  • Efficient: 12B size provides fast inference while maintaining quality
  • Template-dependent: Requires Mistral-V3-Tekken for proper stop behavior

Recommended Settings

Use chat completion mode

  • Temperature: 0.7 (tested and validated)
  • Template: Mistral-V3-Tekken OR ChatML, some users reported better results with ChatML (critical for proper formatting and stop behavior)
  • Context: Handles 15-20+ turn conversations effectively
  • Prompt Preprocessing: Semi-strict, no tools

Strengths

  • Writing Quality: Direct, concrete descriptions without purple prose
  • Natural Dialogue: Speech patterns feel authentic, not performative
  • Emotional Range: Handles vulnerability, humor, and character growth
  • Structural Variety: Avoids formulaic response patterns
  • Show Don't Tell: Trusts the reader, doesn't over-explain emotions

Comparison to Prima-24B

Trouper-12B and Prima-24B are trained on identical data but offer different trade-offs:

Aspect Trouper-12B Prima-24B
Prose Style Direct and concrete Slightly more elaborate
AI Slop Minimal Moderate (some patterns)
Reliability Good (template-sensitive) Excellent
Long Context Good (12B) Better (24B)
Inference Speed Faster (12B) Slower (24B)
Setup Difficulty Moderate (template critical) Easy
Action RP Good Excellent
Emotional RP Excellent Good

Choose Trouper-12B if: You want best prose quality, natural dialogue, and don't mind template setup
Choose Prima-24B if: You want reliability, long context, or action-oriented RP

Comparison to TinyRP-12B

This model addresses several issues found in my previous TinyRP-12B release:

Aspect TinyRP-12B Trouper-12B
Formulaic patterns Yes (after 20+ turns) No
Character stagnation Yes No - characters evolve
Opening variety Repetitive Varied
Training data Original dataset Custom "Actors" dataset
Long conversations Degrades Maintains quality

Known Limitations

  • Template Sensitivity: Without Mistral-V3-Tekken, may generate meta-narration or continue past appropriate stopping points
  • Occasional Meta-Breaks: Rare instances of stepping outside character (regenerate if needed)
  • Context Window: While good for 15-20+ turns, may be outperformed by larger models at 50+ turns. please let me know how it works for you!

Got Feedback?

Issues, questions, or feedback welcome! Particularly interested in:

  • Long conversation quality (20+ turns)
  • Template compatibility findings
  • Comparison with other RP models

Feel free to make a post in the Community tab here!

Downloads last month
16
Safetensors
Model size
12B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DarwinAnim8or/Trouper-12B

Finetuned
(80)
this model
Quantizations
7 models