Painted Fantasy v4

Magistral Small 2509 24B
image

Overview

This is an uncensored model intended to excel at creative character driven RP / ERP.

Feels like a good in between of creativity / dialogue and logic. This version tries to improve on the writing style and intelligence compared to v3.

A small portion of reasoning data was included, so thinking using the [THINK][/THINK] tags should still work. Although I haven't tested it personally, as the model is generally intended to be used without reasoning.

SillyTavern Settings

Recommended Roleplay Format

> Actions: In plaintext
> Dialogue: "In quotes"
> Thoughts: *In asterisks*

Recommended Samplers

> Temp: 0.8
> MinP: 0.05 - 0.075
> TopP: 0.95 - 1.00

Instruct

Mistral v7 Tekken

Quantizations

EXL3

> 3bpw
> 4bpw
> 5bpw
> 6bpw

Creation Process

Creation Process: SFT > DPO

SFT on approx 26 million tokens (18.3 million trainable). Datasets included SFW / NSFW RP, stories, NSFW reddit writing prompts, creative instruct & chat data.

90% of the dataset is without thinking, 10% included thinking, using the [THINK][/THINK] tags.

All RP data and synthetic stories went through rewriting with GLM 4.7 using hand edited examples as guidelines to improve the response. Rewritten responses were discarded if they failed to reduce the slop score for the message. This reduced the slop by about 25% for each RP / story dataset and made the model noticably more creative with some of its descriptions.

Additionally, some extra filtering was run over the datasets, finding about a dozen samples containing uncaught refusals, some messy human data and in general just some conversations that were outliers in low quality that had accumulated since I started building my datasets.

DPO was expanded to include non creative datasets. My usual RP DPO dataset (also rewritten) was included along with cybersecurity and two partial subsets of general assistant / chat preference datasets to help stabalize the model. This worked pretty well. While creativity did take a small hit, enough remained that the improved logic resulted in a notably improved model (IMO).

Downloads last month
28
Safetensors
Model size
24B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for zerofata/MS3.2-PaintedFantasy-v4-24B

Datasets used to train zerofata/MS3.2-PaintedFantasy-v4-24B

Collection including zerofata/MS3.2-PaintedFantasy-v4-24B