Configuration Parsing
Warning:
In config.json: "quantization_config.bits" must be an integer
Writer-Large-2411-v2.1
EXL2-Quant of gghfez/Writer-Large-2411-v2.1
Creative-Writing Control-Vectors available here: gghfez/Writer-Large-2411-v2.1-control-vectors
Overview
This model is built on Mistral-Large-Instruct-2411 and optimized for creative writing purposes. The base model excels at following instructions and handling details in long context when using the new prompt template.
Key Improvements
- Reduced positivity bias
- Reduced AI tropes and repetitive language patterns in story generation
- Enhanced performance with longer context stories (multiple chapters) and roleplay sessions
- Improved steering capabilities for roleplay via [OOC] instructions
- Better handling of "group chat" scenarios
Usage
Prompt Template
The model requires a system prompt in the Mistral-V7 format.
If you omit [SYSTEM_PROMPT] [/SYSTEM_PROMPT]
, the model:
- May not follow instructions properly at short contexts
- Can become repetitive at longer contexts
Example:
[SYSTEM_PROMPT]You are an award winning writer. Assist the user.[/SYSTEM_PROMPT][INST] Write the opening chapter of ... [/INST]
SillyTavern Integration
Story String:
[SYSTEM_PROMPT] {{#if system}}{{system}}[/SYSTEM_PROMPT] [INST]
{{/if}}{{#if wiBefore}}{{wiBefore}}
{{/if}}{{#if description}}{{description}}
{{/if}}{{#if personality}}{{personality}}
{{/if}}{{#if scenario}}{{scenario}}
{{/if}}{{#if wiAfter}}{{wiAfter}}
{{/if}}{{#if persona}}{{persona}}
{{/if}}{{trim}}[/INST] Understood.</s>
For response steering, use [OOC]
commands, e.g.:
[OOC] Have them interrupted by a loud explosion in a nearby factory
[OOC] Have her refuse to sell it and suggest another merchant instead
Technical Details
Training
- QLoRA training at 32768 context
- Merged with gghfez/Mistral-Large-Instruct-2411 at bf16
- jukofyork/Creative writing control vectors were applied during synthetic dataset generation
- Includes standard assistant instruct data for long-context stability
- Note: Performance on code tasks may be reduced compared to base model
- Note: No attempt was made to remove 'Name-Slop', so you'll still encounter Lily and Elara if you don't specify character names
Context Length
- Base model: 131,072 tokens
- Training range: 1024-32728 tokens
- Training context window: 32768 tokens
Testing Environments
Tested with exllamav2 4.5bpw on:
- Downloads last month
- 48
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for gghfez/Writer-Large-2411-v2.1-exl2-4.5bpw
Base model
mistralai/Mistral-Large-Instruct-2411
Finetuned
gghfez/Writer-Large-2411-v2.1