Edit model card

EXL2 quants of crestf411/L3-70B-daybreak-storywriter-v0.4

3.00 bits per weight
3.50 bits per weight
4.00 bits per weight
4.50 bits per weight
5.00 bits per weight
6.00 bits per weight
8.00 bits per weight

Created using the defaults from exllamav2 0.0.21 convert.py
3.0bpw to 6.0bpw head bits = 6
8.0bpw head bits = 8
length = 8192
dataset rows = 200
measurement rows = 32
measurement length = 8192

L3-70B-daybreak-storywriter-v0.4

Daybreak (2024 May 24) v0.4 LoRA on top of https://huggingface.co/tdrussell/Llama-3-70B-Instruct-Storywriter

Dataset curation to remove slop-perceived expressions continues.

The below regexes return 0 matches. Bold entries are new since v0.3.

  • 'barely above a whisper',
  • 'barely audible',
  • 'shiver([s]?) down',
  • ' ministration',
  • 'audible (["'"]?)p[l]?op',
  • 'can't help but',
  • 'buck([s]?) my ',
  • 'buck([s]?) h[ei][rs] ',
  • '[Dd]espite h[ie][mr]self',
  • 'slick slit',
  • 'whatever it takes',
  • 'unlike anything (s?)he',
  • 'a mix([a-z]*) of',
  • 'wave after wave',
  • 'reckless abandon',
  • '[Mm]aybe, just maybe',
  • 'eyes gleaming',
  • 'mischievously',
  • "couldn't help but",

From testing so far, it feels like temperature 0.8-0.9 is a good starting point. I have mostly tested with everything neutralized. Please give feedback on which parameters work good for you.

Downloads last month
0
Inference API
Input a message to start chatting with kim512/L3-70B-daybreak-storywriter-v0.4-3.0bpw-h6-exl2.
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.