Text Generation
Transformers
Safetensors
llama
text-generation-inference
Inference Endpoints
leaderboard-pr-bot's picture
Adding Evaluation Results
c8239fd verified
|
raw
history blame
7.94 kB
metadata
license: other
datasets:
  - adamo1139/rawrr_v2-2_stage1
  - adamo1139/HESOYAM_v0.2
license_name: yi-license
license_link: LICENSE
model-index:
  - name: Yi-34B-200K-HESOYAM-0905
    results:
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: AI2 Reasoning Challenge (25-Shot)
          type: ai2_arc
          config: ARC-Challenge
          split: test
          args:
            num_few_shot: 25
        metrics:
          - type: acc_norm
            value: 65.61
            name: normalized accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=adamo1139/Yi-34B-200K-HESOYAM-0905
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: HellaSwag (10-Shot)
          type: hellaswag
          split: validation
          args:
            num_few_shot: 10
        metrics:
          - type: acc_norm
            value: 83.39
            name: normalized accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=adamo1139/Yi-34B-200K-HESOYAM-0905
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: MMLU (5-Shot)
          type: cais/mmlu
          config: all
          split: test
          args:
            num_few_shot: 5
        metrics:
          - type: acc
            value: 75.3
            name: accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=adamo1139/Yi-34B-200K-HESOYAM-0905
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: TruthfulQA (0-shot)
          type: truthful_qa
          config: multiple_choice
          split: validation
          args:
            num_few_shot: 0
        metrics:
          - type: mc2
            value: 43.35
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=adamo1139/Yi-34B-200K-HESOYAM-0905
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: Winogrande (5-shot)
          type: winogrande
          config: winogrande_xl
          split: validation
          args:
            num_few_shot: 5
        metrics:
          - type: acc
            value: 82.95
            name: accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=adamo1139/Yi-34B-200K-HESOYAM-0905
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: GSM8k (5-shot)
          type: gsm8k
          config: main
          split: test
          args:
            num_few_shot: 5
        metrics:
          - type: acc
            value: 53.68
            name: accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=adamo1139/Yi-34B-200K-HESOYAM-0905
          name: Open LLM Leaderboard

Known Issues

There's something weird going on with tokenizer. EXL2 quant works fine in ooba but not in exui. BNB 4-bit quant works fine in ooba. For best results, use ooba with BOS token being inserted, repp 1.05 and probably exllamav2_HF loader over exllamav2

Model Description

Have you ever wanted a sandbox for text-based social media? A place where you can bully a person, throw arguments or attack someone without any kind of actual harm being done and without any repercussions? All of it fully local, so nobody but you will ever know? No? Well, HESOYAM kinda can do that, but it's not exactly a bully similator, that's just one of ways you could use it. Specify a place on the internet that you want to be in the system prompt and then start a discussion. Will it be engaging or will you be sucked into someone's depression? For now, probably the latter. Still, I had some insightful concrete useful discussions with this model, it's not all gptslopped fluff. It does have a lot of depressive negative tones though, so it might not be for everyone.

To get this model, first, I fine-tuned Yi-34B-200K (xlctx, as in second version of 34B 200K model, not new 1.5) on adamo1139/rawrr_v2-2_stage1 to make it so that base model will forget it's AI assistant programming and behave like a completion model trained on raw corpus of internet. This was done using ORPO and GaLore - all of it handled by Unsloth. I would say it's a moderately successful finetune, I plan to enhance rawrr dataset with richer data to make better finetunes of this kind in the future. Resulting adapter file can be found here and FP16 model file for RAWrr ORPO finetune can be found here.

Once I had good base model, I fine-tuned it on HESOYAM 0.2 dataset. It's a collection of single turn conversations from around 10 subreddits and multi-turn conversations from board /x/. There's also pippa in there. All samples there have system prompts that should tell the model about where discussion is taking place, this will be useful when you will be deciding on where you want to have your sandbox discussion take place. Here, I used classic SFT with GaLore and Unsloth, I wanted to get some results quick so it's trained for just 0.4 epochs. Adapter after that part of fine-tuning can be found here.

Conversation samples - I put in a seed prompt and let the model generate the rest of the conversation.

Results on my base benchmarks - Responses suggests it still has some general assistant capabilities. I don't really want that, maybe I should up the learning rate for next run so that it stays in character more.

Prompt template

It's chatml, like always.

<|im_start|>system
A chat on subreddit /r/pcmasterrace.<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant

Quants

I haven't done them yet. I will maybe upload one EXL2 quant.

Intended uses & limitations

Use is limited by Yi license.
Some datasets that were used prohibit commercial use (no_robots with CC-BY-NC-4.0), so I think you should use non-commercially only, unless you know law better and think it doesn't matter.

Credits

Thanks to unsloth and huggingface team for providing software packages used during fine-tuning.
Thanks to authors of ORPO and GaLore for their innovative fine-tuning strategies.
Thanks to random people who post datasets on hf, you rock!

made with Unsloth

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 67.38
AI2 Reasoning Challenge (25-Shot) 65.61
HellaSwag (10-Shot) 83.39
MMLU (5-Shot) 75.30
TruthfulQA (0-shot) 43.35
Winogrande (5-shot) 82.95
GSM8k (5-shot) 53.68