Edit model card

TeeZee/NEBULA-XB-v1.03

Experiment, can DUS be taken one or more steps further?

Technical notes:

  • pretrained model v03 finetuned on 50k entries from SlimOrca dataset
  • 18 layers removed from both models of finetuned GALAXY-XB-v03
  • model has 108 layers (((48-12)*2)-18)*2 = 108
  • second step in scaling DUS procedure

To evaluate

  • model performance after merge, should be a little lover that GALAXY finetuned on 50k of slimorca

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 53.52
AI2 Reasoning Challenge (25-Shot) 56.66
HellaSwag (10-Shot) 81.78
MMLU (5-Shot) 60.98
TruthfulQA (0-shot) 44.03
Winogrande (5-shot) 77.66
GSM8k (5-shot) 0.00
Downloads last month
748
Safetensors
Model size
23.8B params
Tensor type
BF16
·
Inference API
Input a message to start chatting with TeeZee/NEBULA-XB-v1.0.
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Dataset used to train TeeZee/NEBULA-XB-v1.0

Evaluation results