Add support for batched generation

#18
Files changed (1) hide show
  1. README.md +1 -14
README.md CHANGED
@@ -43,17 +43,4 @@ torchrun --nproc_per_node=4 --master_port=3045 train.py \
43
  --fsdp "shard_grad_op auto_wrap" \
44
  --fsdp_transformer_layer_cls_to_wrap 'LLaMADecoderLayer' \
45
  --tf32 True --report_to="wandb"
46
- ```
47
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
48
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_chavinlo__alpaca-native)
49
-
50
- | Metric | Value |
51
- |-----------------------|---------------------------|
52
- | Avg. | 41.96 |
53
- | ARC (25-shot) | 52.3 |
54
- | HellaSwag (10-shot) | 77.09 |
55
- | MMLU (5-shot) | 41.6 |
56
- | TruthfulQA (0-shot) | 37.58 |
57
- | Winogrande (5-shot) | 69.46 |
58
- | GSM8K (5-shot) | 1.44 |
59
- | DROP (3-shot) | 14.23 |
 
43
  --fsdp "shard_grad_op auto_wrap" \
44
  --fsdp_transformer_layer_cls_to_wrap 'LLaMADecoderLayer' \
45
  --tf32 True --report_to="wandb"
46
+ ```