|
--- |
|
language: |
|
- ru |
|
- ru-RU |
|
tags: |
|
- t5 |
|
inference: |
|
parameters: |
|
no_repeat_ngram_size: 4 |
|
datasets: |
|
- samsum |
|
widget: |
|
- text: | |
|
Jeff: Can I train a 🤗 Transformers model on Amazon SageMaker? |
|
Philipp: Sure you can use the new Hugging Face Deep Learning Container. |
|
Jeff: ok. |
|
Jeff: and how can I get started? |
|
Jeff: where can I find documentation? |
|
Philipp: ok, ok you can find everything here. https://huggingface.co/blog/the-partnership-amazon-sagemaker-and-hugging-face |
|
model-index: |
|
- name: mbart_ruDialogSum |
|
results: |
|
- task: |
|
name: Abstractive Dialogue Summarization |
|
type: abstractive-text-summarization |
|
dataset: |
|
name: "SAMSum Corpus (translated to Russian)" |
|
type: samsum |
|
metrics: |
|
- name: Validation ROGUE-1 |
|
type: rogue-1 |
|
value: 30 |
|
- name: Validation ROGUE-L |
|
type: rogue-l |
|
value: 30 |
|
- name: Test ROGUE-1 |
|
type: rogue-1 |
|
value: 31 |
|
- name: Test ROGUE-L |
|
type: rogue-l |
|
value: 31 |
|
--- |
|
### 📝 Description |