Add evaluation results on the default config and test split of phpthinh/data_1
Beep boop, I am a bot from Hugging Face's automatic model evaluator 👋!
Your model has been evaluated on the default config and test split of the phpthinh/data_1 dataset by
@phpthinh
, using the predictions stored here.
Accept this pull request to see the results displayed on the Hub leaderboard.
Evaluate your model on more datasets here.
Hey again @phpthinh ! Would you mind closing the PRs whenever you run a private evaluation against the models? It would help tremendously!
I wish I could close the PR and run a private evaluation!
It is impossible for me to close the PR since the PR is created by the tool.
I will contact the team about this issue.
Sorry for the annoyance!
No worries, not your fault. If you could do it, that would be great, if you can't, we'll open a feature request so that you can evaluate without having to open a PR I guess.
Hi @phpthinh , excited to see that you're using Evaluation on the Hub for the BLOOM models! We should definitely add a feature to allow people to evaluate models without opening a PR.
In the meantime, a temporary suggestion for mitigating this may be for you to clone the BLOOM model repository to your namespace on the Hub instead — then you can run evaluation jobs against it as much as you'd like without pinging the authors!
That would be great!
Thanks for your suggestion
@mathemakitten
!