Rerun Request for Failed Benchmark

#458
by Ba2han - opened

Hi!

https://huggingface.co/datasets/open-llm-leaderboard/requests/blob/main/Ba2han/BruinsV2-OpHermesNeu-11B_eval_request_False_bfloat16_Original.json

Could you re-submit my merge please? I think it failed during maintenance and I can't do it myself, thank you.

Open LLM Leaderboard org

Hi!
Your model failed because it could not be loaded, did you follow all the steps in the About ?

  File "...python3.10/site-packages/transformers/models/auto/auto_factory.py", line 566, in from_pretrained
    return model_class.from_pretrained(
  File "...python3.10/site-packages/transformers/modeling_utils.py", line 3761, in from_pretrained
    dispatch_model(model, **device_map_kwargs)
  File "...python3.10/site-packages/accelerate/big_modeling.py", line 421, in dispatch_model
    model.to(device)
  File "...python3.10/site-packages/transformers/modeling_utils.py", line 2460, in to
    return super().to(*args, **kwargs)
  File "...python3.10/site-packages/torch/nn/modules/module.py", line 1145, in to
    return self._apply(convert)
  File "...python3.10/site-packages/torch/nn/modules/module.py", line 797, in _apply
    module._apply(fn)
  File "...python3.10/site-packages/torch/nn/modules/module.py", line 797, in _apply
    module._apply(fn)
  File "...python3.10/site-packages/torch/nn/modules/module.py", line 797, in _apply
    module._apply(fn)
  [Previous line repeated 2 more times]
  File "...python3.10/site-packages/torch/nn/modules/module.py", line 820, in _apply
    param_applied = fn(param)
  File "...python3.10/site-packages/torch/nn/modules/module.py", line 1143, in convert
    return t.to(device, dtype if t.is_floating_point() or t.is_complex() else None, non_blocking)
NotImplementedError: Cannot copy out of meta tensor; no data!

Hello again, thank you for sharing this. I am not sure why this happened but I assume my super bad internet may have something to do with it. Solved!

Edit: My other merge got a score, and I didn't do anything different. I am barely scratching the surface of LLMs and Python, so excuse my lack of knowledge if I failed to follow the steps!

Ba2han changed discussion status to closed
Open LLM Leaderboard org
edited Dec 13, 2023

It's probably not a matter of your internet connection since we load models from the hub.

Regarding the steps, we suggest to do a number of things before submitting to ensure model evaluation goes smoothly (converting the model with safetensors, making sure it can be loaded with from_pretrained) - once you followed them for your model, I can relaunch its evaluation if you want.

Thank you for your comment!

It's probably not a matter of your internet connection since we load models from the hub.

Regarding the steps, we suggest to do a number of things before submitting to ensure model evaluation goes smoothly (converting the model with safetensors, making sure it can be loaded with from_pretrained) - once you followed them for your model, I can relaunch its evaluation if you want.

Thank you for your comment!

True, what I failed to mention was that I probably disconnected while uploading a shard due to bad connection. I checked the SHA of safetensors. The last one was wrong:

resim.png

I now fixed it and loaded the model successfully with the provided code:

loading.png

I would appreciate if you could try processing again!

Open LLM Leaderboard org

I see, that makes sense!
I relaunched it :)

Sign up or log in to comment