Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
EmbeddedLLM
/
vLLM-AMD-flash-attn-debug
like
0
Follow
EmbeddedLLM
25
Model card
Files
Files and versions
Community
main
vLLM-AMD-flash-attn-debug
1 contributor
History:
3 commits
tjellm
Upload flash-attn-api-error.zip with huggingface_hub
2b4be78
verified
3 months ago
.gitattributes
Safe
1.52 kB
initial commit
3 months ago
flash-attn-api-error.zip
pickle
Detected Pickle imports (14)
"torch.storage._load_from_bytes"
,
"vllm.sampling_params.SamplingParams"
,
"torch._utils._rebuild_tensor_v2"
,
"array.array"
,
"collections.OrderedDict"
,
"vllm.worker.model_runner.ModelInputForGPUWithSamplingMetadata"
,
"vllm.sequence.SequenceStage"
,
"vllm.sequence.SequenceData"
,
"vllm.sampling_params.RequestOutputKind"
,
"vllm.model_executor.sampling_metadata.SequenceGroupToSample"
,
"vllm.attention.backends.rocm_flash_attn.ROCmFlashAttentionMetadata"
,
"vllm.model_executor.sampling_metadata.SamplingMetadata"
,
"array._array_reconstructor"
,
"vllm.sampling_params.SamplingType"
How to fix it?
13.6 kB
LFS
Upload flash-attn-api-error.zip with huggingface_hub
3 months ago