Text Generation
Transformers
PyTorch
Safetensors
English
mixtral
conversational
Inference Endpoints
text-generation-inference

llama.ts: failed to load model. Error: create_tensor: tensor 'blk.0.ffn_gate.weight' not found

#21
by AlexanderWillamowski - opened

Hi community,

constantly I am facing the following issue. I downloaded version dolphin-2.5-mixtral-8x7b.Q4_K_M.gguf on my Linux Debian 12 machine, 32 RAM both with GPT4all (v2.5.4) and LM Studio (AppImage LM+Studio-0.2.8-beta-v1) and on both applications, the model throws the following error wenn I want to load the model:

Error loading model: create_tensor: tensor 'blk.0.ffn_gate.weight' not found
llama_load_model_from_file: failed to load model
llama_init_from_gpt_params: error: failed to load model '/media/berion/daten/gpt4all/models/TheBloke/dolphin-2.5-mixtral-8x7b-GGUF/dolphin-2.5-mixtral-8x7b.Q4_K_M.gguf'
load: error: failed to load model '/media/berion/daten/gpt4all/models/TheBloke/dolphin-2.5-mixtral-8x7b-GGUF/dolphin-2.5-mixtral-8x7b.Q4_K_M.gguf'
Debug #4/media/berion/daten/gpt4all/models/TheBloke/dolphin-2.5-mixtral-8x7b-GGUF/dolphin-2.5-mixtral-8x7b.Q4_K_M.gguf
llama.ts: failed to load model. Error: create_tensor: tensor 'blk.0.ffn_gate.weight' not found
[modelprocess.ts] Error in loadModel: Error: create_tensor: tensor 'blk.0.ffn_gate.weight' not found
on message 2 [remotellm.ts] LLM process error: Error: create_tensor: tensor 'blk.0.ffn_gate.weight' not found
[modelprocesscontainer.ts] error handler: Error: Error: create_tensor: tensor 'blk.0.ffn_gate.weight' not found
[appserver] Error from LLM: Error: Error: create_tensor: tensor 'blk.0.ffn_gate.weight' not found
[appserver] No inferencingReplyPort to send error to
Error: create_tensor: tensor 'blk.0.ffn_gate.weight' not found
    at t.Llama.<anonymous> (/tmp/.mount_LM+Stu4ZNZrJ/resources/app/.webpack/main/utility.js:2:261927)
    at Generator.next (<anonymous>)
    at r (/tmp/.mount_LM+Stu4ZNZrJ/resources/app/.webpack/main/utility.js:2:260462)
[modelprocesscontainer.ts] #forkUtilityProcessAndLoadModel Error loading model: Error: create_tensor: tensor 'blk.0.ffn_gate.weight' not found
[modelprocesscontainer.ts] error handler: Error: Model failed (exit code: 42). Please try loading it again.
[appserver] Error from LLM: Error: Model failed (exit code: 42). Please try loading it again.
[appserver] No inferencingReplyPort to send error to
[remotellm.ts] INSIDE LOAD PROMISE EXIT LLM process exited with code 42.

I tried it also with a newer version like dolphin-2.6-mixtral-8x7b.Q3_K_M.gguf ending up the same and
the minimal model dolphin-2.5-mixtral-8x7b.Q2_K.gguf

I've also checked the SHA256 checksum. Everything should be fine with the downloaded GGUF file.

$ sha256sum dolphin-2.5-mixtral-8x7b.Q4_K_M.gguf 
ee4474c3f6f9c20d7ba738ab63c7c19d3e1471749a72e8624e5b9db0c827b525  dolphin-2.5-mixtral-8x7b.Q4_K_M.gguf

Any suggestions how to fix this issue?

My memory configurations:

$ cat /proc/meminfo
MemTotal:       32643500 kB
MemFree:        10057616 kB
MemAvailable:   26913816 kB
Buffers:         6538016 kB
Cached:         11745012 kB
SwapCached:            0 kB
Active:          1800968 kB
Inactive:       18772184 kB
Active(anon):       8208 kB
Inactive(anon):  3581172 kB
Active(file):    1792760 kB
Inactive(file): 15191012 kB
Unevictable:     1229592 kB
Mlocked:            5964 kB
SwapTotal:       7812092 kB
SwapFree:        7812092 kB
Zswap:                 0 kB
Zswapped:              0 kB
Dirty:              5600 kB
Writeback:             0 kB
AnonPages:       3469052 kB
Mapped:           768908 kB
Shmem:           1294052 kB
KReclaimable:     341452 kB
Slab:             511628 kB
SReclaimable:     341452 kB
SUnreclaim:       170176 kB
KernelStack:       25984 kB
PageTables:        50308 kB
SecPageTables:         0 kB
NFS_Unstable:          0 kB
Bounce:                0 kB
WritebackTmp:          0 kB
CommitLimit:    24133840 kB
Committed_AS:   12432884 kB
VmallocTotal:   34359738367 kB
VmallocUsed:       72596 kB
VmallocChunk:          0 kB
Percpu:             7520 kB
HardwareCorrupted:     0 kB
AnonHugePages:    811008 kB
ShmemHugePages:  1105920 kB
ShmemPmdMapped:        0 kB
FileHugePages:         0 kB
FilePmdMapped:         0 kB
HugePages_Total:       0
HugePages_Free:        0
HugePages_Rsvd:        0
HugePages_Surp:        0
Hugepagesize:       2048 kB
Hugetlb:               0 kB
DirectMap4k:      190068 kB
DirectMap2M:     3780608 kB
DirectMap1G:    30408704 kB

Sign up or log in to comment