Matthew Hendrey
mrhendrey
AI & ML interests
None yet
Recent Activity
updated
a model
about 13 hours ago
mrhendrey/Llama-3.1-8B-Instruct-quantized.w8a8
published
a model
about 14 hours ago
mrhendrey/Llama-3.1-8B-Instruct-quantized.w8a8
new activity
about 2 months ago
neuralmagic/Sparse-Llama-3.1-8B-ultrachat_200k-2of4-quantized.w4a16:Model only outputs "!!!!!!!!!!"
Organizations
None yet
mrhendrey's activity
New activity in
neuralmagic/Sparse-Llama-3.1-8B-ultrachat_200k-2of4-quantized.w4a16
about 2 months ago
Model only outputs "!!!!!!!!!!"
1
#1 opened about 2 months ago
by
mrhendrey
VRAM consumption when using GPU (CUDA)
3
#37 opened 7 months ago
by
Sunjay353
Batch: inefficient memory
1
#50 opened 7 months ago
by
SinanAkkoyun
Any chance your team is working on a 4-bit Llama-3.2-90B-Vision-Instruct-quantized.w4a16 version?
1
#1 opened 4 months ago
by
mrhendrey