Fuli Luo
luofuli
AI & ML interests
None yet
Organizations
None yet
luofuli's activity
What's the diff with deepseek-ai/deepseek-moe-16b-chat ?
1
#3 opened 7 months ago
by
JohnSaxon
training: fix type mismatch when training
1
#6 opened 7 months ago
by
Jack477
function/tool calling support
7
#5 opened 8 months ago
by
kaijietti
Adding `safetensors` variant of this model
#3 opened 10 months ago
by
SFconvertbot
Adding `safetensors` variant of this model
#4 opened 9 months ago
by
abalogh
Intermediate Pretraining Checkpoints?
1
#3 opened 10 months ago
by
RylanSchaeffer
Intermediate Pretraining Checkpoints?
1
#1 opened 10 months ago
by
RylanSchaeffer
Intermediate Pretraining Checkpoints?
1
#5 opened 10 months ago
by
RylanSchaeffer
max_positional_embeddings
1
#2 opened 10 months ago
by
ehartford
Missing `tokenizer.model`
3
#3 opened 10 months ago
by
AlienKevin
Please publish LLM 1.7b base
2
#2 opened 10 months ago
by
SinanAkkoyun
llm-1.3b-base
1
#2 opened 10 months ago
by
SinanAkkoyun
4-bit quant?
2
#3 opened 9 months ago
by
Neman
tokenizer.model
2
#26 opened 9 months ago
by
BigDeeper
Why do we need the line trust_remote_code=True?
2
#23 opened 10 months ago
by
Rubiel1
excellent results
1
#1 opened 11 months ago
by
Tonic
Failed to Deploy this model in Inference Endpoints
3
#19 opened 12 months ago
by
calvinball
Does it matter if the prompt is incomplete?
1
#4 opened about 1 year ago
by
Hamlyn
tokenizer.model
1
#6 opened about 1 year ago
by
RonanMcGovern
Exllamav2 need tokenizer.model to load
1
#8 opened about 1 year ago
by
CTXEE