Fuli Luo
luofuli
AI & ML interests
None yet
Recent Activity
authored
a paper
about 2 months ago
DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via
Reinforcement Learning
liked
a model
3 months ago
deepseek-ai/DeepSeek-V3-Base
liked
a model
5 months ago
mistralai/Mamba-Codestral-7B-v0.1
Organizations
None yet
luofuli's activity
What's the diff with deepseek-ai/deepseek-moe-16b-chat ?
1
#3 opened 10 months ago
by
JohnSaxon
training: fix type mismatch when training
1
#6 opened 10 months ago
by
Jack477
function/tool calling support
8
#5 opened 10 months ago
by
kaijietti
Adding `safetensors` variant of this model
#3 opened about 1 year ago
by
SFconvertbot

Adding `safetensors` variant of this model
#4 opened 12 months ago
by
abalogh
Intermediate Pretraining Checkpoints?
1
#3 opened about 1 year ago
by
RylanSchaeffer
Intermediate Pretraining Checkpoints?
1
#1 opened about 1 year ago
by
RylanSchaeffer
Intermediate Pretraining Checkpoints?
1
#5 opened about 1 year ago
by
RylanSchaeffer
max_positional_embeddings
1
#2 opened about 1 year ago
by
ehartford

Missing `tokenizer.model`
3
#3 opened about 1 year ago
by
AlienKevin

Please publish LLM 1.7b base
2
#2 opened about 1 year ago
by
SinanAkkoyun

llm-1.3b-base
1
#2 opened about 1 year ago
by
SinanAkkoyun

4-bit quant?
2
#3 opened 12 months ago
by
Neman

tokenizer.model
2
#26 opened 12 months ago
by
BigDeeper
Why do we need the line trust_remote_code=True?
2
#23 opened about 1 year ago
by
Rubiel1
excellent results
1
#1 opened about 1 year ago
by
Tonic

Failed to Deploy this model in Inference Endpoints
3
#19 opened about 1 year ago
by
calvinball
Does it matter if the prompt is incomplete?
1
#4 opened over 1 year ago
by
Hamlyn
tokenizer.model
1
#6 opened over 1 year ago
by
RonanMcGovern

Exllamav2 need tokenizer.model to load
1
#8 opened over 1 year ago
by
CTXEE