Max's picture

Max

moock
Β·

AI & ML interests

None yet

Recent Activity

Organizations

MXMK's profile picture huggingPartyParis's profile picture

moock's activity

upvoted an article 19 days ago
view article
Article

Use Models from the Hugging Face Hub in LM Studio

By yagilb β€’
β€’ 127
reacted to clem's post with πŸš€ 6 months ago
view post
Post
5781
5,000 new repos (models, datasets, spaces) are created EVERY DAY on HF now. The community is amazing!
replied to lunarflu's post 7 months ago
view reply

It would be fun to have a prediction of my future daily activities πŸͺ„

reacted to lunarflu's post with πŸ”₯ 7 months ago
view post
Post
1933
cooking up something....anyone interested in a daily activity tracker for HF?
Β·
reacted to singhsidhukuldeep's post with πŸ‘ 7 months ago
view post
Post
2085
🎭 You picked an LLM for your work but then you find out it hallucinates! πŸ€–

πŸ€” Your first thought might be to fine-tune it on more training data.... but should you? πŸ› οΈ

πŸ“œ This is what @Google is exploring in the paper "Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?" πŸ•΅οΈβ€β™‚οΈ

πŸ“˜ When LLMs undergo supervised fine-tuning with new factual knowledge not present in their initial training data, there is a risk they might "hallucinate" or produce factually incorrect information. 🚨

πŸ” The paper investigates how fine-tuning LLMs with new facts influences their ability to leverage pre-existing knowledge and the extent to which they generate errors. πŸ“Š

βš™οΈTechnical Setup:

πŸ”§ Approach: They introduce a system named SliCK (this stands for Sampling-based Categorization of Knowledge, don't even bother understanding how) to categorize knowledge into four levels (HighlyKnown, MaybeKnown, WeaklyKnown, and Unknown) based on how well the model's generated responses agree with known facts. πŸ—‚οΈ

πŸ“ Experimental Setup: The study uses a controlled setup focusing on closed-book QA, adjusting the proportion of fine-tuning examples that introduce new facts versus those that do not. πŸ§ͺ

πŸ‘‰ Here is the gist of the findings:

🚸 LLMs struggle to integrate new factual knowledge during fine-tuning, and such examples are learned slower than those consistent with the model's pre-existing knowledge. 🐒

πŸ“ˆ As LLMs learn from examples containing new knowledge, their propensity to hallucinate increases. πŸ‘»

⏱️ Early stopping during training can mitigate the risks of hallucinations by minimizing exposure to unlearned new facts. πŸ›‘

🧠 Training LLMs mostly with known examples leads to better utilization of pre-existing knowledge, whereas examples introducing new knowledge increase the risk of generating incorrect information. πŸ—οΈ

πŸ“„ Paper: Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations? (2405.05904) πŸ“š
  • 2 replies
Β·
New activity in yanze/PuLID 7 months ago
liked a Space 8 months ago