Rakshit Aralimatti

RakshitAralimatti

AI & ML interests

generative AI , NLP

Recent Activity

Organizations

None yet

RakshitAralimatti's activity

Getting Error

3
#1 opened 2 months ago by RakshitAralimatti
New activity in LLM360/TxT360 11 days ago

Size of the Dataset?

6
#9 opened 16 days ago by RakshitAralimatti
New activity in mlfoundations/dclm-baseline-1.0 16 days ago
New activity in varma007ut/Indian_Legal_Assitant about 2 months ago

Dataset Used for Fine-tuning

#1 opened about 2 months ago by RakshitAralimatti
Reacted to bartowski's post with โค๏ธ 3 months ago
view post
Post
10014
So turns out I've been spreading a bit of misinformation when it comes to imatrix in llama.cpp

It starts true; imatrix runs the model against a corpus of text and tracks the activation of weights to determine which are most important

However what the quantization then does with that information is where I was wrong.

I think I made the accidental connection between imatrix and exllamav2's measuring, where ExLlamaV2 decides how many bits to assign to which weight depending on the goal BPW

Instead, what llama.cpp with imatrix does is it attempts to select a scale for a quantization block that most accurately returns the important weights to their original values, ie minimizing the dequantization error based on the importance of activations

The mildly surprising part is that it actually just does a relatively brute force search, it picks a bunch of scales and tries each and sees which one results in the minimum error for weights deemed important in the group

But yeah, turns out, the quantization scheme is always the same, it's just that the scaling has a bit more logic to it when you use imatrix

Huge shoutout to @compilade for helping me wrap my head around it - feel free to add/correct as well if I've messed something up
ยท
Reacted to as-cle-bert's post with ๐Ÿš€ 4 months ago
view post
Post
5046
Hi HF Community!๐Ÿค—

In the past days, OpenAI announced their search engine, SearchGPT: today, I'm glad to introduce you SearchPhi, an AI-powered and open-source web search tool that aims to reproduce similar features to SearchGPT, built upon microsoft/Phi-3-mini-4k-instruct, llama.cpp๐Ÿฆ™ and Streamlit.
Although not as capable as SearchGPT, SearchPhi v0.0-beta.0 is a first step toward a fully functional and multimodal search engine :)
If you want to know more, head over to the GitHub repository (https://github.com/AstraBert/SearchPhi) and, to test it out, use this HF space: as-cle-bert/SearchPhi
Have fun!๐Ÿฑ
Reacted to victor's post with โค๏ธ 5 months ago
Reacted to KingNish's post with ๐Ÿ”ฅโž• 6 months ago
view post
Post
4626
Microsoft Just Launched 3 Powerful Models

1. Phi 3 Medium (4k and 128k): A 14b Instruct tuned models that outperformed big models like Command R+ (104b), GPT 3.5 Pro, Gemini Pro, and is highly competitive with top models such as Mixtral 8x22b, Llama3 70B, and GPT 4.
microsoft/Phi-3-medium-4k-instruct
DEMO: https://huggingface.co/spaces/Walmart-the-bag/Phi-3-Medium

2. Phi 3 Mini Vision 128k: A 4.5 billion-parameter, instruction-tuned vision model that has outperformed models such as Llava3 and Claude 3, and is providing stiff competition to Gemini 1Pro Vision.
microsoft/Phi-3-vision-128k-instruct

3. Phi3 Small (8k and 128k): Better than Llama3 8b, Mixtral 8x7b and GPT 3.5 turbo.
microsoft/Phi-3-small-128k-instruct
ยท