Pruna AI
company
AI & ML interests
Efficient machine learning for any model and hardware: pruning, quantization, compilation, and more.
Organization Card
About org cards
Simply make AI models faster, cheaper, smaller, greener!
Pruna AI makes, in one line of code, any AI model faster, cheaper, smaller, greener on any hardware. It covers CV, NLP, audio, graphs for predictive and generative AI.
We provide two packages. You can read their documentations to know more here.
pruna
: Package to smash your AI model to make it more efficient without losing quality. For this, you only need to callpruna.smash()
. If you want to compress models on your side, you can request access here.pruna_engine
: Package to run your AI model more efficiently without changing your pipeline. If you want to run models publicly shared on HuggingFace, you can install it from here.
models
3120
PrunaAI/nvidia-Llama3-ChatQA-1.5-8B-GGUF-smashed
Updated
•
12k
PrunaAI/rinna-japanese-gpt-neox-3.6b-HQQ-1bit-smashed
Text Generation
•
Updated
•
1
PrunaAI/codellama-CodeLlama-13b-Python-hf-HQQ-2bit-smashed
Text Generation
•
Updated
PrunaAI/maywell-Llama-3-Ko-8B-Instruct-AWQ-4bit-smashed
Text Generation
•
Updated
•
1
PrunaAI/zjunlp-OceanGPT-7b-v0.1-QUANTO-int2bit-smashed
Updated
PrunaAI/zjunlp-OceanGPT-7b-v0.1-HQQ-4bit-smashed
Text Generation
•
Updated
PrunaAI/zjunlp-OceanGPT-7b-v0.1-HQQ-2bit-smashed
Text Generation
•
Updated
PrunaAI/zjunlp-OceanGPT-7b-v0.1-HQQ-1bit-smashed
Text Generation
•
Updated
PrunaAI/zjunlp-OceanGPT-7b-v0.1-bnb-4bit-smashed
Text Generation
•
Updated
PrunaAI/nickypro-tinyllama-15M-bnb-4bit-smashed
Text Generation
•
Updated
•
1
datasets
None public yet