--- tags: - finetuned - quantized - 4-bit - AWQ - transformers - pytorch - safetensors - mistral - text-generation - finetuned - conversational - arxiv:2310.06825 - license:apache-2.0 - autotrain_compatible - has_space - text-generation-inference - region:us model_name: Mistral-7B-Instruct-v0.1-AWQ base_model: mistralai/Mistral-7B-Instruct-v0.1 inference: false model_creator: mistralai pipeline_tag: text-generation quantized_by: MaziyarPanahi --- # Description [MaziyarPanahi/Mistral-7B-Instruct-v0.1-AWQ](https://huggingface.co/MaziyarPanahi/Mistral-7B-Instruct-v0.1-AWQ) is a quantized (AWQ) version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) ## How to use ### Install the necessary packages ``` pip install --upgrade accelerate autoawq transformers ``` ### Example Python code ```python from transformers import AutoTokenizer, AutoModelForCausalLM model_id = "MaziyarPanahi/Mistral-7B-Instruct-v0.1-AWQ" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained(model_id).to(0) text = "User:\nHello can you provide me with top-3 cool places to visit in Paris?\n\nAssistant:\n" inputs = tokenizer(text, return_tensors="pt").to(0) out = model.generate(**inputs, max_new_tokens=300) print(tokenizer.decode(out[0], skip_special_tokens=True)) ```