ekshat commited on
Commit
78294ad
1 Parent(s): 9cb7102

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -15
README.md CHANGED
@@ -3,30 +3,30 @@ license: apache-2.0
3
  ---
4
  This repository hosts both the standard and quantized versions of the Zephyr 7B model, allowing users to choose the version that best fits their resource constraints and performance needs.
5
 
6
- Model Details
7
- Model Name: Zephyr 7B
8
- Model Size: 7 billion parameters
9
- Architecture: Transformer-based
10
- Languages: Primarily English, with support for multilingual text
11
- Quantized Version: Available for reduced memory footprint and faster inference
12
-
13
- Performance and Efficiency
14
  The quantized version of Zephyr 7B is optimized for environments with limited computational resources. It offers:
15
 
16
- Reduced Memory Usage: The model size is significantly smaller, making it suitable for deployment on devices with limited RAM.
17
- Faster Inference: Quantized models can perform faster inference, providing quicker responses in real-time applications.
18
 
19
- Fine-Tuning
20
  You can fine-tune the Zephyr 7B model on your own dataset to better suit specific tasks or domains. Refer to the Huggingface documentation for guidance on how to fine-tune transformer models.
21
 
22
- Contributing
23
  We welcome contributions to improve the Zephyr 7B model. Please submit pull requests or open issues for any enhancements or bugs you encounter.
24
 
25
- License
26
  This model is licensed under the MIT License.
27
 
28
- Acknowledgments
29
  Special thanks to the Huggingface team for providing the transformers library and to the broader AI community for their continuous support and contributions.
30
 
31
- Contact
32
  For any questions or inquiries, please contact us at akshayhedaoo7246@gmail.com.
 
3
  ---
4
  This repository hosts both the standard and quantized versions of the Zephyr 7B model, allowing users to choose the version that best fits their resource constraints and performance needs.
5
 
6
+ # Model Details
7
+ ## Model Name: Zephyr 7B
8
+ ## Model Size: 7 billion parameters
9
+ ## Architecture: Transformer-based
10
+ ## Languages: Primarily English, with support for multilingual text
11
+ ## Quantized Version: Available for reduced memory footprint and faster inference
12
+
13
+ # Performance and Efficiency
14
  The quantized version of Zephyr 7B is optimized for environments with limited computational resources. It offers:
15
 
16
+ ## Reduced Memory Usage: The model size is significantly smaller, making it suitable for deployment on devices with limited RAM.
17
+ ## Faster Inference: Quantized models can perform faster inference, providing quicker responses in real-time applications.
18
 
19
+ # Fine-Tuning
20
  You can fine-tune the Zephyr 7B model on your own dataset to better suit specific tasks or domains. Refer to the Huggingface documentation for guidance on how to fine-tune transformer models.
21
 
22
+ # Contributing
23
  We welcome contributions to improve the Zephyr 7B model. Please submit pull requests or open issues for any enhancements or bugs you encounter.
24
 
25
+ # License
26
  This model is licensed under the MIT License.
27
 
28
+ # Acknowledgments
29
  Special thanks to the Huggingface team for providing the transformers library and to the broader AI community for their continuous support and contributions.
30
 
31
+ # Contact
32
  For any questions or inquiries, please contact us at akshayhedaoo7246@gmail.com.