GGUF
Composer
MosaicML
llm-foundry
maddes8cht commited on
Commit
97dc9e2
·
1 Parent(s): 6db6b73

"Update README.md"

Browse files
Files changed (1) hide show
  1. README.md +5 -12
README.md CHANGED
@@ -16,21 +16,14 @@ I'm constantly enhancing these model descriptions to provide you with the most r
16
  - Model creator: [mosaicml](https://huggingface.co/mosaicml)
17
  - Original model: [mpt-7b-storywriter](https://huggingface.co/mosaicml/mpt-7b-storywriter)
18
 
19
- # Important Update for Falcon Models in llama.cpp Versions After October 18, 2023
20
 
21
- As noted on the [Llama.cpp GitHub repository](https://github.com/ggerganov/llama.cpp#hot-topics), all new Llama.cpp releases after October 18, 2023, will require a re-quantization due to the new BPE tokenizer.
22
 
23
- **Good news!** I am glad that my re-quantization process for Falcon Models is nearly complete. Download the latest quantized models to ensure compatibility with recent llama.cpp software.
24
-
25
- **Key Points:**
26
-
27
- - **Stay Informed:** Keep an eye on software application release schedules using llama.cpp libraries.
28
- - **Monitor Upload Times:** Re-quantization is *almost* done. Watch for updates on my Hugging Face Model pages.
29
-
30
- **Important Compatibility Note:** Old software will work with old Falcon models, but expect updated software to exclusively support the new models.
31
-
32
- This change primarily affects **Falcon** and **Starcoder** models, with other models remaining unaffected.
33
 
 
34
 
35
 
36
 
 
16
  - Model creator: [mosaicml](https://huggingface.co/mosaicml)
17
  - Original model: [mpt-7b-storywriter](https://huggingface.co/mosaicml/mpt-7b-storywriter)
18
 
19
+ MPT-7b and MPT-30B are part of the family of Mosaic Pretrained Transformer (MPT) models, which use a modified transformer architecture optimized for efficient training and inference.
20
 
 
21
 
22
+ ---
23
+ # Brief
24
+ The MPT-7B and MPT-30B Models are part of the family of Mosaic Pretrained Transformer (MPT) models, which use a modified transformer architecture optimized for efficient training and inference.
 
 
 
 
 
 
 
25
 
26
+ ---
27
 
28
 
29