Transformers
mpt
Composer
MosaicML
llm-foundry
text-generation-inference
TheBloke commited on
Commit
e1bcf08
1 Parent(s): 60df632

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -5
README.md CHANGED
@@ -46,11 +46,7 @@ Please note that these GGMLs are **not compatible with llama.cpp, or currently w
46
 
47
  [KoboldCpp](https://github.com/LostRuins/koboldcpp) just added GPU accelerated (OpenCL) support for MPT models, so that is the client I recommend using for these models.
48
 
49
- **Note**: There is currently a bug with loading this model in KoboldCpp Release 1.32: it will wrongly detect it as a GPT-NeoX model.
50
-
51
- To resolve this, add argument `--forceversion 500`
52
-
53
- This should be fixed in the next release of KoboldCpp, so if you are running a version later than 1.32 it should not be necessary.
54
 
55
  ## Repositories available
56
 
 
46
 
47
  [KoboldCpp](https://github.com/LostRuins/koboldcpp) just added GPU accelerated (OpenCL) support for MPT models, so that is the client I recommend using for these models.
48
 
49
+ **Note**: Please make sure you're using KoboldCpp version 1.32.3 or later, as a number of MPT-related bugs are fixed.
 
 
 
 
50
 
51
  ## Repositories available
52