kartikmosaicml commited on
Commit
05067c3
1 Parent(s): ec40b36

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -4
README.md CHANGED
@@ -43,7 +43,6 @@ The following models are finetuned on MPT-30B:
43
  * [MPT-30B-Instruct](https://huggingface.co/mosaicml/mpt-30b-instruct): a model for short-form instruction following.
44
  Built by finetuning MPT-30B on several carefully curated datasets.
45
  * License: _CC-By-NC-SA-3.0_
46
- * [Demo on Hugging Face Spaces](https://huggingface.co/spaces/mosaicml/mpt-30b-instruct)
47
 
48
  * [MPT-30B-Chat](https://huggingface.co/mosaicml/mpt-30b-chat): a chatbot-like model for dialogue generation.
49
  Built by finetuning MPT-30B on [ShareGPT-Vicuna](https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered), [Camel-AI](https://huggingface.co/camel-ai),
@@ -61,7 +60,7 @@ Apache-2.0
61
 
62
  ## Documentation
63
 
64
- * [Blog post: MPT-30B: Raising the bar for open-source commercial foundation models](https://www.mosaicml.com/blog/mpt-30b)
65
  * [Codebase (mosaicml/llm-foundry repo)](https://github.com/mosaicml/llm-foundry/)
66
  * Questions: Feel free to contact us via the [MosaicML Community Slack](https://mosaicml.me/slack)!
67
 
@@ -236,8 +235,8 @@ Please cite this model using the following format:
236
  ```
237
  @online{MosaicML2023Introducing,
238
  author = {MosaicML NLP Team},
239
- title = {Introducing MPT-30B: A New Standard for Open-Source,
240
- ly Usable LLMs},
241
  year = {2023},
242
  url = {www.mosaicml.com/blog/mpt-30b},
243
  note = {Accessed: 2023-06-22},
 
43
  * [MPT-30B-Instruct](https://huggingface.co/mosaicml/mpt-30b-instruct): a model for short-form instruction following.
44
  Built by finetuning MPT-30B on several carefully curated datasets.
45
  * License: _CC-By-NC-SA-3.0_
 
46
 
47
  * [MPT-30B-Chat](https://huggingface.co/mosaicml/mpt-30b-chat): a chatbot-like model for dialogue generation.
48
  Built by finetuning MPT-30B on [ShareGPT-Vicuna](https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered), [Camel-AI](https://huggingface.co/camel-ai),
 
60
 
61
  ## Documentation
62
 
63
+ * [Blog post: MPT-30B: Raising the bar for open-source foundation models](https://www.mosaicml.com/blog/mpt-30b)
64
  * [Codebase (mosaicml/llm-foundry repo)](https://github.com/mosaicml/llm-foundry/)
65
  * Questions: Feel free to contact us via the [MosaicML Community Slack](https://mosaicml.me/slack)!
66
 
 
235
  ```
236
  @online{MosaicML2023Introducing,
237
  author = {MosaicML NLP Team},
238
+ title = {Introducing MPT-30B: Raising the bar
239
+ for open-source foundation models},
240
  year = {2023},
241
  url = {www.mosaicml.com/blog/mpt-30b},
242
  note = {Accessed: 2023-06-22},