guoday commited on
Commit
1c90537
1 Parent(s): a5db2fc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -3
README.md CHANGED
@@ -12,7 +12,7 @@ Deepseek Coder comprises a series of code language models trained on both 87% co
12
 
13
  - **Massive Training Data**: Trained on 2T tokens, including 87% code and 13% linguistic data in both English and Chinese languages.
14
 
15
- - **Highly Flexible & Scalable**: Offered in model sizes of 1B, 7B, and 33B, enabling users to choose the setup most suitable for their requirements.
16
 
17
  - **Superior Model Performance**: State-of-the-art performance among publicly available code models on HumanEval, MultiPL-E, MBPP, DS-1000, and APPS benchmarks.
18
 
@@ -167,8 +167,6 @@ outputs = model.generate(**inputs, max_new_tokens=140)
167
  print(tokenizer.decode(outputs[0]))
168
  ```
169
 
170
- ---
171
- In the following scenario, the Deepseek-Coder 7B model effectively calls a class **IrisClassifier** and its member function from the `model.py` file, and also utilizes functions from the `utils.py` file, to correctly complete the **main** function in`main.py` file for model training and evaluation.
172
 
173
 
174
  ### 4. Lincense
 
12
 
13
  - **Massive Training Data**: Trained on 2T tokens, including 87% code and 13% linguistic data in both English and Chinese languages.
14
 
15
+ - **Highly Flexible & Scalable**: Offered in model sizes of 1.3B, 5.7B, 6.7B, and 33B, enabling users to choose the setup most suitable for their requirements.
16
 
17
  - **Superior Model Performance**: State-of-the-art performance among publicly available code models on HumanEval, MultiPL-E, MBPP, DS-1000, and APPS benchmarks.
18
 
 
167
  print(tokenizer.decode(outputs[0]))
168
  ```
169
 
 
 
170
 
171
 
172
  ### 4. Lincense