JosephusCheung
commited on
Commit
•
61c626f
1
Parent(s):
a0fec25
Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,7 @@ tags:
|
|
11 |
|
12 |
This is the LLaMAfied version of [Qwen/Qwen-7B-Chat](https://huggingface.co/Qwen/Qwen-7B-Chat), recalibrated to fit the original LLaMA/LLaMA-2-like model structure.
|
13 |
|
14 |
-
You can use LlamaForCausalLM for model inference, which is the same as LLaMA/LLaMA-2 models (
|
15 |
|
16 |
The model has been edited to be white-labelled, meaning the model will no longer call itself a Qwen.
|
17 |
|
@@ -27,7 +27,7 @@ Issue: Compared to the original Qwen-Chat scoring 53.9, the MMLU score dropped s
|
|
27 |
|
28 |
这是 [通义千问 Qwen/Qwen-7B-Chat](https://huggingface.co/Qwen/Qwen-7B-Chat) 的 LLaMA 化版本,经过重新校准以适应原始的类似 LLaMA/LLaMA-2 的模型结构。
|
29 |
|
30 |
-
您可以使用 LlamaCausalLM 进行模型推理,和 LLaMA/LLaMA-2
|
31 |
|
32 |
模型已经被编辑实现白标化,不再自称通义千问。
|
33 |
|
|
|
11 |
|
12 |
This is the LLaMAfied version of [Qwen/Qwen-7B-Chat](https://huggingface.co/Qwen/Qwen-7B-Chat), recalibrated to fit the original LLaMA/LLaMA-2-like model structure.
|
13 |
|
14 |
+
You can use LlamaForCausalLM for model inference, which is the same as LLaMA/LLaMA-2 models (using GPT2Tokenizer converted from the original tiktoken, by [vonjack](https://huggingface.co/vonjack)).
|
15 |
|
16 |
The model has been edited to be white-labelled, meaning the model will no longer call itself a Qwen.
|
17 |
|
|
|
27 |
|
28 |
这是 [通义千问 Qwen/Qwen-7B-Chat](https://huggingface.co/Qwen/Qwen-7B-Chat) 的 LLaMA 化版本,经过重新校准以适应原始的类似 LLaMA/LLaMA-2 的模型结构。
|
29 |
|
30 |
+
您可以使用 LlamaCausalLM 进行模型推理,和 LLaMA/LLaMA-2 保持一致(使用由 [vonjack](https://huggingface.co/vonjack) 从原始 tiktoken 转换而来的 GPT2Tokenizer 分词器)。
|
31 |
|
32 |
模型已经被编辑实现白标化,不再自称通义千问。
|
33 |
|