JosephusCheung commited on
Commit
915124b
1 Parent(s): 508743a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -11,7 +11,9 @@ tags:
11
 
12
  This is the LLaMAfied version of [Qwen/Qwen-7B-Chat](https://huggingface.co/Qwen/Qwen-7B-Chat), recalibrated to fit the original LLaMA/LLaMA-2-like model structure.
13
 
14
- You can use LlamaCausalLM for model inference, which is the same as LLaMA/LLaMA-2 models (the tokenizer remains the same, so you still need to allow external codes when loading, eg: `AutoTokenizer.from_pretrained(llama_model_path, use_fast=False, trust_remote_code=True)`).
 
 
15
 
16
  SPOILOR: Further finetuning is in progress, the current version is a work-in-progress, some knowledge may be biased and illusory due to structural changes. Will be updated very, very sooooooooooon.
17
 
@@ -27,6 +29,8 @@ Issue: Compared to the original Qwen-Chat scoring 53.9, the MMLU score dropped s
27
 
28
  您可以使用 LlamaCausalLM 进行模型推理,和 LLaMA/LLaMA-2 保持一致(分词器保持不变,因此加载时仍然需要允许外部代码,例如:`AutoTokenizer.from_pretrained(llama_model_path, use_fast=False, trust_remote_code=True)`)。
29
 
 
 
30
  剧透: 进一步的微调正在进行中,当前版本是一个正在进行的工作,一些知识可能由于结构变化而产生偏见和幻觉。 会更新,很快,非常非常非常快。
31
 
32
  PROMPT 格式: [chatml](https://github.com/openai/openai-python/blob/main/chatml.md)
 
11
 
12
  This is the LLaMAfied version of [Qwen/Qwen-7B-Chat](https://huggingface.co/Qwen/Qwen-7B-Chat), recalibrated to fit the original LLaMA/LLaMA-2-like model structure.
13
 
14
+ You can use LlamaForCausalLM for model inference, which is the same as LLaMA/LLaMA-2 models (the tokenizer remains the same, so you still need to allow external codes when loading, eg: `AutoTokenizer.from_pretrained(llama_model_path, use_fast=False, trust_remote_code=True)`).
15
+
16
+ The model has been edited to be white-labelled, meaning the model will no longer call itself a Qwen.
17
 
18
  SPOILOR: Further finetuning is in progress, the current version is a work-in-progress, some knowledge may be biased and illusory due to structural changes. Will be updated very, very sooooooooooon.
19
 
 
29
 
30
  您可以使用 LlamaCausalLM 进行模型推理,和 LLaMA/LLaMA-2 保持一致(分词器保持不变,因此加载时仍然需要允许外部代码,例如:`AutoTokenizer.from_pretrained(llama_model_path, use_fast=False, trust_remote_code=True)`)。
31
 
32
+ 模型已经被编辑实现白标化,不再自称通义千问。
33
+
34
  剧透: 进一步的微调正在进行中,当前版本是一个正在进行的工作,一些知识可能由于结构变化而产生偏见和幻觉。 会更新,很快,非常非常非常快。
35
 
36
  PROMPT 格式: [chatml](https://github.com/openai/openai-python/blob/main/chatml.md)