From 8b2d2b0dc2952c6e376c973d1702a03b17c8c4aa Mon Sep 17 00:00:00 2001 From: Yang An Date: Tue, 12 Sep 2023 11:30:53 +0800 Subject: [PATCH] Update README_CN.md --- README_CN.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/README_CN.md b/README_CN.md index 8e95b8a..501d8ec 100644 --- a/README_CN.md +++ b/README_CN.md @@ -175,9 +175,9 @@ print(tokenizer.decode(pred.cpu()[0], skip_special_tokens=True)) from modelscope import AutoModelForCausalLM, AutoTokenizer from modelscope import GenerationConfig -tokenizer = AutoTokenizer.from_pretrained("qwen/Qwen-7B-Chat", revision = 'v1.0.5',trust_remote_code=True) -model = AutoModelForCausalLM.from_pretrained("qwen/Qwen-7B-Chat", revision = 'v1.0.5',device_map="auto", trust_remote_code=True,fp16 = True).eval() -model.generation_config = GenerationConfig.from_pretrained("Qwen/Qwen-7B-Chat",revision = 'v1.0.5', trust_remote_code=True) # 可指定不同的生成长度、top_p等相关超参 +tokenizer = AutoTokenizer.from_pretrained("qwen/Qwen-7B-Chat", revision='v1.0.5', trust_remote_code=True) +model = AutoModelForCausalLM.from_pretrained("qwen/Qwen-7B-Chat", revision='v1.0.5', device_map="auto", trust_remote_code=True, fp16=True).eval() +model.generation_config = GenerationConfig.from_pretrained("Qwen/Qwen-7B-Chat", revision='v1.0.5', trust_remote_code=True) # 可指定不同的生成长度、top_p等相关超参 response, history = model.chat(tokenizer, "你好", history=None) print(response)