diff --git a/cli_demo.py b/cli_demo.py index bc1efb5..76f88c4 100644 --- a/cli_demo.py +++ b/cli_demo.py @@ -7,7 +7,7 @@ from utils import load_model_on_gpus tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True) model = AutoModel.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True, device='cuda') -# 多显卡支持,注释掉下面三行,将num_gpus改为你实际的显卡数量 +# 多显卡支持,使用下面三行代替上面两行,将num_gpus改为你实际的显卡数量 # model_path = "THUDM/chatglm2-6b" # tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) # model = load_model_on_gpus(model_path, num_gpus=2) diff --git a/web_demo.py b/web_demo.py index eaa0c9c..27bfaa4 100644 --- a/web_demo.py +++ b/web_demo.py @@ -4,7 +4,7 @@ import mdtex2html tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True) model = AutoModel.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True, device='cuda') -# 多显卡支持,注释掉下面三行,将num_gpus改为你实际的显卡数量 +# 多显卡支持,使用下面三行代替上面两行,将num_gpus改为你实际的显卡数量 # model_path = "THUDM/chatglm2-6b" # tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) # model = load_model_on_gpus(model_path, num_gpus=2)