**Usage** # 设置环境 git clone https://github.com/hiyouga/ChatGLM-Efficient-Tuning.git conda create -n chatglm_etuning python=3.10 conda activate chatglm_etuning cd ChatGLM-Efficient-Tuning pip install -r requirements.txt # 模型部署,新建文件夹path_to_checkpoint,下载Laurie/ChatGLM-Efficient-Tuning-SFT里的文件到里面 import sys sys.path.append("src") from src import load_pretrained, ModelArguments model_args = ModelArguments(checkpoint_dir=path_to_checkpoint) model, tokenizer = load_pretrained(model_args) model = model.cuda() model.eval() # model.generate, model.chat()... # 命令行进行测试 CUDA_VISIBLE_DEVICES=0 python src/infer.py \ --checkpoint_dir path_to_checkpoint