| | |
| | |
| | |
| | |
| | |
| | FROM nvidia/cuda:11.3.1-runtime-ubuntu20.04 |
| | ARG useProxyNetwork='' |
| | RUN apt-get update |
| | RUN apt-get install -y curl proxychains curl |
| | RUN apt-get install -y git python python3 python-dev python3-dev --fix-missing |
| |
|
| | |
| | |
| | RUN $useProxyNetwork curl cip.cc |
| | RUN sed -i '$ d' /etc/proxychains.conf |
| | RUN sed -i '$ d' /etc/proxychains.conf |
| | |
| | RUN echo "socks5 127.0.0.1 10880" >> /etc/proxychains.conf |
| | ARG useProxyNetwork=proxychains |
| | |
| |
|
| |
|
| | |
| | RUN curl -sS https://bootstrap.pypa.io/get-pip.py | python3.8 |
| | |
| | RUN $useProxyNetwork python3 -m pip install torch --extra-index-url https://download.pytorch.org/whl/cu113 |
| | |
| | WORKDIR /gpt |
| | RUN $useProxyNetwork git clone https://github.com/binary-husky/chatgpt_academic.git |
| | WORKDIR /gpt/chatgpt_academic |
| | RUN $useProxyNetwork python3 -m pip install -r requirements.txt |
| | RUN $useProxyNetwork python3 -m pip install -r request_llm/requirements_chatglm.txt |
| | RUN $useProxyNetwork python3 -m pip install -r request_llm/requirements_newbing.txt |
| |
|
| | |
| | RUN echo ' \n\ |
| | from transformers import AutoModel, AutoTokenizer \n\ |
| | chatglm_tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True) \n\ |
| | chatglm_model = AutoModel.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True).float() ' >> warm_up_chatglm.py |
| | RUN python3 -u warm_up_chatglm.py |
| |
|
| | |
| | ADD "https://www.random.org/cgi-bin/randbyte?nbytes=10&format=h" skipcache |
| | RUN $useProxyNetwork git pull |
| |
|
| | |
| | RUN python3 -c 'from check_proxy import warm_up_modules; warm_up_modules()' |
| |
|
| | |
| | |
| | |
| | |
| | |
| | RUN echo ' \n\ |
| | API_KEY = "sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx,fkxxxxxx-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" \n\ |
| | USE_PROXY = True \n\ |
| | LLM_MODEL = "chatglm" \n\ |
| | LOCAL_MODEL_DEVICE = "cuda" \n\ |
| | proxies = { "http": "socks5h://localhost:10880", "https": "socks5h://localhost:10880", } ' >> config_private.py |
| |
|
| | |
| | CMD ["python3", "-u", "main.py"] |
| |
|