| |
| |
| |
| |
| |
| FROM nvidia/cuda:11.3.1-runtime-ubuntu20.04 |
| ARG useProxyNetwork='' |
| RUN apt-get update |
| RUN apt-get install -y curl proxychains curl |
| RUN apt-get install -y git python python3 python-dev python3-dev --fix-missing |
|
|
| |
| |
| RUN $useProxyNetwork curl cip.cc |
| RUN sed -i '$ d' /etc/proxychains.conf |
| RUN sed -i '$ d' /etc/proxychains.conf |
| |
| RUN echo "socks5 127.0.0.1 10880" >> /etc/proxychains.conf |
| ARG useProxyNetwork=proxychains |
| |
|
|
|
|
| |
| RUN curl -sS https://bootstrap.pypa.io/get-pip.py | python3.8 |
| |
| RUN $useProxyNetwork python3 -m pip install torch --extra-index-url https://download.pytorch.org/whl/cu113 |
| |
| WORKDIR /gpt |
| RUN $useProxyNetwork git clone https://github.com/binary-husky/chatgpt_academic.git |
| WORKDIR /gpt/chatgpt_academic |
| RUN $useProxyNetwork python3 -m pip install -r requirements.txt |
| RUN $useProxyNetwork python3 -m pip install -r request_llm/requirements_chatglm.txt |
| RUN $useProxyNetwork python3 -m pip install -r request_llm/requirements_newbing.txt |
|
|
| |
| RUN echo ' \n\ |
| from transformers import AutoModel, AutoTokenizer \n\ |
| chatglm_tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True) \n\ |
| chatglm_model = AutoModel.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True).float() ' >> warm_up_chatglm.py |
| RUN python3 -u warm_up_chatglm.py |
|
|
| |
| ADD "https://www.random.org/cgi-bin/randbyte?nbytes=10&format=h" skipcache |
| RUN $useProxyNetwork git pull |
|
|
| |
| RUN python3 -c 'from check_proxy import warm_up_modules; warm_up_modules()' |
|
|
| |
| |
| |
| |
| |
| RUN echo ' \n\ |
| API_KEY = "sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx,fkxxxxxx-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" \n\ |
| USE_PROXY = True \n\ |
| LLM_MODEL = "chatglm" \n\ |
| LOCAL_MODEL_DEVICE = "cuda" \n\ |
| proxies = { "http": "socks5h://localhost:10880", "https": "socks5h://localhost:10880", } ' >> config_private.py |
|
|
| |
| CMD ["python3", "-u", "main.py"] |
|
|