zheyishine commited on
Commit
b01ee29
·
verified ·
1 Parent(s): 3a54ea1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -33,7 +33,7 @@ To enable deployment of [Ring-Linear-2.0](https://github.com/inclusionAI/Ring-V2
33
 
34
  Since the Pull Request (PR) has not been submitted to the vLLM community at this stage, please prepare the environment by following the steps below.
35
 
36
- First, create a Conda environment with Python 3.10 and CUDA 12.8 (Use the root or admin account, or ensure the current user has access to /home/admin/logs):
37
  ```shell
38
  conda create -n vllm python=3.10
39
  conda activate vllm
@@ -41,12 +41,12 @@ conda activate vllm
41
 
42
  Next, install our vLLM wheel package:
43
  ```shell
44
- pip install https://media.githubusercontent.com/media/inclusionAI/Ring-V2/refs/heads/main/hybrid_linear/whls/vllm-0.8.5%2Bcuda12_8_gcc10_2_1-cp310-cp310-linux_x86_64.whl --force-reinstall
45
  ```
46
 
47
  Finally, install compatible versions of PyTorch and Torchvision after vLLM is installed:
48
  ```shell
49
- pip install torch==2.7.0 torchvision==0.22.0
50
  ```
51
 
52
  #### Offline Inference
 
33
 
34
  Since the Pull Request (PR) has not been submitted to the vLLM community at this stage, please prepare the environment by following the steps below.
35
 
36
+ First, create a Conda environment with Python 3.10 and CUDA 12.8:
37
  ```shell
38
  conda create -n vllm python=3.10
39
  conda activate vllm
 
41
 
42
  Next, install our vLLM wheel package:
43
  ```shell
44
+ pip install https://media.githubusercontent.com/media/zheyishine/vllm_whl/refs/heads/main/vllm-0.8.5.post2.dev28%2Bgd327eed71.cu128-cp310-cp310-linux_x86_64.whl --force-reinstall
45
  ```
46
 
47
  Finally, install compatible versions of PyTorch and Torchvision after vLLM is installed:
48
  ```shell
49
+ pip install transformers==4.51.1
50
  ```
51
 
52
  #### Offline Inference