LH-Tech-AI commited on
Commit
f8d4fb3
·
verified ·
1 Parent(s): 44f99f0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -25,7 +25,7 @@ Then, use the prepare-script and the finetuning script in the files list of this
25
 
26
  # How to use it
27
  You can directly download the final model as ONNX format - so it runs without the need to install a huge Python environment with PyTorch, CUDA, etc... - as INT8 and in full precision.
28
- Use `inference.py` for local inference on CUDA or CPU!
29
 
30
 
31
  Have fun! :D
 
25
 
26
  # How to use it
27
  You can directly download the final model as ONNX format - so it runs without the need to install a huge Python environment with PyTorch, CUDA, etc... - as INT8 and in full precision.
28
+ Use `inference.py` for local inference on CUDA or CPU! First, install `pip install onnxruntime-gpu tiktoken numpy nvidia-cudnn-cu12 nvidia-cublas-cu12` on your system (in a Python VENV for Linux users).
29
 
30
 
31
  Have fun! :D