Polushinm commited on
Commit
ac63253
·
verified ·
1 Parent(s): 9b738bb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -1
README.md CHANGED
@@ -4,7 +4,7 @@ language:
4
  - ru
5
  - en
6
  pipeline_tag: text-generation
7
- license: other
8
  license_name: apache-2.0
9
  license_link: https://huggingface.co/MTSAIR/Kodify-Nano/blob/main/Apache%20License%20MTS%20AI.docx
10
  ---
@@ -22,6 +22,13 @@ Kodify-Nano is a lightweight LLM designed for code development tasks with minima
22
  ```bash
23
  python3 -m vllm.entrypoints.openai.api_server --model MTSAIR/Kodify-Nano --port 8985
24
  ```
 
 
 
 
 
 
 
25
  ---
26
 
27
  ## Using the Ollama Image
 
4
  - ru
5
  - en
6
  pipeline_tag: text-generation
7
+ license: apache-2.0
8
  license_name: apache-2.0
9
  license_link: https://huggingface.co/MTSAIR/Kodify-Nano/blob/main/Apache%20License%20MTS%20AI.docx
10
  ---
 
22
  ```bash
23
  python3 -m vllm.entrypoints.openai.api_server --model MTSAIR/Kodify-Nano --port 8985
24
  ```
25
+
26
+ > **Important!** If you encounter the **"CUDA out of memory. Tried to allocate..."** error despite having sufficient GPU memory, try one of these solutions:
27
+ > 1. Add the --enforce-eager argument
28
+ > 2. Reduce GPU memory utilization (for example --gpu-memory-utilization 0.8)
29
+ >
30
+ > Note: This may decrease model performance.
31
+
32
  ---
33
 
34
  ## Using the Ollama Image