Kodify-Nano
Collection
Kodify-Nano is a lightweight LLM designed for code development tasks with minimal resource usage. It is optimized for fast and efficient interaction, • 3 items • Updated • 3
Kodify-Nano-GGUF - GGUF версия модели MTSAIR/Kodify-Nano, оптимизированная для CPU/GPU-инференса и использованием Ollama/llama.cpp. Легковесная LLM для задач разработки кода с минимальными ресурсами.
Kodify-Nano-GGUF - GGUF version of MTSAIR/Kodify-Nano, optimized for CPU/GPU inference with Ollama/llama.cpp. Lightweight LLM for code development tasks with minimal resource requirements.
You can run Kodify Nano on OLLAMA in two ways:
docker run -e OLLAMA_HOST=0.0.0.0:8985 -p 8985:8985 --name ollama -d ollama/ollama
docker run --runtime nvidia -e OLLAMA_HOST=0.0.0.0:8985 -p 8985:8985 --name ollama -d ollama/ollama
Important:
- Ensure Docker is installed and running
- If port 8985 is occupied, replace it with any available port and update plugin configuration
docker exec ollama ollama pull hf.co/MTSAIR/Kodify-Nano-GGUF
docker exec ollama ollama cp hf.co/MTSAIR/Kodify-Nano-GGUF kodify_nano
docker exec ollama ollama run kodify_nano
Download OLLAMA:
https://ollama.com/download
Set the port:
export OLLAMA_HOST=0.0.0.0:8985
Note: If port 8985 is occupied, replace it and update plugin configuration
ollama serve &
ollama pull hf.co/MTSAIR/Kodify-Nano-GGUF
ollama cp hf.co/MTSAIR/Kodify-Nano-GGUF kodify_nano
ollama run kodify_nano
If you changed the Docker port from 8985, update the plugin's config.json:
Ctrl+L (Cmd+L on Mac).Ctrl+J (Cmd+J on Mac).config.json file:apiBase port under tabAutocompleteModel and models.Ctrl+S or File > Save).Download using huggingface_hub:
pip install huggingface-hub
python -c "from huggingface_hub import hf_hub_download; hf_hub_download(repo_id='MTSAIR/Kodify-Nano-GGUF', filename='Kodify_Nano_q4_k_s.gguf', local_dir='./models')"
Install Ollama Python library:
pip install ollama
Example code:
import ollama
response = ollama.generate(
model="kodify-nano",
prompt="Write a Python function to calculate factorial",
options={
"temperature": 0.4,
"top_p": 0.8,
"num_ctx": 8192
}
)
print(response['response'])
response = ollama.generate(
model="kodify-nano",
prompt="""<s>[INST]
Write a Python function that:
1. Accepts a list of numbers
2. Returns the median value
[/INST]""",
options={"max_tokens": 512}
)
### Code Refactoring
response = ollama.generate(
model="kodify-nano",
prompt="""<s>[INST]
Refactor this Python code:
def calc(a,b):
s = a + b
d = a - b
p = a * b
return s, d, p
[/INST]""",
options={"temperature": 0.3}
)