Update README.md
Browse files
README.md
CHANGED
|
@@ -102,7 +102,7 @@ print(f"Usando device: {device}")
|
|
| 102 |
# ----------------------------
|
| 103 |
# Scarica tokenizer e modello da HF
|
| 104 |
# ----------------------------
|
| 105 |
-
repo_id = "Mattimax/PicoDAC
|
| 106 |
|
| 107 |
tokenizer_path = hf_hub_download(repo_id, "tokenizer.json")
|
| 108 |
model_path = hf_hub_download(repo_id, "model.safetensors")
|
|
@@ -205,14 +205,14 @@ def generate(prompt, max_new_tokens=50, temperature=0.7):
|
|
| 205 |
# ----------------------------
|
| 206 |
# Loop chat
|
| 207 |
# ----------------------------
|
| 208 |
-
print("Chat PicoDAC
|
| 209 |
while True:
|
| 210 |
user_input = input("Tu: ").strip()
|
| 211 |
if user_input.lower() in ["exit", "quit"]:
|
| 212 |
print("Chiusura chat. Ciao!")
|
| 213 |
break
|
| 214 |
response = generate(user_input)
|
| 215 |
-
print(f"PicoDAC
|
| 216 |
```
|
| 217 |
|
| 218 |
**Suggerimenti pratici:**
|
|
|
|
| 102 |
# ----------------------------
|
| 103 |
# Scarica tokenizer e modello da HF
|
| 104 |
# ----------------------------
|
| 105 |
+
repo_id = "Mattimax/PicoDAC"
|
| 106 |
|
| 107 |
tokenizer_path = hf_hub_download(repo_id, "tokenizer.json")
|
| 108 |
model_path = hf_hub_download(repo_id, "model.safetensors")
|
|
|
|
| 205 |
# ----------------------------
|
| 206 |
# Loop chat
|
| 207 |
# ----------------------------
|
| 208 |
+
print("Chat PicoDAC (scrivi 'exit' per uscire)")
|
| 209 |
while True:
|
| 210 |
user_input = input("Tu: ").strip()
|
| 211 |
if user_input.lower() in ["exit", "quit"]:
|
| 212 |
print("Chiusura chat. Ciao!")
|
| 213 |
break
|
| 214 |
response = generate(user_input)
|
| 215 |
+
print(f"PicoDAC: {response}")
|
| 216 |
```
|
| 217 |
|
| 218 |
**Suggerimenti pratici:**
|