| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| model_name = "TinyLlama/TinyLlama-1.1B-Chat-v1.0" | |
| model = AutoModelForCausalLM.from_pretrained(model_name) | |
| tokenizer = AutoTokenizer.from_pretrained(model_name) | |
| # Сохраняем модель и токенизатор локально | |
| model.save_pretrained("./tinyllama") | |
| tokenizer.save_pretrained("./tinyllama") |