# download_models.py from transformers import AutoModelForCausalLM, AutoTokenizer from langchain_huggingface import HuggingFaceEmbeddings from config import EMBEDDING_MODEL, LLM_MODEL print("⏳ Downloading Embedding Model...") HuggingFaceEmbeddings(model_name=EMBEDDING_MODEL) print(f"⏳ Downloading LLM: {LLM_MODEL}...") # Direct download to cache AutoTokenizer.from_pretrained(LLM_MODEL) AutoModelForCausalLM.from_pretrained(LLM_MODEL) print("✅ Models cached successfully")