| from langchain_community.llms import Ollama | |
| from langchain_community.embeddings import HuggingFaceEmbeddings | |
| from llama_index.core import Settings | |
| # Create an Ollama instance with the model configuration | |
| llm = Ollama(model='tinyllama') | |
| # Use the llm instance directly where needed | |
| Settings.llm = llm | |
| # Configure the embedding model for your settings | |
| Settings.embed_model = HuggingFaceEmbeddings( | |
| model_name="BAAI/bge-small-en-v1.5" | |
| ) | |