import requests def check_ollama(): try: response = requests.get("http://localhost:11434/") if response.status_code == 200: print("✅ Ollama is running.") else: print(f"⚠️ Ollama returned status code: {response.status_code}") # Check for models response = requests.get("http://localhost:11434/api/tags") if response.status_code == 200: models = [m['name'] for m in response.json().get('models', [])] print(f"Available models: {models}") if "llama3:latest" in models or "llama3" in models: print("✅ llama3 model found.") else: print("❌ llama3 model NOT found. Please run 'ollama pull llama3'") else: print("❌ Could not list models.") except Exception as e: print(f"❌ Error connecting to Ollama: {e}") if __name__ == "__main__": check_ollama()