File size: 977 Bytes
3e30d53 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 | import requests
def check_ollama():
try:
response = requests.get("http://localhost:11434/")
if response.status_code == 200:
print("✅ Ollama is running.")
else:
print(f"⚠️ Ollama returned status code: {response.status_code}")
# Check for models
response = requests.get("http://localhost:11434/api/tags")
if response.status_code == 200:
models = [m['name'] for m in response.json().get('models', [])]
print(f"Available models: {models}")
if "llama3:latest" in models or "llama3" in models:
print("✅ llama3 model found.")
else:
print("❌ llama3 model NOT found. Please run 'ollama pull llama3'")
else:
print("❌ Could not list models.")
except Exception as e:
print(f"❌ Error connecting to Ollama: {e}")
if __name__ == "__main__":
check_ollama()
|