Spaces:
Running
Running
gemma
Browse files
app.py
CHANGED
|
@@ -12,7 +12,7 @@ logging.basicConfig(level=logging.INFO)
|
|
| 12 |
logger = logging.getLogger(__name__)
|
| 13 |
|
| 14 |
app = FastAPI()
|
| 15 |
-
MODEL_NAME = '
|
| 16 |
|
| 17 |
@lru_cache()
|
| 18 |
def get_llm():
|
|
|
|
| 12 |
logger = logging.getLogger(__name__)
|
| 13 |
|
| 14 |
app = FastAPI()
|
| 15 |
+
MODEL_NAME = 'gemma2:2b'
|
| 16 |
|
| 17 |
@lru_cache()
|
| 18 |
def get_llm():
|
start.sh
CHANGED
|
@@ -9,8 +9,8 @@ export CUDA_VISIBLE_DEVICES=0 # Use the first GPU if available
|
|
| 9 |
ollama serve & # Use GPU 0 if available
|
| 10 |
|
| 11 |
# Pull the model if not already present
|
| 12 |
-
if ! ollama list | grep -q "
|
| 13 |
-
ollama pull
|
| 14 |
fi
|
| 15 |
|
| 16 |
# Wait for Ollama to start up (use a more robust check)
|
|
|
|
| 9 |
ollama serve & # Use GPU 0 if available
|
| 10 |
|
| 11 |
# Pull the model if not already present
|
| 12 |
+
if ! ollama list | grep -q "gemma2:2b"; then
|
| 13 |
+
ollama pull gemma2:2b
|
| 14 |
fi
|
| 15 |
|
| 16 |
# Wait for Ollama to start up (use a more robust check)
|