Upload folder using huggingface_hub
Browse files- Dockerfile +1 -1
- app.py +1 -1
Dockerfile
CHANGED
|
@@ -7,7 +7,7 @@ WORKDIR /app
|
|
| 7 |
COPY . .
|
| 8 |
# Install dependencies from the requirements file without using cache to reduce image size
|
| 9 |
RUN pip install --no-cache-dir --upgrade -r requirements.txt
|
| 10 |
-
|
| 11 |
# Define the command to start the application using Gunicorn with 4 worker processes
|
| 12 |
# - `-w 4`: Uses 4 worker processes for handling requests
|
| 13 |
# - `-b 0.0.0.0:7860`: Binds the server to port 7860 on all network interfaces
|
|
|
|
| 7 |
COPY . .
|
| 8 |
# Install dependencies from the requirements file without using cache to reduce image size
|
| 9 |
RUN pip install --no-cache-dir --upgrade -r requirements.txt
|
| 10 |
+
#
|
| 11 |
# Define the command to start the application using Gunicorn with 4 worker processes
|
| 12 |
# - `-w 4`: Uses 4 worker processes for handling requests
|
| 13 |
# - `-b 0.0.0.0:7860`: Binds the server to port 7860 on all network interfaces
|
app.py
CHANGED
|
@@ -27,7 +27,7 @@ def predict_price():
|
|
| 27 |
prediction = loaded_model.predict(input_data)
|
| 28 |
# Return the prediction as a JSON response
|
| 29 |
return jsonify({'Price': prediction})
|
| 30 |
-
|
| 31 |
if __name__ == '__main__':
|
| 32 |
# Run the app on all available interfaces on port 5000
|
| 33 |
app.run(debug=True, host='0.0.0.0', port=5000)
|
|
|
|
| 27 |
prediction = loaded_model.predict(input_data)
|
| 28 |
# Return the prediction as a JSON response
|
| 29 |
return jsonify({'Price': prediction})
|
| 30 |
+
##
|
| 31 |
if __name__ == '__main__':
|
| 32 |
# Run the app on all available interfaces on port 5000
|
| 33 |
app.run(debug=True, host='0.0.0.0', port=5000)
|