Spaces:
Running
Running
fix
Browse files
app.py
CHANGED
|
@@ -641,18 +641,14 @@ def create_ui():
|
|
| 641 |
# Initialize
|
| 642 |
def initialize():
|
| 643 |
"""Initialize the application"""
|
| 644 |
-
|
|
|
|
|
|
|
|
|
|
| 645 |
load_status()
|
| 646 |
-
|
| 647 |
-
# Check and setup llama.cpp
|
| 648 |
check_llama_cpp()
|
| 649 |
-
|
| 650 |
-
# Schedule regular checks for updates
|
| 651 |
scheduler = BackgroundScheduler()
|
| 652 |
-
scheduler.add_job(check_and_process, 'interval', minutes=60)
|
| 653 |
-
scheduler.start()
|
| 654 |
-
|
| 655 |
-
# Run initial check
|
| 656 |
threading.Thread(target=check_and_process).start()
|
| 657 |
|
| 658 |
if __name__ == "__main__":
|
|
|
|
| 641 |
# Initialize
|
| 642 |
def initialize():
|
| 643 |
"""Initialize the application"""
|
| 644 |
+
global QUANT_CONFIGS
|
| 645 |
+
# Sort configurations by size (smallest first)
|
| 646 |
+
QUANT_CONFIGS = sorted(QUANT_CONFIGS, key=lambda x: x["size_gb"])
|
| 647 |
+
# Rest of the initialization code...
|
| 648 |
load_status()
|
|
|
|
|
|
|
| 649 |
check_llama_cpp()
|
|
|
|
|
|
|
| 650 |
scheduler = BackgroundScheduler()
|
| 651 |
+
scheduler.add_job(check_and_process, 'interval', minutes=60)
|
|
|
|
|
|
|
|
|
|
| 652 |
threading.Thread(target=check_and_process).start()
|
| 653 |
|
| 654 |
if __name__ == "__main__":
|