Spaces:
Runtime error
Runtime error
Commit
·
111d85e
1
Parent(s):
634acb1
fix
Browse files
app.py
CHANGED
|
@@ -1,6 +1,8 @@
|
|
| 1 |
import os
|
| 2 |
import time
|
| 3 |
import traceback
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from config_store import (
|
| 6 |
get_process_config,
|
|
@@ -44,14 +46,16 @@ MODELS_TO_TASKS = {
|
|
| 44 |
TASKS = set(TASKS_TO_OVMODEL.keys()) & set(TASKS_TO_MODEL_LOADERS.keys())
|
| 45 |
|
| 46 |
|
| 47 |
-
def run_benchmark(kwargs, oauth_token: gr.OAuthToken):
|
| 48 |
-
if oauth_token.token is None:
|
| 49 |
gr.Error("Please login to be able to run the benchmark.")
|
| 50 |
return tuple(None for _ in BACKENDS)
|
| 51 |
|
|
|
|
|
|
|
| 52 |
timestamp = time.strftime("%Y-%m-%d-%H-%M-%S")
|
| 53 |
-
|
| 54 |
-
repo_id = f"{
|
| 55 |
token = oauth_token.token
|
| 56 |
|
| 57 |
create_repo(repo_id, token=token, repo_type="dataset", exist_ok=True)
|
|
@@ -124,7 +128,7 @@ def run_benchmark(kwargs, oauth_token: gr.OAuthToken):
|
|
| 124 |
benchmark.push_to_hub(
|
| 125 |
repo_id=repo_id, subfolder=benchmark_name, token=oauth_token.token
|
| 126 |
)
|
| 127 |
-
gr.Info(f"Pushed benchmark to {
|
| 128 |
|
| 129 |
outputs[backend] = f"\n{benchmark_report.to_markdown_text()}"
|
| 130 |
|
|
@@ -137,11 +141,13 @@ def run_benchmark(kwargs, oauth_token: gr.OAuthToken):
|
|
| 137 |
|
| 138 |
yield tuple(outputs[b] for b in BACKENDS)
|
| 139 |
|
|
|
|
|
|
|
| 140 |
|
| 141 |
def build_demo():
|
| 142 |
with gr.Blocks() as demo:
|
| 143 |
# add login button
|
| 144 |
-
gr.LoginButton(
|
| 145 |
|
| 146 |
# add image
|
| 147 |
gr.HTML(
|
|
@@ -237,11 +243,8 @@ def build_demo():
|
|
| 237 |
return demo
|
| 238 |
|
| 239 |
|
| 240 |
-
demo = build_demo()
|
| 241 |
-
|
| 242 |
if __name__ == "__main__":
|
| 243 |
os.environ["LOG_TO_FILE"] = "0"
|
| 244 |
os.environ["LOG_LEVEL"] = "INFO"
|
| 245 |
-
setup_logging(level="INFO", prefix="MAIN-PROCESS")
|
| 246 |
|
| 247 |
-
|
|
|
|
| 1 |
import os
|
| 2 |
import time
|
| 3 |
import traceback
|
| 4 |
+
import logging
|
| 5 |
+
from typing import Optional
|
| 6 |
|
| 7 |
from config_store import (
|
| 8 |
get_process_config,
|
|
|
|
| 46 |
TASKS = set(TASKS_TO_OVMODEL.keys()) & set(TASKS_TO_MODEL_LOADERS.keys())
|
| 47 |
|
| 48 |
|
| 49 |
+
def run_benchmark(kwargs, oauth_token: Optional[gr.OAuthToken] = None):
|
| 50 |
+
if oauth_token.token is None or oauth_token.token == "":
|
| 51 |
gr.Error("Please login to be able to run the benchmark.")
|
| 52 |
return tuple(None for _ in BACKENDS)
|
| 53 |
|
| 54 |
+
setup_logging(level="INFO", prefix="MAIN-PROCESS")
|
| 55 |
+
|
| 56 |
timestamp = time.strftime("%Y-%m-%d-%H-%M-%S")
|
| 57 |
+
name = whoami(oauth_token.token)["name"]
|
| 58 |
+
repo_id = f"{name}/benchmarks"
|
| 59 |
token = oauth_token.token
|
| 60 |
|
| 61 |
create_repo(repo_id, token=token, repo_type="dataset", exist_ok=True)
|
|
|
|
| 128 |
benchmark.push_to_hub(
|
| 129 |
repo_id=repo_id, subfolder=benchmark_name, token=oauth_token.token
|
| 130 |
)
|
| 131 |
+
gr.Info(f"Pushed benchmark to {repo_id}/{benchmark_name}")
|
| 132 |
|
| 133 |
outputs[backend] = f"\n{benchmark_report.to_markdown_text()}"
|
| 134 |
|
|
|
|
| 141 |
|
| 142 |
yield tuple(outputs[b] for b in BACKENDS)
|
| 143 |
|
| 144 |
+
logging.getLogger().setLevel(logging.NOTSET)
|
| 145 |
+
|
| 146 |
|
| 147 |
def build_demo():
|
| 148 |
with gr.Blocks() as demo:
|
| 149 |
# add login button
|
| 150 |
+
gr.LoginButton()
|
| 151 |
|
| 152 |
# add image
|
| 153 |
gr.HTML(
|
|
|
|
| 243 |
return demo
|
| 244 |
|
| 245 |
|
|
|
|
|
|
|
| 246 |
if __name__ == "__main__":
|
| 247 |
os.environ["LOG_TO_FILE"] = "0"
|
| 248 |
os.environ["LOG_LEVEL"] = "INFO"
|
|
|
|
| 249 |
|
| 250 |
+
build_demo().queue(max_size=10).launch()
|