emsesc commited on
Commit
07805d7
·
1 Parent(s): 8352ce2

switch loading method

Browse files
Files changed (2) hide show
  1. app.py +22 -4
  2. requirements.txt +2 -1
app.py CHANGED
@@ -1,6 +1,8 @@
1
  from dash import Dash, html, dcc, Input, Output, State
2
  import pandas as pd
3
  import dash_mantine_components as dmc
 
 
4
  from graphs.leaderboard import (
5
  create_leaderboard,
6
  get_top_n_leaderboard,
@@ -12,10 +14,26 @@ app = Dash()
12
  server = app.server
13
 
14
  # Load parquet file from Hugging Face
15
- print("Loading data...")
16
- hf_parquet_url = "https://huggingface.co/datasets/emsesc/open_model_evolution_data/resolve/main/filtered_df.parquet"
17
- filtered_df = pd.read_parquet(hf_parquet_url)
18
- print("Data loaded.")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  # List columns for reference
21
  print(filtered_df.columns.tolist())
 
1
  from dash import Dash, html, dcc, Input, Output, State
2
  import pandas as pd
3
  import dash_mantine_components as dmc
4
+ from dataasets import load_dataset
5
+ import time
6
  from graphs.leaderboard import (
7
  create_leaderboard,
8
  get_top_n_leaderboard,
 
14
  server = app.server
15
 
16
  # Load parquet file from Hugging Face
17
+ HF_DATASET_ID = "emsesc/open_model_evolution_data"
18
+ hf_parquet_url = "https://huggingface.co/datasets/emsesc/open_model_evolution_data/resolve/main/"
19
+ data_files = {
20
+ "filtered_df": hf_parquet_url + "filtered_df.parquet",
21
+ # "weekly_df": hf_parquet_url + "weekly_df.parquet",
22
+ }
23
+ filtered_df = pd.DataFrame()
24
+
25
+ print(f"Attempting to load dataset from Hugging Face Hub: {HF_DATASET_ID}")
26
+ try:
27
+ overall_start_time = time.time()
28
+ dataset = load_dataset("parquet", data_files=data_files)
29
+ df = dataset["filtered_df"].to_pandas()
30
+ filtered_df = df.copy()
31
+
32
+ msg = f"Successfully loaded dataset in {time.time() - overall_start_time:.2f}s."
33
+ print(msg)
34
+ except Exception as e:
35
+ err_msg = f"Failed to load dataset. Error: {e}"
36
+ print(err_msg)
37
 
38
  # List columns for reference
39
  print(filtered_df.columns.tolist())
requirements.txt CHANGED
@@ -5,4 +5,5 @@ gunicorn
5
  dash-mantine-components
6
  dash-bootstrap-components
7
  pyarrow
8
- dash-iconify
 
 
5
  dash-mantine-components
6
  dash-bootstrap-components
7
  pyarrow
8
+ dash-iconify
9
+ datasets