File size: 12,139 Bytes
4974a11
411153e
 
 
 
 
 
 
 
 
ed7d966
411153e
 
 
1f2bf05
411153e
 
1f2bf05
ed7d966
c4fcf32
 
 
411153e
c4fcf32
f433418
1f2bf05
 
411153e
 
9d82914
 
 
1f2bf05
9d82914
411153e
 
9d82914
 
411153e
9d82914
 
 
411153e
 
9d82914
 
411153e
c4fcf32
ed7d966
 
 
 
411153e
ed7d966
411153e
 
ed7d966
 
 
 
 
1f2bf05
ed7d966
 
 
 
 
1f2bf05
ed7d966
 
 
 
 
 
02a80e2
411153e
 
 
1f2bf05
741d530
411153e
 
 
 
 
 
 
 
 
 
4974a11
460fe63
1f2bf05
 
 
 
 
 
 
 
 
 
 
 
 
 
ed7d966
1f2bf05
 
 
 
 
 
ed7d966
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1f2bf05
411153e
 
1f2bf05
 
 
ed7d966
 
 
411153e
ed7d966
4974a11
411153e
1f2bf05
411153e
 
 
 
 
1f2bf05
411153e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1f2bf05
411153e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1f2bf05
411153e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1f2bf05
411153e
 
 
 
1f2bf05
411153e
 
4974a11
9c61038
 
 
 
 
 
 
 
 
 
e31ac6d
9c61038
741d530
9c61038
e31ac6d
9c61038
 
 
e31ac6d
411153e
3e0c8fa
4974a11
460fe63
411153e
c4fcf32
e31ac6d
 
 
c4fcf32
ed7d966
 
c4fcf32
ed7d966
 
e31ac6d
 
 
c4fcf32
ed7d966
 
411153e
 
4974a11
411153e
741d530
4974a11
411153e
 
 
 
4974a11
411153e
ed7d966
 
 
 
411153e
 
 
 
 
 
 
 
 
 
 
 
 
4974a11
 
ed7d966
 
1f2bf05
ed7d966
c4fcf32
ed7d966
 
 
 
 
2567a65
 
e31ac6d
ed7d966
 
 
 
 
460fe63
ed7d966
 
 
 
 
 
2567a65
 
e31ac6d
ed7d966
 
 
 
 
144f1bb
ed7d966
 
 
 
 
 
2567a65
 
e31ac6d
ed7d966
 
 
 
 
 
e31ac6d
ed7d966
1f2bf05
ed7d966
 
e31ac6d
460fe63
ed7d966
 
 
1f2bf05
ed7d966
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2567a65
 
460fe63
ed7d966
 
 
 
 
 
02a80e2
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
import streamlit as st
import pandas as pd
import plotly.graph_objects as go
import plotly.express as px
from io import StringIO
import json

# Page config
st.set_page_config(
    page_title="NaviTrace Leaderboard",
    layout="centered",
    initial_sidebar_state="collapsed"
)

# Custom CSS for Nerfies-style design
st.markdown("""
<style>
    /* Import Font Awesome */
    @import url('https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.4.0/css/all.min.css');

    /* Headings */
    h1 {
        text-align: center;
        font-size: 4.5rem !important;
        font-weight: 500;
        margin-top: 1rem;
        margin-bottom: 2rem;
    }
    
    /* Links container */
    .links-container {
        text-align: center;
        margin-bottom: 3rem;
        font-size: 1.1rem;
    }
    
    .links-container a {
        margin: 0 1rem;
        text-decoration: none;
        color: #667eea;
        font-weight: 600;
        transition: color 0.3s;
    }
    
    .links-container a:hover {
        color: #764ba2;
    }
        
    /* Instructions styling */
    .instruction-item {
        display: flex;
        gap: 1.5rem;
        margin: 2rem 0;
        align-items: flex-start;
    }
    
    .instruction-number {
        flex-shrink: 0;
        width: 40px;
        height: 40px;
        border-radius: 50%;
        background: linear-gradient(135deg, #667eea 0%, #764ba2 100%);
        color: white;
        display: flex;
        align-items: center;
        justify-content: center;
        font-weight: 700;
        font-size: 1.2rem;
    }
    
    .instruction-content {
        flex-grow: 1;
        padding-top: 0.3rem;
    }

</style>
""", unsafe_allow_html=True)

# Sample data - Replace with your actual data
def load_data():
    return pd.DataFrame({
        'Model': ['GPT-4', 'Claude-3.5-Sonnet', 'Gemini-Pro', 'Llama-3-70B', 'Mistral-Large'],
        'Total Score': [87.5, 85.2, 82.1, 78.3, 75.6],
        'Embodiment-A': [90.2, 87.5, 84.3, 80.1, 77.8],
        'Embodiment-B': [85.8, 84.1, 81.2, 77.9, 74.5],
        'Embodiment-C': [86.5, 84.0, 80.8, 76.9, 74.5],
        'Category-Spatial': [88.9, 86.7, 83.5, 79.8, 76.9],
        'Category-Temporal': [86.3, 84.2, 81.0, 77.5, 75.1],
        'Category-Object': [87.3, 84.7, 81.8, 77.6, 74.8],
    })

def calculate_score(results_df):
    """
    Calculate score using private test split ground truth.
    This function should:
    1. Load the private test split ground truth (not exposed to users)
    2. Compare uploaded predictions with ground truth
    3. Calculate metrics per embodiment and category
    4. Return detailed scores
    
    Args:
        results_df: DataFrame with columns ['sample_id', 'prediction', ...]
    
    Returns:
        dict: Scores breakdown or None if error
    """
    try:
        # TODO: Implement your scoring logic here
        # Example structure:
        # ground_truth = load_private_test_split()  # From secure location
        # scores = evaluate_predictions(results_df, ground_truth)
        
        # Placeholder - replace with actual calculation
        scores = {
            'Total Score': 85.0,
            'Embodiment-A': 87.0,
            'Embodiment-B': 84.0,
            'Embodiment-C': 84.0,
            'Category-Spatial': 86.0,
            'Category-Temporal': 85.0,
            'Category-Object': 84.0,
        }
        return scores
    except Exception as e:
        st.error(f"Error calculating score: {str(e)}")
        return None

def validate_tsv_format(uploaded_file):
    """Validate that the uploaded TSV has the correct format"""
    try:
        df = pd.read_csv(uploaded_file, sep='\t')
        # TODO: Add your specific validation logic
        # Check for required columns, data types, etc.
        required_cols = ['sample_id', 'prediction']  # Adjust as needed
        if not all(col in df.columns for col in required_cols):
            return False, f"Missing required columns. Expected: {required_cols}"
        return True, df
    except Exception as e:
        return False, f"Error reading file: {str(e)}"

def create_bar_chart(df, view_type):
    """Create interactive bar chart based on view type"""
    if view_type == "Total Score":
        fig = go.Figure(data=[
            go.Bar(
                x=df['Model'],
                y=df['Total Score'],
                marker_color=px.colors.sequential.Purples_r,
                text=df['Total Score'].round(1),
                textposition='outside',
            )
        ])
        fig.update_layout(
            title="Model Performance - Total Score",
            xaxis_title="Model",
            yaxis_title="Score",
            yaxis_range=[0, 100],
            height=500,
        )
    
    elif view_type == "Per Embodiment":
        embodiment_cols = [col for col in df.columns if col.startswith('Embodiment-')]
        fig = go.Figure()
        for col in embodiment_cols:
            fig.add_trace(go.Bar(
                name=col.replace('Embodiment-', ''),
                x=df['Model'],
                y=df[col],
                text=df[col].round(1),
                textposition='outside',
            ))
        fig.update_layout(
            title="Model Performance - Per Embodiment",
            xaxis_title="Model",
            yaxis_title="Score",
            yaxis_range=[0, 100],
            barmode='group',
            height=500,
        )
    
    else:  # Per Category
        category_cols = [col for col in df.columns if col.startswith('Category-')]
        fig = go.Figure()
        for col in category_cols:
            fig.add_trace(go.Bar(
                name=col.replace('Category-', ''),
                x=df['Model'],
                y=df[col],
                text=df[col].round(1),
                textposition='outside',
            ))
        fig.update_layout(
            title="Model Performance - Per Category",
            xaxis_title="Model",
            yaxis_title="Score",
            yaxis_range=[0, 100],
            barmode='group',
            height=500,
        )
    
    # Common styling
    fig.update_layout(
        plot_bgcolor='rgba(0,0,0,0)',
        paper_bgcolor='rgba(0,0,0,0)',
        font=dict(size=12),
        showlegend=(view_type != "Total Score"),
        margin=dict(t=80, b=60, l=60, r=60),
    )
    fig.update_xaxes(showgrid=False)
    fig.update_yaxes(showgrid=True, gridcolor='lightgray', gridwidth=0.5)
    
    return fig

# TODO remove # Serve only the chart as JSON if parameter "only_chart" is set
# # E.g. https://huggingface.co/spaces/leggedrobotics/navitrace_leaderboard/?only_chart=total_score
# params = st.query_params
# if "only_chart" in params and params["only_chart"] in ["total_score", "per_embodiment", "per_category"]:
#     if params["only_chart"] == "total_score":
#         view_type = "Total Score"
#     elif params["only_chart"] == "per_embodiment":
#         view_type = "Per Embodiment"
#     elif params["only_chart"] == "per_category":
#         view_type = "Per Category"

#     # Create chart
#     df = load_data()
#     fig = create_bar_chart(df, view_type)

#     # Only output JSON
#     st.write(fig.to_json())
#     st.stop()

# Main content
st.title("NaviTrace Leaderboard")

# Links
st.markdown("""
<div class="links-container">
    <a href="https://leggedrobotics.github.io/navitrace_webpage/" target="_blank">
        <i class="fas fa-house"></i> Project
    </a>
    <a href="https://your-paper-website.com" target="_blank">
        <i class="fas fa-file-pdf"></i> Paper
    </a>
    <a href="https://github.com/your-username/navitrace" target="_blank">
        <i class="fab fa-github"></i> Code
    </a>
    <a href="https://huggingface.co/datasets/your-username/navitrace" target="_blank">
        <i class="fas fa-database"></i> Dataset
    </a>
    <a href="https://your-demo-link.com" target="_blank">
        <i class="far fa-images"></i> Demo
    </a>
</div>
""", unsafe_allow_html=True)

# Load data
df = load_data()

# Add user's model if it exists in session state
if 'user_results' in st.session_state:
    user_row = pd.DataFrame([st.session_state.user_results])
    df = pd.concat([user_row, df], ignore_index=True)

# View selector
view_type = st.selectbox(
    "Select View",
    ["Total Score", "Per Embodiment", "Per Category"],
)

# Display chart
fig = create_bar_chart(df, view_type)
st.plotly_chart(fig, use_container_width=True, config={
    'displayModeBar': True,
    'displaylogo': False,
    'toImageButtonOptions': {
        'format': 'png',
        'filename': 'navitrace_leaderboard',
        'height': 600,
        'width': 1200,
        'scale': 2
    }
})

# Detailed table
with st.expander("View Detailed Scores"):
    st.dataframe(df.style.background_gradient(cmap='Purples', subset=df.columns[1:]), use_container_width=True)

with st.expander("How to Test Your Model", expanded=True):
    # Step 1
    st.markdown("""
    <div class="instruction-item">
        <div class="instruction-number">1</div>
        <div class="instruction-content">
            <div><b>Run Evaluation</b></div>
            <div>
                Download and run our evaluation notebook adjusted to your model. The notebook will generate a TSV file with your model's predictions on the test set.
            </div>
        </div>
    </div>
    """, unsafe_allow_html=True)
    
    st.link_button("📓 Open Evaluation Notebook", "https://colab.research.google.com/your-notebook-link", use_container_width=True)
    
    # Step 2
    st.markdown("""
    <div class="instruction-item">
        <div class="instruction-number">2</div>
        <div class="instruction-content">
            <div><b>Upload Results</b></div>
            <div>
                Upload the TSV file generated by the evaluation notebook.
            </div>
        </div>
    </div>
    """, unsafe_allow_html=True)
    
    uploaded_file = st.file_uploader("Upload your TSV file with results", type=['tsv', 'txt'], label_visibility="collapsed")
    
    # Step 3
    st.markdown("""
    <div class="instruction-item">
        <div class="instruction-number">3</div>
        <div class="instruction-content">
            <div><b>Calculate Score</b></div>
            <div>
                Click the button below to evaluate your predictions. Scores are calculated using hidden test set ground-truths.
            </div>
        </div>
    </div>
    """, unsafe_allow_html=True)
    
    if uploaded_file is not None:
        if st.button("🧮 Calculate Score", use_container_width=True):
            with st.spinner("Validating and calculating scores..."):
                # Validate format
                is_valid, result = validate_tsv_format(uploaded_file)
                if is_valid:
                    # Calculate score using hidden ground-truth
                    scores = calculate_score(result)
                    if scores is not None:
                        st.success(f"✅ Score calculated successfully: **{scores['Total Score']:.1f}**")
                        
                        # Store in session state
                        st.session_state.user_results = {
                            'Model': 'Your Model',
                            **scores
                        }
                        st.info("👆 Scroll up to see your model on the leaderboard!")
                        st.rerun()
                else:
                    st.error(f"❌ Invalid file format: {result}")
    else:
        st.info("👆 Upload a TSV file to calculate your score")
    
    # Step 4
    st.markdown("""
    <div class="instruction-item">
        <div class="instruction-number">4</div>
        <div class="instruction-content">
            <div><b>Submit to Official Leaderboard</b></div>
            <div>
                Happy with your score? Submit your model to appear on the official leaderboard.
                Fill out the form below with your model details and results.
            </div>
        </div>
    </div>
    """, unsafe_allow_html=True)
    
    st.link_button("🗳️ Submit Model", "https://forms.gle/your-google-form-link", use_container_width=True)