ln2697 commited on
Commit
e4a312e
·
1 Parent(s): 42e0a50
Files changed (2) hide show
  1. app.py +17 -9
  2. results.json +569 -45
app.py CHANGED
@@ -2,26 +2,34 @@ import json
2
  import gradio as gr
3
 
4
  with open("results.json") as f:
5
- data = json.load(f)
6
 
7
- data = sorted(data, key=lambda x: x["driving_score"], reverse=True)
 
 
 
 
8
 
9
  rows_html = ""
10
  for idx, d in enumerate(data, 1):
11
- paper = f'<a href="{d["paper_link"]}" target="_blank">📄 {d["paper_title"]}</a>' if d["paper_link"] else ""
 
 
 
12
 
13
- # Handle repository - check for null, "-", or empty string
14
- repo_link = d.get("repository_link")
15
- if repo_link and repo_link != "-" and repo_link != "":
16
- repo = f'<a href="{repo_link}" target="_blank"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 16 16" fill="currentColor"><path d="M8 0C3.58 0 0 3.58 0 8c0 3.54 2.29 6.53 5.47 7.59.4.07.55-.17.55-.38 0-.19-.01-.82-.01-1.49-2.01.37-2.53-.49-2.69-.94-.09-.23-.48-.94-.82-1.13-.28-.15-.68-.52-.01-.53.63-.01 1.08.58 1.23.82.72 1.21 1.87.87 2.33.66.07-.52.28-.87.51-1.07-1.78-.2-3.64-.89-3.64-3.95 0-.87.31-1.59.82-2.15-.08-.2-.36-1.02.08-2.12 0 0 .67-.21 2.2.82.64-.18 1.32-.27 2-.27.68 0 1.36.09 2 .27 1.53-1.04 2.2-.82 2.2-.82.44 1.1.16 1.92.08 2.12.51.56.82 1.27.82 2.15 0 3.07-1.87 3.75-3.65 3.95.29.25.54.73.54 1.48 0 1.07-.01 1.93-.01 2.2 0 .21.15.46.55.38A8.013 8.013 0 0016 8c0-4.42-3.58-8-8-8z"></path></svg></a>'
 
17
  else:
18
  repo = "-"
19
 
20
  row_class = "row-even" if idx % 2 == 0 else "row-odd"
21
  rows_html += f"""
22
  <tr class="{row_class}">
23
- <td class="model-name">{d["name"]}</td>
24
- <td class="score">{d["driving_score"]}</td>
25
  <td class="paper-cell">{paper}</td>
26
  <td class="repo-cell">{repo}</td>
27
  </tr>
 
2
  import gradio as gr
3
 
4
  with open("results.json") as f:
5
+ raw_data = json.load(f)
6
 
7
+ # Extract rows from the new format
8
+ data = raw_data["datasets"][0]["sota"]["rows"]
9
+
10
+ # Sort by driving score
11
+ data = sorted(data, key=lambda x: float(x["metrics"]["Driving Score"]), reverse=True)
12
 
13
  rows_html = ""
14
  for idx, d in enumerate(data, 1):
15
+ # Extract paper info
16
+ paper_title = d.get("paper_title", "")
17
+ paper_url = d.get("paper_url", "")
18
+ paper = f'<a href="{paper_url}" target="_blank">📄 {paper_title}</a>' if paper_url and paper_title else ""
19
 
20
+ # Extract repository info
21
+ code_links = d.get("code_links", [])
22
+ if code_links and len(code_links) > 0:
23
+ repo_url = code_links[0]["url"]
24
+ repo = f'<a href="{repo_url}" target="_blank"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 16 16" fill="currentColor"><path d="M8 0C3.58 0 0 3.58 0 8c0 3.54 2.29 6.53 5.47 7.59.4.07.55-.17.55-.38 0-.19-.01-.82-.01-1.49-2.01.37-2.53-.49-2.69-.94-.09-.23-.48-.94-.82-1.13-.28-.15-.68-.52-.01-.53.63-.01 1.08.58 1.23.82.72 1.21 1.87.87 2.33.66.07-.52.28-.87.51-1.07-1.78-.2-3.64-.89-3.64-3.95 0-.87.31-1.59.82-2.15-.08-.2-.36-1.02.08-2.12 0 0 .67-.21 2.2.82.64-.18 1.32-.27 2-.27.68 0 1.36.09 2 .27 1.53-1.04 2.2-.82 2.2-.82.44 1.1.16 1.92.08 2.12.51.56.82 1.27.82 2.15 0 3.07-1.87 3.75-3.65 3.95.29.25.54.73.54 1.48 0 1.07-.01 1.93-.01 2.2 0 .21.15.46.55.38A8.013 8.013 0 0016 8c0-4.42-3.58-8-8-8z"></path></svg></a>'
25
  else:
26
  repo = "-"
27
 
28
  row_class = "row-even" if idx % 2 == 0 else "row-odd"
29
  rows_html += f"""
30
  <tr class="{row_class}">
31
+ <td class="model-name">{d["model_name"]}</td>
32
+ <td class="score">{d["metrics"]["Driving Score"]}</td>
33
  <td class="paper-cell">{paper}</td>
34
  <td class="repo-cell">{repo}</td>
35
  </tr>
results.json CHANGED
@@ -1,45 +1,569 @@
1
- [
2
- {
3
- "name": "HiP-AD",
4
- "driving_score": 86.77,
5
- "paper_title": "HiP-AD: Hierarchical and Multi-Granularity Planning with Deformable Attention for Autonomous Driving in a Single Decoder",
6
- "paper_link": "https://arxiv.org/abs/2503.08612",
7
- "repository": "nullmax-vision/HiP-AD",
8
- "repository_link": "https://github.com/nullmax-vision/HiP-AD"
9
- },
10
- {
11
- "name": "BridgeDrive",
12
- "driving_score": 86.87,
13
- "paper_title": "BridgeDrive: Diffusion Bridge Policy for Closed-Loop Trajectory Planning in Autonomous Driving",
14
- "paper_link": "https://arxiv.org/abs/2509.23589",
15
- "repository_link": null
16
- },
17
- {
18
- "name": "SimLingo",
19
- "driving_score": 85.94 ,
20
- "paper_title": "SimLingo: Vision-Only Closed-Loop Autonomous Driving with Language-Action Alignment",
21
- "paper_link": "https://arxiv.org/abs/2503.09594",
22
- "repository_link": "https://github.com/RenzKa/simlingo"
23
- },
24
- {
25
- "name": "TransFuser++",
26
- "driving_score": 84.21 ,
27
- "paper_title": "Hidden Biases of End-to-End Driving Datasets",
28
- "paper_link": "https://arxiv.org/abs/2412.09602",
29
- "repository_link": "https://github.com/autonomousvision/carla_garage"
30
- },
31
- {
32
- "name": "AutoVLA",
33
- "driving_score": 78.84,
34
- "paper_title": "AutoVLA: A Vision-Language-Action Model for End-to-End Autonomous Driving with Adaptive Reasoning and Reinforcement Fine-Tuning",
35
- "paper_link": "https://arxiv.org/abs/2506.13757",
36
- "repository_link": "https://github.com/ucla-mobility/AutoVLA"
37
- },
38
- {
39
- "name": "ORION",
40
- "driving_score": 77.74 ,
41
- "paper_title": "ORION: A Holistic End-to-End Autonomous Driving Framework by Vision-Language Instructed Action Generation",
42
- "paper_link": "https://arxiv.org/abs/2503.19755",
43
- "repository_link": "https://github.com/xiaomi-mlab/Orion"
44
- }
45
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "task": "Bench2Drive",
3
+ "categories": [],
4
+ "description": "Bench2Drive is an autonomous driving benchmark based on the CARLA leaderboard 2.0. It consists of 220 short routes featuring safety critical scenarios. The evaluation is performed closed-loop in the CARLA simulator. The performance of an entire driving stack is being evaluated.",
5
+ "subtasks": [],
6
+ "synonyms": [],
7
+ "source_link": null,
8
+ "datasets": [
9
+ {
10
+ "sota": {
11
+ "rows": [
12
+ {
13
+ "code_links": [
14
+ {
15
+ "url": "https://github.com/nullmax-vision/hip-ad",
16
+ "title": "nullmax-vision/hip-ad"
17
+ }
18
+ ],
19
+ "model_links": [],
20
+ "model_name": "HiP-AD",
21
+ "metrics": {
22
+ "Driving Score": "86.77"
23
+ },
24
+ "paper_date": "2025-03-11",
25
+ "paper_title": "HiP-AD: Hierarchical and Multi-Granularity Planning with Deformable Attention for Autonomous Driving in a Single Decoder",
26
+ "uses_additional_data": false,
27
+ "paper_url": "https://arxiv.org/abs/2503.08612v1"
28
+ },
29
+ {
30
+ "code_links": [],
31
+ "model_links": [],
32
+ "model_name": "R2SE",
33
+ "metrics": {
34
+ "Driving Score": "86.28"
35
+ },
36
+ "paper_date": null,
37
+ "paper_title": "",
38
+ "uses_additional_data": false,
39
+ "paper_url": ""
40
+ },
41
+ {
42
+ "code_links": [
43
+ {
44
+ "url": "https://github.com/RenzKa/simlingo",
45
+ "title": "RenzKa/simlingo"
46
+ }
47
+ ],
48
+ "model_links": [],
49
+ "model_name": "SimLingo-Base (CarLLaVa)",
50
+ "metrics": {
51
+ "Driving Score": "85.94"
52
+ },
53
+ "paper_date": "2024-06-14",
54
+ "paper_title": "CarLLaVA: Vision language models for camera-only closed-loop driving",
55
+ "uses_additional_data": false,
56
+ "paper_url": "https://arxiv.org/abs/2406.10165v1"
57
+ },
58
+ {
59
+ "code_links": [
60
+ {
61
+ "url": "https://github.com/autonomousvision/carla_garage",
62
+ "title": "autonomousvision/carla_garage"
63
+ }
64
+ ],
65
+ "model_links": [],
66
+ "model_name": "TransFuser++",
67
+ "metrics": {
68
+ "Driving Score": "84.21"
69
+ },
70
+ "paper_date": "2023-06-13",
71
+ "paper_title": "Hidden Biases of End-to-End Driving Models",
72
+ "uses_additional_data": false,
73
+ "paper_url": "https://arxiv.org/abs/2306.07957v2"
74
+ },
75
+ {
76
+ "code_links": [],
77
+ "model_links": [],
78
+ "model_name": "GaussianFusion",
79
+ "metrics": {
80
+ "Driving Score": "79.4"
81
+ },
82
+ "paper_date": "2025-05-27",
83
+ "paper_title": "GaussianFusion: Gaussian-Based Multi-Sensor Fusion for End-to-End Autonomous Driving",
84
+ "uses_additional_data": false,
85
+ "paper_url": "https://arxiv.org/abs/2506.00034v1"
86
+ },
87
+ {
88
+ "code_links": [],
89
+ "model_links": [],
90
+ "model_name": "ORION",
91
+ "metrics": {
92
+ "Driving Score": "77.7"
93
+ },
94
+ "paper_date": "2025-03-25",
95
+ "paper_title": "ORION: A Holistic End-to-End Autonomous Driving Framework by Vision-Language Instructed Action Generation",
96
+ "uses_additional_data": false,
97
+ "paper_url": "https://arxiv.org/abs/2503.19755v1"
98
+ },
99
+ {
100
+ "code_links": [],
101
+ "model_links": [],
102
+ "model_name": "Raw2Drive",
103
+ "metrics": {
104
+ "Driving Score": "74.36"
105
+ },
106
+ "paper_date": "2025-05-22",
107
+ "paper_title": "Raw2Drive: Reinforcement Learning with Aligned World Models for End-to-End Autonomous Driving (in CARLA v2)",
108
+ "uses_additional_data": false,
109
+ "paper_url": "https://arxiv.org/abs/2505.16394v1"
110
+ },
111
+ {
112
+ "code_links": [],
113
+ "model_links": [],
114
+ "model_name": "ETA",
115
+ "metrics": {
116
+ "Driving Score": "74.33"
117
+ },
118
+ "paper_date": null,
119
+ "paper_title": "",
120
+ "uses_additional_data": false,
121
+ "paper_url": ""
122
+ },
123
+ {
124
+ "code_links": [],
125
+ "model_links": [],
126
+ "model_name": "DriveMoE",
127
+ "metrics": {
128
+ "Driving Score": "74.22"
129
+ },
130
+ "paper_date": "2025-05-22",
131
+ "paper_title": "DriveMoE: Mixture-of-Experts for Vision-Language-Action Model in End-to-End Autonomous Driving",
132
+ "uses_additional_data": false,
133
+ "paper_url": "https://arxiv.org/abs/2505.16278v1"
134
+ },
135
+ {
136
+ "code_links": [
137
+ {
138
+ "url": "https://github.com/woxihuanjiangguo/hydra-next",
139
+ "title": "woxihuanjiangguo/hydra-next"
140
+ }
141
+ ],
142
+ "model_links": [],
143
+ "model_name": "Hydra-NeXt",
144
+ "metrics": {
145
+ "Driving Score": "73.86"
146
+ },
147
+ "paper_date": "2025-03-15",
148
+ "paper_title": "Hydra-NeXt: Robust Closed-Loop Driving with Open-Loop Training",
149
+ "uses_additional_data": false,
150
+ "paper_url": "https://arxiv.org/abs/2503.12030v1"
151
+ },
152
+ {
153
+ "code_links": [],
154
+ "model_links": [],
155
+ "model_name": "VL (on failure)",
156
+ "metrics": {
157
+ "Driving Score": "73.29"
158
+ },
159
+ "paper_date": "2024-06-03",
160
+ "paper_title": "Validity Learning on Failures: Mitigating the Distribution Shift in Autonomous Vehicle Planning",
161
+ "uses_additional_data": false,
162
+ "paper_url": "https://arxiv.org/abs/2406.01544v2"
163
+ },
164
+ {
165
+ "code_links": [],
166
+ "model_links": [],
167
+ "model_name": "DRIVER",
168
+ "metrics": {
169
+ "Driving Score": "68.90"
170
+ },
171
+ "paper_date": null,
172
+ "paper_title": "",
173
+ "uses_additional_data": false,
174
+ "paper_url": ""
175
+ },
176
+ {
177
+ "code_links": [],
178
+ "model_links": [],
179
+ "model_name": "DiffAD",
180
+ "metrics": {
181
+ "Driving Score": "67.92"
182
+ },
183
+ "paper_date": "2025-03-15",
184
+ "paper_title": "DiffAD: A Unified Diffusion Modeling Approach for Autonomous Driving",
185
+ "uses_additional_data": false,
186
+ "paper_url": "https://arxiv.org/abs/2503.12170v1"
187
+ },
188
+ {
189
+ "code_links": [],
190
+ "model_links": [],
191
+ "model_name": "NavigationDrive",
192
+ "metrics": {
193
+ "Driving Score": "67.17"
194
+ },
195
+ "paper_date": null,
196
+ "paper_title": "",
197
+ "uses_additional_data": false,
198
+ "paper_url": ""
199
+ },
200
+ {
201
+ "code_links": [
202
+ {
203
+ "url": "https://github.com/Kguo-cs/iPad",
204
+ "title": "Kguo-cs/iPad"
205
+ }
206
+ ],
207
+ "model_links": [],
208
+ "model_name": "iPad",
209
+ "metrics": {
210
+ "Driving Score": "65.02"
211
+ },
212
+ "paper_date": "2025-05-21",
213
+ "paper_title": "iPad: Iterative Proposal-centric End-to-End Autonomous Driving",
214
+ "uses_additional_data": false,
215
+ "paper_url": "https://arxiv.org/abs/2505.15111v1"
216
+ },
217
+ {
218
+ "code_links": [
219
+ {
220
+ "url": "https://github.com/opendrivelab/driveadapter",
221
+ "title": "opendrivelab/driveadapter"
222
+ }
223
+ ],
224
+ "model_links": [],
225
+ "model_name": "DriveAdapter",
226
+ "metrics": {
227
+ "Driving Score": "64.22"
228
+ },
229
+ "paper_date": "2023-08-01",
230
+ "paper_title": "DriveAdapter: Breaking the Coupling Barrier of Perception and Planning in End-to-End Autonomous Driving",
231
+ "uses_additional_data": false,
232
+ "paper_url": "https://arxiv.org/abs/2308.00398v2"
233
+ },
234
+ {
235
+ "code_links": [
236
+ {
237
+ "url": "https://github.com/liuxueyi/reasonplan",
238
+ "title": "liuxueyi/reasonplan"
239
+ }
240
+ ],
241
+ "model_links": [],
242
+ "model_name": "ReasonPlan",
243
+ "metrics": {
244
+ "Driving Score": "64.01"
245
+ },
246
+ "paper_date": "2025-05-26",
247
+ "paper_title": "ReasonPlan: Unified Scene Prediction and Decision Reasoning for Closed-loop Autonomous Driving",
248
+ "uses_additional_data": false,
249
+ "paper_url": "https://arxiv.org/abs/2505.20024v1"
250
+ },
251
+ {
252
+ "code_links": [
253
+ {
254
+ "url": "https://github.com/thinklab-sjtu/drivetransformer",
255
+ "title": "thinklab-sjtu/drivetransformer"
256
+ }
257
+ ],
258
+ "model_links": [],
259
+ "model_name": "Drivetransformer-Large",
260
+ "metrics": {
261
+ "Driving Score": "63.46"
262
+ },
263
+ "paper_date": "2025-03-07",
264
+ "paper_title": "DriveTransformer: Unified Transformer for Scalable End-to-End Autonomous Driving",
265
+ "uses_additional_data": false,
266
+ "paper_url": "https://arxiv.org/abs/2503.07656v1"
267
+ },
268
+ {
269
+ "code_links": [
270
+ {
271
+ "url": "https://github.com/opendrivelab/thinktwice",
272
+ "title": "opendrivelab/thinktwice"
273
+ }
274
+ ],
275
+ "model_links": [],
276
+ "model_name": "ThinkTwice",
277
+ "metrics": {
278
+ "Driving Score": "62.44"
279
+ },
280
+ "paper_date": "2023-05-10",
281
+ "paper_title": "Think Twice before Driving: Towards Scalable Decoders for End-to-End Autonomous Driving",
282
+ "uses_additional_data": false,
283
+ "paper_url": "https://arxiv.org/abs/2305.06242v1"
284
+ },
285
+ {
286
+ "code_links": [
287
+ {
288
+ "url": "https://github.com/OpenPerceptionX/TCP",
289
+ "title": "OpenPerceptionX/TCP"
290
+ }
291
+ ],
292
+ "model_links": [],
293
+ "model_name": "TCP-traj",
294
+ "metrics": {
295
+ "Driving Score": "59.90"
296
+ },
297
+ "paper_date": "2022-06-16",
298
+ "paper_title": "Trajectory-guided Control Prediction for End-to-end Autonomous Driving: A Simple yet Strong Baseline",
299
+ "uses_additional_data": false,
300
+ "paper_url": "https://arxiv.org/abs/2206.08129v2"
301
+ },
302
+ {
303
+ "code_links": [
304
+ {
305
+ "url": "https://github.com/suhaisheng/difsd",
306
+ "title": "suhaisheng/difsd"
307
+ }
308
+ ],
309
+ "model_links": [],
310
+ "model_name": "DiFSD",
311
+ "metrics": {
312
+ "Driving Score": "52.02"
313
+ },
314
+ "paper_date": "2024-09-15",
315
+ "paper_title": "DiFSD: Ego-Centric Fully Sparse Paradigm with Uncertainty Denoising and Iterative Refinement for Efficient End-to-End Self-Driving",
316
+ "uses_additional_data": false,
317
+ "paper_url": "https://arxiv.org/abs/2409.09777v4"
318
+ },
319
+ {
320
+ "code_links": [],
321
+ "model_links": [],
322
+ "model_name": "X-Driver",
323
+ "metrics": {
324
+ "Driving Score": "51.70"
325
+ },
326
+ "paper_date": "2025-05-08",
327
+ "paper_title": "X-Driver: Explainable Autonomous Driving with Vision-Language Models",
328
+ "uses_additional_data": false,
329
+ "paper_url": "https://arxiv.org/abs/2505.05098v2"
330
+ },
331
+ {
332
+ "code_links": [
333
+ {
334
+ "url": "https://github.com/OpenPerceptionX/TCP",
335
+ "title": "OpenPerceptionX/TCP"
336
+ }
337
+ ],
338
+ "model_links": [],
339
+ "model_name": "TCP-traj w/o distillation",
340
+ "metrics": {
341
+ "Driving Score": "49.30"
342
+ },
343
+ "paper_date": "2022-06-16",
344
+ "paper_title": "Trajectory-guided Control Prediction for End-to-end Autonomous Driving: A Simple yet Strong Baseline",
345
+ "uses_additional_data": false,
346
+ "paper_url": "https://arxiv.org/abs/2206.08129v2"
347
+ },
348
+ {
349
+ "code_links": [],
350
+ "model_links": [],
351
+ "model_name": "CogAD",
352
+ "metrics": {
353
+ "Driving Score": "48.30"
354
+ },
355
+ "paper_date": "2025-05-27",
356
+ "paper_title": "CogAD: Cognitive-Hierarchy Guided End-to-End Autonomous Driving",
357
+ "uses_additional_data": false,
358
+ "paper_url": "https://arxiv.org/abs/2505.21581v2"
359
+ },
360
+ {
361
+ "code_links": [],
362
+ "model_links": [],
363
+ "model_name": "MomAD",
364
+ "metrics": {
365
+ "Driving Score": "47.91"
366
+ },
367
+ "paper_date": null,
368
+ "paper_title": "",
369
+ "uses_additional_data": false,
370
+ "paper_url": ""
371
+ },
372
+ {
373
+ "code_links": [
374
+ {
375
+ "url": "https://github.com/opendrivelab/uniad",
376
+ "title": "opendrivelab/uniad"
377
+ }
378
+ ],
379
+ "model_links": [],
380
+ "model_name": "UniAD-Base",
381
+ "metrics": {
382
+ "Driving Score": "45.81"
383
+ },
384
+ "paper_date": "2022-12-20",
385
+ "paper_title": "Planning-oriented Autonomous Driving",
386
+ "uses_additional_data": false,
387
+ "paper_url": "https://arxiv.org/abs/2212.10156v2"
388
+ },
389
+ {
390
+ "code_links": [],
391
+ "model_links": [],
392
+ "model_name": "TTOG",
393
+ "metrics": {
394
+ "Driving Score": "45.23"
395
+ },
396
+ "paper_date": "2025-04-17",
397
+ "paper_title": "Two Tasks, One Goal: Uniting Motion and Planning for Excellent End To End Autonomous Driving Performance",
398
+ "uses_additional_data": false,
399
+ "paper_url": "https://arxiv.org/abs/2504.12667v1"
400
+ },
401
+ {
402
+ "code_links": [
403
+ {
404
+ "url": "https://github.com/wzzheng/genad",
405
+ "title": "wzzheng/genad"
406
+ }
407
+ ],
408
+ "model_links": [],
409
+ "model_name": "GenAD",
410
+ "metrics": {
411
+ "Driving Score": "44.81"
412
+ },
413
+ "paper_date": "2024-02-18",
414
+ "paper_title": "GenAD: Generative End-to-End Autonomous Driving",
415
+ "uses_additional_data": false,
416
+ "paper_url": "https://arxiv.org/abs/2402.11502v3"
417
+ },
418
+ {
419
+ "code_links": [
420
+ {
421
+ "url": "https://github.com/swc-17/sparsedrive",
422
+ "title": "swc-17/sparsedrive"
423
+ },
424
+ {
425
+ "url": "https://github.com/linxuewu/sparse4d",
426
+ "title": "linxuewu/sparse4d"
427
+ }
428
+ ],
429
+ "model_links": [],
430
+ "model_name": "SparseDrive",
431
+ "metrics": {
432
+ "Driving Score": "44.54"
433
+ },
434
+ "paper_date": "2024-05-30",
435
+ "paper_title": "SparseDrive: End-to-End Autonomous Driving via Sparse Scene Representation",
436
+ "uses_additional_data": false,
437
+ "paper_url": "https://arxiv.org/abs/2405.19620v2"
438
+ },
439
+ {
440
+ "code_links": [
441
+ {
442
+ "url": "https://github.com/hustvl/vad",
443
+ "title": "hustvl/vad"
444
+ },
445
+ {
446
+ "url": "https://github.com/yangyucheng000/papercode-2/tree/main/WS-VAD-mindspore-main",
447
+ "title": "yangyucheng000/papercode-2"
448
+ }
449
+ ],
450
+ "model_links": [],
451
+ "model_name": "VAD",
452
+ "metrics": {
453
+ "Driving Score": "42.35"
454
+ },
455
+ "paper_date": "2023-03-21",
456
+ "paper_title": "VAD: Vectorized Scene Representation for Efficient Autonomous Driving",
457
+ "uses_additional_data": false,
458
+ "paper_url": "https://arxiv.org/abs/2303.12077v3"
459
+ },
460
+ {
461
+ "code_links": [
462
+ {
463
+ "url": "https://github.com/opendrivelab/uniad",
464
+ "title": "opendrivelab/uniad"
465
+ }
466
+ ],
467
+ "model_links": [],
468
+ "model_name": "UniAD-Tiny",
469
+ "metrics": {
470
+ "Driving Score": "40.73"
471
+ },
472
+ "paper_date": "2022-12-20",
473
+ "paper_title": "Planning-oriented Autonomous Driving",
474
+ "uses_additional_data": false,
475
+ "paper_url": "https://arxiv.org/abs/2212.10156v2"
476
+ },
477
+ {
478
+ "code_links": [
479
+ {
480
+ "url": "https://github.com/OpenPerceptionX/TCP",
481
+ "title": "OpenPerceptionX/TCP"
482
+ }
483
+ ],
484
+ "model_links": [],
485
+ "model_name": "TCP",
486
+ "metrics": {
487
+ "Driving Score": "40.70"
488
+ },
489
+ "paper_date": "2022-06-16",
490
+ "paper_title": "Trajectory-guided Control Prediction for End-to-end Autonomous Driving: A Simple yet Strong Baseline",
491
+ "uses_additional_data": false,
492
+ "paper_url": "https://arxiv.org/abs/2206.08129v2"
493
+ },
494
+ {
495
+ "code_links": [],
496
+ "model_links": [],
497
+ "model_name": "VAD + SERA",
498
+ "metrics": {
499
+ "Driving Score": "35.64"
500
+ },
501
+ "paper_date": "2025-05-28",
502
+ "paper_title": "From Failures to Fixes: LLM-Driven Scenario Repair for Self-Evolving Autonomous Driving",
503
+ "uses_additional_data": false,
504
+ "paper_url": "https://arxiv.org/abs/2505.22067v1"
505
+ },
506
+ {
507
+ "code_links": [
508
+ {
509
+ "url": "https://github.com/OpenPerceptionX/TCP",
510
+ "title": "OpenPerceptionX/TCP"
511
+ }
512
+ ],
513
+ "model_links": [],
514
+ "model_name": "TCP-ctrl",
515
+ "metrics": {
516
+ "Driving Score": "30.47"
517
+ },
518
+ "paper_date": "2022-06-16",
519
+ "paper_title": "Trajectory-guided Control Prediction for End-to-end Autonomous Driving: A Simple yet Strong Baseline",
520
+ "uses_additional_data": false,
521
+ "paper_url": "https://arxiv.org/abs/2206.08129v2"
522
+ },
523
+ {
524
+ "code_links": [
525
+ {
526
+ "url": "https://github.com/Thinklab-SJTU/Bench2Drive",
527
+ "title": "Thinklab-SJTU/Bench2Drive"
528
+ },
529
+ {
530
+ "url": "https://github.com/autonomousvision/carla_garage",
531
+ "title": "autonomousvision/carla_garage"
532
+ },
533
+ {
534
+ "url": "https://github.com/thinklab-sjtu/bench2drivezoo",
535
+ "title": "thinklab-sjtu/bench2drivezoo"
536
+ },
537
+ {
538
+ "url": "https://github.com/RenzKa/simlingo",
539
+ "title": "RenzKa/simlingo"
540
+ }
541
+ ],
542
+ "model_links": [],
543
+ "model_name": "AD-MLP",
544
+ "metrics": {
545
+ "Driving Score": "18.05"
546
+ },
547
+ "paper_date": "2024-06-06",
548
+ "paper_title": "Bench2Drive: Towards Multi-Ability Benchmarking of Closed-Loop End-To-End Autonomous Driving",
549
+ "uses_additional_data": false,
550
+ "paper_url": "https://arxiv.org/abs/2406.03877v3"
551
+ }
552
+ ],
553
+ "metrics": [
554
+ "Driving Score"
555
+ ]
556
+ },
557
+ "dataset_links": [
558
+ {
559
+ "url": "https://paperswithcode.com/sota/bench2drive-on-bench2drive",
560
+ "title": "Papers with Code Leaderboard URL"
561
+ }
562
+ ],
563
+ "description": "",
564
+ "subdatasets": [],
565
+ "dataset_citations": [],
566
+ "dataset": "Bench2Drive"
567
+ }
568
+ ]
569
+ }