File size: 3,874 Bytes
2b3b9b6 0686082 2b3b9b6 0686082 2b3b9b6 0686082 b060c23 2b3b9b6 0686082 2b3b9b6 e4b5dd1 2b3b9b6 e4b5dd1 2b3b9b6 b060c23 2b3b9b6 0686082 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 | ---
license: other
license_name: license
license_link: LICENSE
pipeline_tag: depth-estimation
---
<div align="center">
<h1>WildCross: A Cross-Modal Large Scale Benchmark for Place Recognition and Metric Depth Estimation in Natural Environments</h1>
[**Joshua Knights**](https://scholar.google.com/citations?user=RxbGr2EAAAAJ&hl=en)<sup>1,2</sup> 路 **Joseph Reid**<sup>1</sup> 路 [**Mark Cox**](https://scholar.google.com/citations?user=Bk3UD4EAAAAJ&hl=en)<sup>1</sup>
<br>
[**Kaushik Roy**](https://bit0123.github.io/)<sup>1</sup> 路 [**David Hall**](https://scholar.google.com/citations?user=dosODoQAAAAJ&hl=en)<sup>1</sup> 路 [**Peyman Moghadam**](https://scholar.google.com.au/citations?user=QAVcuWUAAAAJ&hl=en)<sup>1,2</sup>
<sup>1</sup>DATA61, CSIRO   <sup>2</sup>Queensland University of Technology
<br>
<a href="https://huggingface.co/papers/2603.01475"><img src='https://img.shields.io/badge/arXiv-WildCross-red' alt='Paper PDF'></a>
<a href='https://csiro-robotics.github.io/WildCross'><img src='https://img.shields.io/badge/Project_Page-WildCross-green' alt='Project Page'></a>
<a href='https://doi.org/10.25919/5fmy-yg37'><img src='https://img.shields.io/badge/Dataset_Download-WildCross-blue'></a>
</div>
This repository contains the pre-trained checkpoints for a variety of tasks on the WildCross benchmark.

If you find this repository useful or use the WildCross dataset in your work, please cite us using the following:
```
@inproceedings{wildcross2026,
title={{WildCross: A Cross-Modal Large Scale Benchmark for Place Recognition and Metric Depth Estimation in Natural Environments}},
author={Joshua Knights, Joseph Reid, Kaushik Roy, David Hall, Mark Cox, Peyman Moghadam},
booktitle={Proceedings-IEEE International Conference on Robotics and Automation},
pages={},
year={2026}
}
```
## Download Instructions
Our dataset can be downloaded through the [**CSIRO Data Access Portal**](https://doi.org/10.25919/5fmy-yg37). Detailed instructions for downloading the dataset can be found in the README file provided on the data access portal page.
## Training and Benchmarking
Here we provide pre-trained checkpoints for a variety of tasks on WildCross.
**Visual Place Recognition**
### Checkpoints
| Model | Checkpoint Folder|
|------------|------------|
| NetVlad | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/VPR/NetVLAD) |
| MixVPR | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/VPR/MixVPR) |
| SALAD | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/VPR/SALAD) |
| BoQ | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/VPR/BoQ) |
**Cross Modal Place Recognition**
### Checkpoints
| Model | Checkpoint Folder|
|------------|------------|
| Lip-Loc (ResNet50) | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/crossmodal/resnet50) |
| Lip-Loc (Dino-v2) | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/crossmodal/dinov2) |
| Lip-Loc (Dino-v3) | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/crossmodal/dinov3) |
**Metric Depth Estimation**
### Checkpoints
| Model | Checkpoint Folder|
|------------|------------|
| DepthAnythingV2-vits | [Link](https://huggingface.co/CSIRORobotics/WildCross/resolve/main/DepthAnythingV2/finetuned/vits.pth) |
| DepthAnythingV2-vitb | [Link](https://huggingface.co/CSIRORobotics/WildCross/resolve/main/DepthAnythingV2/finetuned/vitb.pth) |
| DepthAnythingV2-vitl | [Link](https://huggingface.co/CSIRORobotics/WildCross/resolve/main/DepthAnythingV2/finetuned/vitl.pth) |
For instructions on how to use these checkpoints for training or evaluation, further instructions can be found on the [WildCross GitHub repository](https://github.com/csiro-robotics/WildCross). |