Improve model card: Add pipeline tag, paper, code, and usage
Browse filesThis PR significantly enhances the model card for the VISION model by:
- Adding the `pipeline_tag: image-to-image`, which correctly categorizes the model and improves its discoverability on the Hugging Face Hub.
- Including a direct link to the associated paper: [VISION: Prompting Ocean Vertical Velocity Reconstruction from Incomplete Observations](https://huggingface.co/papers/2509.21477).
- Providing the paper's abstract for a comprehensive overview.
- Adding a link to the official GitHub repository for easy access to the code and resources.
- Including a "Quick Start" section with installation and inference code snippets directly from the GitHub README, enabling users to quickly get started with the model.
- Incorporating the main illustrative image from the GitHub repository and a proper BibTeX citation.
Please review these additions and merge if they align with the repository's guidelines.
|
@@ -1,3 +1,69 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
--
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
pipeline_tag: image-to-image
|
| 4 |
+
---
|
| 5 |
+
|
| 6 |
+
# VISION: Prompting Ocean Vertical Velocity Reconstruction from Incomplete Observations
|
| 7 |
+
|
| 8 |
+
This model is presented in the paper [VISION: Prompting Ocean Vertical Velocity Reconstruction from Incomplete Observations](https://huggingface.co/papers/2509.21477).
|
| 9 |
+
|
| 10 |
+
<div align="center">
|
| 11 |
+
<img src="https://raw.githubusercontent.com/YuanGao-YG/VISION/main/img/vision_main.jpg" width="1080">
|
| 12 |
+
</div>
|
| 13 |
+
|
| 14 |
+
## Abstract
|
| 15 |
+
Reconstructing subsurface ocean dynamics, such as vertical velocity fields, from incomplete surface observations poses a critical challenge in Earth science, a field long hampered by the lack of standardized, analysis-ready benchmarks. To systematically address this issue and catalyze research, we first build and release KD48, a high-resolution ocean dynamics benchmark derived from petascale simulations and curated with expert-driven denoising. Building on this benchmark, we introduce VISION, a novel reconstruction paradigm based on Dynamic Prompting designed to tackle the core problem of missing data in real-world observations. The essence of VISION lies in its ability to generate a visual prompt on-the-fly from any available subset of observations, which encodes both data availability and the ocean's physical state. More importantly, we design a State-conditioned Prompting module that efficiently injects this prompt into a universal backbone, endowed with geometry- and scale-aware operators, to guide its adaptive adjustment of computational strategies. This mechanism enables VISION to precisely handle the challenges posed by varying input combinations. Extensive experiments on the KD48 benchmark demonstrate that VISION not only substantially outperforms state-of-the-art models but also exhibits strong generalization under extreme data missing scenarios. By providing a high-quality benchmark and a robust model, our work establishes a solid infrastructure for ocean science research under data uncertainty.
|
| 16 |
+
|
| 17 |
+
## Code
|
| 18 |
+
The official code is available on the [GitHub repository](https://github.com/YuanGao-YG/VISION).
|
| 19 |
+
|
| 20 |
+
## Quick Start
|
| 21 |
+
|
| 22 |
+
### Installation
|
| 23 |
+
|
| 24 |
+
- cuda 11.8
|
| 25 |
+
|
| 26 |
+
```bash
|
| 27 |
+
# git clone this repository
|
| 28 |
+
git clone https://github.com/YuanGao-YG/VISION.git
|
| 29 |
+
cd VISION
|
| 30 |
+
|
| 31 |
+
# create new anaconda env
|
| 32 |
+
conda env create -f environment.yml
|
| 33 |
+
conda activate vision
|
| 34 |
+
```
|
| 35 |
+
|
| 36 |
+
### Inference
|
| 37 |
+
|
| 38 |
+
Pre-trained weights and test data are available on [Hugging Face](https://huggingface.co/YuanGao-YG/VISION/tree/main). Download them and place them in the same location.
|
| 39 |
+
|
| 40 |
+
1. Inference using Incomplete Observations (SSH)
|
| 41 |
+
```bash
|
| 42 |
+
python inference_io_ssh_vision.py
|
| 43 |
+
```
|
| 44 |
+
|
| 45 |
+
3. Inference using Incomplete Observations (SSH + U + V)
|
| 46 |
+
```bash
|
| 47 |
+
python inference_io_ssh_u_v_vision.py
|
| 48 |
+
```
|
| 49 |
+
|
| 50 |
+
2. Inference using Complete Observations (SSH + U + V + B)
|
| 51 |
+
```bash
|
| 52 |
+
python inference_co_ssh_u_v_b_vision.py
|
| 53 |
+
```
|
| 54 |
+
|
| 55 |
+
## Citation
|
| 56 |
+
If you find our work useful, please consider citing it:
|
| 57 |
+
|
| 58 |
+
```bibtex
|
| 59 |
+
@article{gao2025vision,
|
| 60 |
+
title={VISION: Prompting Ocean Vertical Velocity Reconstruction from Incomplete Observations},
|
| 61 |
+
author={Gao, Yuan and Wu, Hao and Wen, Qingsong and Wang, Kun and Wu, Xian and Huang, Xiaomeng},
|
| 62 |
+
journal={arXiv preprint arXiv:2509.21477},
|
| 63 |
+
year={2025},
|
| 64 |
+
eprint={2509.21477},
|
| 65 |
+
archivePrefix={arXiv},
|
| 66 |
+
primaryClass={cs.CL},
|
| 67 |
+
url={https://arxiv.org/abs/2509.21477}
|
| 68 |
+
}
|
| 69 |
+
```
|