Update README.md
Browse files
README.md
CHANGED
|
@@ -7,22 +7,21 @@ license: cc-by-nc-sa-4.0
|
|
| 7 |

|
| 8 |

|
| 9 |
|
| 10 |
-
---
|
| 11 |
## Model Description
|
| 12 |
|
| 13 |
VL-LN Bench is the first benchmark for **Interactive Instance Object Navigation (IION)**, where an embodied agent must locate a specific object instance in a realistic 3D home while engaging in **free-form natural-language dialogue**. It also provides an **automated data-collection pipeline** that generates large-scale training data for learning interactive navigation behaviors. Using this dataset, we train an **IION base model** that shares the same architecture as **InternVLA-N1**.
|
| 14 |
|
| 15 |
The resulting model demonstrates baseline competence on IION: it can search for a specific instance in **previously unseen** environments. During exploration, the agent can either **move** by predicting a pixel-goal waypoint or **ask** a question to reduce ambiguity and improve task success and efficiency.
|
| 16 |
|
| 17 |
-
|
| 18 |
-
###
|
| 19 |
|
| 20 |
[](https://github.com/InternRobotics/InternNav)
|
| 21 |
[](https://arxiv.org/abs/2512.08186)
|
| 22 |
[](https://0309hws.github.io/VL-LN.github.io/)
|
| 23 |
[](https://huggingface.co/datasets/InternRobotics/InternData-N1)
|
| 24 |
|
| 25 |
-
|
| 26 |
## Usage
|
| 27 |
|
| 28 |
-
For inference and evaluation please refer to the [VL-LN-Bench repository](https://github.com/InternRobotics/InternNav).
|
|
|
|
| 7 |

|
| 8 |

|
| 9 |
|
|
|
|
| 10 |
## Model Description
|
| 11 |
|
| 12 |
VL-LN Bench is the first benchmark for **Interactive Instance Object Navigation (IION)**, where an embodied agent must locate a specific object instance in a realistic 3D home while engaging in **free-form natural-language dialogue**. It also provides an **automated data-collection pipeline** that generates large-scale training data for learning interactive navigation behaviors. Using this dataset, we train an **IION base model** that shares the same architecture as **InternVLA-N1**.
|
| 13 |
|
| 14 |
The resulting model demonstrates baseline competence on IION: it can search for a specific instance in **previously unseen** environments. During exploration, the agent can either **move** by predicting a pixel-goal waypoint or **ask** a question to reduce ambiguity and improve task success and efficiency.
|
| 15 |
|
| 16 |
+
|
| 17 |
+
### Resources
|
| 18 |
|
| 19 |
[](https://github.com/InternRobotics/InternNav)
|
| 20 |
[](https://arxiv.org/abs/2512.08186)
|
| 21 |
[](https://0309hws.github.io/VL-LN.github.io/)
|
| 22 |
[](https://huggingface.co/datasets/InternRobotics/InternData-N1)
|
| 23 |
|
| 24 |
+
|
| 25 |
## Usage
|
| 26 |
|
| 27 |
+
For inference and evaluation, please refer to the [VL-LN-Bench repository](https://github.com/InternRobotics/InternNav).
|