File size: 3,396 Bytes
100afe5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
---
library_name: transformers
license: apache-2.0
base_model:
  - Qwen/Qwen3-VL-8B-Instruct
tags:
  - multimodal
  - vision-language
  - tool-use
  - agentic
  - qwen3_vl
  - sft
datasets:
  - Accio-Lab/Metis-ColdStart
language:
  - en
pipeline_tag: image-text-to-text
---

# Metis-8B-ColdStart

**Act Wisely: Cultivating Meta-Cognitive Tool Use in Agentic Multimodal Models**

Metis-8B-ColdStart is the **SFT (Supervised Fine-Tuning) checkpoint** of the Metis framework, fine-tuned from [Qwen3-VL-8B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-8B-Instruct) on the curated [Metis-ColdStart](https://huggingface.co/datasets/Accio-Lab/Metis-ColdStart) dataset. This checkpoint serves as the starting point for HDPO reinforcement learning, which produces the final [Metis-8B-RL](https://huggingface.co/Accio-Lab/Metis-8B-RL) model.

[[Paper (arXiv)]](https://arxiv.org/abs/2604.08545) | [[GitHub]](https://github.com/Accio-Lab/Metis) | [[RL Model]](https://huggingface.co/Accio-Lab/Metis-8B-RL) | [[ColdStart Data]](https://huggingface.co/datasets/Accio-Lab/Metis-ColdStart) | [[RL Data]](https://huggingface.co/datasets/Accio-Lab/Metis-RL)

## Model Details

| Attribute | Value |
|---|---|
| Base model | [Qwen3-VL-8B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-8B-Instruct) |
| Training stage | Supervised Fine-Tuning (Cold Start) |
| Training data | [Metis-ColdStart](https://huggingface.co/datasets/Accio-Lab/Metis-ColdStart) (~27K samples) |
| Next stage | → [Metis-8B-RL](https://huggingface.co/Accio-Lab/Metis-8B-RL) (HDPO reinforcement learning) |
| License | Apache-2.0 |

## Cold Start Data Curation Pipeline

The SFT corpus is curated from publicly available tool-augmented multimodal trajectories (DeepEyesV2, V-Interaction, Thyme, OpenMMReasoner) through a rigorous three-stage pipeline:

1. **Eradicating hallucinated environmental dynamics** — Execute all code in a sandbox environment; discard trajectories with execution failures.
2. **Isolating genuine tool necessity** — Filter out samples where the base model achieves pass@8 = 1 without any tools, ensuring only genuinely tool-dependent samples remain.
3. **Multidimensional meta-cognitive filtering** — An LLM judge evaluates visual relevance, reasoning coherence, and tool-use rationale to ensure high quality.

## Training Pipeline

```
Qwen3-VL-8B-Instruct

        ▼  SFT on Metis-ColdStart (~27K samples)
  Metis-8B-ColdStart  ← (this checkpoint)

        ▼  HDPO on Metis-RL (~5K prompts)
   Metis-8B-RL  (final model)
```

## Usage

Please refer to the [GitHub repository](https://github.com/Accio-Lab/Metis) for full installation and inference instructions.

### Installation

```bash
git clone https://github.com/Accio-Lab/Metis.git
cd Metis
pip install -e verl
pip install -e ".[vllm,search_tool,python_code_dep]"
```

## Citation

```bibtex
@article{yan2026metis,
  title={Act Wisely: Cultivating Meta-Cognitive Tool Use in Agentic Multimodal Models},
  author={Yan, Shilin and Tong, Jintao and Xue, Hongwei and Tang, Xiaojun and Wang, Yangyang and Shi, Kunyu and Zhang, Guannan and Li, Ruixuan and Zou, Yixiong},
  journal={arXiv preprint arXiv:2604.08545},
  year={2026}
}
```

## Acknowledgments

Metis is built upon [verl](https://github.com/volcengine/verl), [verl-tool](https://github.com/TIGER-AI-Lab/verl-tool), and [Qwen3-VL](https://github.com/QwenLM/Qwen3-VL).