Zery commited on
Commit
f32bf82
·
verified ·
1 Parent(s): 7b78303

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -48
README.md CHANGED
@@ -1,56 +1,15 @@
1
  ---
2
- datasets: boot_osworld_vscode
 
3
  library_name: transformers
4
  tags:
5
  - generated_from_trainer
6
  - R1-V
7
  licence: license
 
 
 
 
8
  ---
9
 
10
- # Model Card for None
11
-
12
- This model is a fine-tuned version of [None](https://huggingface.co/None) on the [boot_osworld_vscode](https://huggingface.co/datasets/boot_osworld_vscode) dataset.
13
- It has been trained using [TRL](https://github.com/huggingface/trl).
14
-
15
- ## Quick start
16
-
17
- ```python
18
- from transformers import pipeline
19
-
20
- question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
21
- generator = pipeline("text-generation", model="None", device="cuda")
22
- output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
23
- print(output["generated_text"])
24
- ```
25
-
26
- ## Training procedure
27
-
28
-
29
-
30
-
31
- This model was trained with SFT.
32
-
33
- ### Framework versions
34
-
35
- - TRL: 0.17.0
36
- - Transformers: 4.51.3
37
- - Pytorch: 2.6.0
38
- - Datasets: 3.5.1
39
- - Tokenizers: 0.21.1
40
-
41
- ## Citations
42
-
43
-
44
-
45
- Cite TRL as:
46
-
47
- ```bibtex
48
- @misc{vonwerra2022trl,
49
- title = {{TRL: Transformer Reinforcement Learning}},
50
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
51
- year = 2020,
52
- journal = {GitHub repository},
53
- publisher = {GitHub},
54
- howpublished = {\url{https://github.com/huggingface/trl}}
55
- }
56
- ```
 
1
  ---
2
+ datasets:
3
+ - OS-Copilot/ScienceBoard-Traj
4
  library_name: transformers
5
  tags:
6
  - generated_from_trainer
7
  - R1-V
8
  licence: license
9
+ license: apache-2.0
10
+ base_model:
11
+ - Qwen/Qwen2.5-VL-32B-Instruct
12
+ pipeline_tag: image-text-to-text
13
  ---
14
 
15
+ Check out our [repo](https://github.com/OpenIXCLab/CODA) and [paper](https://arxiv.org/abs/2508.20096) for more implementation details!