Improve model card: add metadata, GitHub link, and sample usage

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +36 -0
README.md CHANGED
@@ -1,5 +1,41 @@
 
 
 
 
 
 
 
 
 
1
  This is the 14B model for the paper [Large-Scale Terminal Agentic Trajectory Generation from Dockerized Environments](https://huggingface.co/papers/2602.01244).
2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ## Citation
4
 
5
  **BibTeX:**
 
1
+ ---
2
+ library_name: transformers
3
+ pipeline_tag: text-generation
4
+ datasets:
5
+ - m-a-p/TerminalTraj
6
+ ---
7
+
8
+ # TerminalTraj-14B
9
+
10
  This is the 14B model for the paper [Large-Scale Terminal Agentic Trajectory Generation from Dockerized Environments](https://huggingface.co/papers/2602.01244).
11
 
12
+ **TerminalTraj** is a scalable pipeline designed to generate high-quality terminal trajectories that capture realistic long-horizon interactions across diverse domains. It addresses the challenges of executability and verifiability by (i) filtering high-quality repositories to construct Dockerized execution environments, (ii) generating Docker-aligned task instances, and (iii) synthesizing agent trajectories with executable validation code.
13
+
14
+ The model is based on the **Qwen2.5-Coder** backbone and demonstrates significant performance improvements on terminal-based agentic tasks (TerminalBench).
15
+
16
+ - **GitHub Repository:** [multimodal-art-projection/TerminalTraj](https://github.com/multimodal-art-projection/TerminalTraj)
17
+ - **Paper:** [Large-Scale Terminal Agentic Trajectory Generation from Dockerized Environments](https://huggingface.co/papers/2602.01244)
18
+ - **Dataset:** [m-a-p/TerminalTraj](https://huggingface.co/datasets/m-a-p/TerminalTraj)
19
+
20
+ ## Usage
21
+
22
+ You can use this model with the `transformers` library:
23
+
24
+ ```python
25
+ from transformers import AutoTokenizer, AutoModelForCausalLM
26
+ import torch
27
+
28
+ model_id = "m-a-p/TerminalTraj-14B"
29
+
30
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
31
+
32
+ model = AutoModelForCausalLM.from_pretrained(
33
+ model_id,
34
+ torch_dtype=torch.float16, # 14B建议用fp16或bf16
35
+ device_map="auto" # 自动分配GPU
36
+ )
37
+ ```
38
+
39
  ## Citation
40
 
41
  **BibTeX:**