nielsr HF Staff commited on
Commit
ebae4ec
·
verified ·
1 Parent(s): c826080

Update pipeline tag to graph-ml and refine model card formatting

Browse files

This PR improves the model card for `PKU-ML/G1-Direct-SFT-7B` by:
- Updating the `pipeline_tag` from `text-generation` to `graph-ml` in the metadata. This change more accurately reflects the model's primary capability in graph reasoning tasks, improving its discoverability on the Hugging Face Hub.
- Removing the redundant `

Files changed (1) hide show
  1. README.md +26 -19
README.md CHANGED
@@ -1,41 +1,40 @@
1
  ---
2
- license: apache-2.0
 
3
  datasets:
4
  - PKU-ML/Erdos
5
  language:
6
  - en
 
 
7
  metrics:
8
  - accuracy
9
- base_model:
10
- - Qwen/Qwen2.5-7B-Instruct
11
- pipeline_tag: text-generation
12
  tags:
13
  - graph
14
  - chat
15
- library_name: transformers
16
  ---
17
 
18
-
19
  # G1-Direct-SFT-7B
20
 
21
  ## Introduction
22
 
23
- G1 is the series of large language models trained on our benchmark [Erdos](https://huggingface.co/datasets/PKU-ML/Erdos) for solving graph reasoning tasks, based on Qwen2.5-Instruct.
24
  We apply Group Relative Policy Optimization (GRPO) for reinforcement learning with supervised finetuning as a prelimary step.
25
 
26
  G1 brings the following improvements:
27
 
28
- - **Significant improvement on graph reasoning**: G1 models achieve up to 46% improvement over baselines on Erdős, with the 7B variant matching OpenAI’s o3-mini and the 3B model surpassing Qwen2.5-72B-Instruct by notable margins.
29
- - **Strong Generalization to unseen graph tasks**: G1 exhibits zero-shot generalization on unseen graph tasks, improving performance on *other graph reasoning benchmarks* (GraphWiz, GraphArena) and *real-world graphs* (Cora, PubMed).
30
- - **NO Compromise on general reasoning**: Crucially, G1 preserves general reasoning ability (GSM8K, MATH, MMLU-Pro), proving its versatility.
31
 
32
 
33
  **This repo contains the G1-Direct-SFT-7B model**, which has the following features:
34
- - Type: Causal Language Models
35
- - Training Stage: SFT
36
- - Architecture: the same with Qwen2.5-Instruct
37
- - Number of Parameters: 7.62B
38
- - Context Length: Full 32,768 tokens and generation 8192 tokens
39
 
40
  For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
41
 
@@ -72,10 +71,18 @@ model = AutoModelForCausalLM.from_pretrained(
72
  )
73
  tokenizer = AutoTokenizer.from_pretrained(model_name)
74
 
75
- prompt = "The task is to determine the degree centrality of a node in the graph.\n\n"\
76
- "Degree centrality for a node is the fraction of nodes it is connected to.\n\n"\
77
- "Here is an undirected graph containing nodes from 1 to 15. The edges are: (1, 15), (15, 11), (2, 3), (2, 6), (3, 6), (3, 7), (6, 7), (6, 8), (7, 8), (7, 14), (4, 10), (10, 5), (10, 12), (8, 14), (8, 9), (12, 11), (12, 13).\n\n"\
78
- "Question: What is the degree centrality of node 2 in the graph?\n\n"\
 
 
 
 
 
 
 
 
79
  "You need to format your answer as a float number."
80
  messages = [
81
  {"role": "user", "content": INSTRUCTION_TEMPLATE.format(instruction=prompt)}
 
1
  ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-7B-Instruct
4
  datasets:
5
  - PKU-ML/Erdos
6
  language:
7
  - en
8
+ library_name: transformers
9
+ license: apache-2.0
10
  metrics:
11
  - accuracy
12
+ pipeline_tag: graph-ml
 
 
13
  tags:
14
  - graph
15
  - chat
 
16
  ---
17
 
 
18
  # G1-Direct-SFT-7B
19
 
20
  ## Introduction
21
 
22
+ G1 is the series of large language models trained on our benchmark [Erdos](https://huggingface.co/datasets/PKU-ML/Erdos) for solving graph reasoning tasks, based on Qwen2.5-Instruct.
23
  We apply Group Relative Policy Optimization (GRPO) for reinforcement learning with supervised finetuning as a prelimary step.
24
 
25
  G1 brings the following improvements:
26
 
27
+ - **Significant improvement on graph reasoning**: G1 models achieve up to 46% improvement over baselines on Erdős, with the 7B variant matching OpenAI’s o3-mini and the 3B model surpassing Qwen2.5-72B-Instruct by notable margins.
28
+ - **Strong Generalization to unseen graph tasks**: G1 exhibits zero-shot generalization on unseen graph tasks, improving performance on *other graph reasoning benchmarks* (GraphWiz, GraphArena) and *real-world graphs* (Cora, PubMed).
29
+ - **NO Compromise on general reasoning**: Crucially, G1 preserves general reasoning ability (GSM8K, MATH, MMLU-Pro), proving its versatility.
30
 
31
 
32
  **This repo contains the G1-Direct-SFT-7B model**, which has the following features:
33
+ - Type: Causal Language Models
34
+ - Training Stage: SFT
35
+ - Architecture: the same with Qwen2.5-Instruct
36
+ - Number of Parameters: 7.62B
37
+ - Context Length: Full 32,768 tokens and generation 8192 tokens
38
 
39
  For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
40
 
 
71
  )
72
  tokenizer = AutoTokenizer.from_pretrained(model_name)
73
 
74
+ prompt = "The task is to determine the degree centrality of a node in the graph.
75
+
76
+ "\
77
+ "Degree centrality for a node is the fraction of nodes it is connected to.
78
+
79
+ "\
80
+ "Here is an undirected graph containing nodes from 1 to 15. The edges are: (1, 15), (15, 11), (2, 3), (2, 6), (3, 6), (3, 7), (6, 7), (6, 8), (7, 8), (7, 14), (4, 10), (10, 5), (10, 12), (8, 14), (8, 9), (12, 11), (12, 13).
81
+
82
+ "\
83
+ "Question: What is the degree centrality of node 2 in the graph?
84
+
85
+ "\
86
  "You need to format your answer as a float number."
87
  messages = [
88
  {"role": "user", "content": INSTRUCTION_TEMPLATE.format(instruction=prompt)}