siheng xiong commited on
Commit
14a4cd0
·
1 Parent(s): 84e9801

update README

Browse files
README.md CHANGED
@@ -9,4 +9,57 @@ base_model:
9
  ---
10
  # Model Card for TG-LLM
11
 
12
- These are the supervised fine-tuned models for temporal reasoning with LLMs. Please check [here](https://github.com/xiongsiheng/TG-LLM) for more details.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  ---
10
  # Model Card for TG-LLM
11
 
12
+ **TG-LLM** consists of supervised fine-tuned models designed for **temporal reasoning** with large language models (LLMs).
13
+ It includes two primary tasks:
14
+
15
+ 1. **Story-to-Temporal-Graph Translation (story_TG_trans)** – converting a narrative into its corresponding temporal graph.
16
+ 2. **Temporal-Graph Reasoning (TGR)** – reasoning over a given temporal graph to answer questions.
17
+
18
+
19
+ ## **Model Details**
20
+
21
+ #### **TGQA_story_TG_trans**
22
+
23
+ * **Base Model:** `meta-llama/Llama-2-13b-chat-hf`
24
+ * **LoRA Configuration:**
25
+
26
+ * `lora_alpha`: 8
27
+ * `r`: 8
28
+ * `target_modules`: `["q_proj", "k_proj", "o_proj", "v_proj"]`
29
+ * `bias`: `"none"`
30
+
31
+ #### **TGQA_TGR**
32
+
33
+ * **Base Model:** `meta-llama/Llama-2-13b-chat-hf`
34
+ * **LoRA Configuration:**
35
+
36
+ * `lora_alpha`: 8
37
+ * `r`: 8
38
+ * `target_modules`: `["q_proj", "k_proj", "o_proj", "v_proj"]`
39
+ * `bias`: `"none"`
40
+
41
+
42
+ For more details, please visit the [TG-LLM GitHub repository](https://github.com/xiongsiheng/TG-LLM).
43
+
44
+
45
+ ## Citation
46
+ ```
47
+ @inproceedings{xiong-etal-2024-large,
48
+ title = "Large Language Models Can Learn Temporal Reasoning",
49
+ author = "Xiong, Siheng and
50
+ Payani, Ali and
51
+ Kompella, Ramana and
52
+ Fekri, Faramarz",
53
+ editor = "Ku, Lun-Wei and
54
+ Martins, Andre and
55
+ Srikumar, Vivek",
56
+ booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
57
+ month = aug,
58
+ year = "2024",
59
+ address = "Bangkok, Thailand",
60
+ publisher = "Association for Computational Linguistics",
61
+ url = "https://aclanthology.org/2024.acl-long.563",
62
+ doi = "10.18653/v1/2024.acl-long.563",
63
+ pages = "10452--10470"
64
+ }
65
+ ```
TGQA_TGR/final/README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: ~/.cache/huggingface/hub/models--meta-llama--Llama-2-13b-chat-hf
3
  library_name: peft
4
  ---
5
 
 
1
  ---
2
+ base_model: meta-llama/Llama-2-13b-chat-hf
3
  library_name: peft
4
  ---
5
 
TGQA_TGR/final/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "/root/autodl-tmp/cache/huggingface/hub/models--meta-llama--Llama-2-13b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
TGQA_story_TG_trans/final/README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: ~/.cache/huggingface/hub/models--meta-llama--Llama-2-13b-chat-hf
3
  library_name: peft
4
  ---
5
 
 
1
  ---
2
+ base_model: meta-llama/Llama-2-13b-chat-hf
3
  library_name: peft
4
  ---
5
 
TGQA_story_TG_trans/final/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "/root/autodl-tmp/cache/huggingface/hub/models--meta-llama--Llama-2-13b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,