yolay commited on
Commit
c2f9de6
·
verified ·
1 Parent(s): b2f5e9c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -6
README.md CHANGED
@@ -1,12 +1,9 @@
1
  ---
2
  base_model:
3
- - meta-llama/Llama-3.1-8B-Instruct
4
  datasets:
5
  - yolay/RAIF-ComplexInstruction-DeepSeek
6
- language:
7
- - en
8
  library_name: transformers
9
- license: apache-2.0
10
  metrics:
11
  - accuracy
12
  pipeline_tag: text-generation
@@ -14,8 +11,6 @@ pipeline_tag: text-generation
14
 
15
  This model belongs to the official implementation of the paper "Incentivizing Reasoning for Advanced Instruction-Following of Large Language Models".
16
 
17
- Project Page: https://yanqval.github.io/PAE/
18
-
19
  You can find the paper at https://huggingface.co/papers/2506.01413.
20
 
21
  Existing large language models (LLMs) face challenges of following complex instructions, especially when multiple constraints are present and organized in paralleling, chaining, and branching structures. One intuitive solution, namely chain-of-thought (CoT), is expected to universally improve capabilities of LLMs. However, we find that the vanilla CoT exerts a negative impact on performance due to its superficial reasoning pattern of simply paraphrasing the instructions. It fails to peel back the compositions of constraints for identifying their relationship across hierarchies of types and dimensions.
 
1
  ---
2
  base_model:
3
+ - deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
4
  datasets:
5
  - yolay/RAIF-ComplexInstruction-DeepSeek
 
 
6
  library_name: transformers
 
7
  metrics:
8
  - accuracy
9
  pipeline_tag: text-generation
 
11
 
12
  This model belongs to the official implementation of the paper "Incentivizing Reasoning for Advanced Instruction-Following of Large Language Models".
13
 
 
 
14
  You can find the paper at https://huggingface.co/papers/2506.01413.
15
 
16
  Existing large language models (LLMs) face challenges of following complex instructions, especially when multiple constraints are present and organized in paralleling, chaining, and branching structures. One intuitive solution, namely chain-of-thought (CoT), is expected to universally improve capabilities of LLMs. However, we find that the vanilla CoT exerts a negative impact on performance due to its superficial reasoning pattern of simply paraphrasing the instructions. It fails to peel back the compositions of constraints for identifying their relationship across hierarchies of types and dimensions.