nielsr HF Staff commited on
Commit
7e65a70
·
verified ·
1 Parent(s): 1b6df59

Add paper link and sample usage for Fill-in-the-Middle (FIM)

Browse files

Hi! I'm Niels from the Hugging Face community science team. I've opened this pull request to improve your model card.

Based on the official GitHub repository and the technical report, I've added a link to the paper and included a sample usage code snippet for the "Fill-in-the-Middle" (FIM) task. FIM is a key feature for code base models, allowing them to complete code given both a prefix and a suffix.

Files changed (1) hide show
  1. README.md +41 -1
README.md CHANGED
@@ -7,6 +7,9 @@ pipeline_tag: text-generation
7
 
8
  # Qwen3-Coder-Next-Base
9
 
 
 
 
10
  ## Highlights
11
 
12
  Today, we're announcing **Qwen3-Coder-Next-Base**, an open-weight language model designed specifically for coding agents and local development. It features the following key enhancements:
@@ -45,6 +48,43 @@ Today, we're announcing **Qwen3-Coder-Next-Base**, an open-weight language model
45
 
46
  For more details, including benchmark evaluation, hardware requirements, and inference performance, please refer to our [blog](https://qwen.ai/blog?id=qwen3-coder-next), [GitHub](https://github.com/QwenLM/Qwen3-Coder), and [Documentation](https://qwen.readthedocs.io/en/latest/).
47
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
  ## Best Practices
49
 
50
  To achieve optimal performance, we recommend the following sampling parameters: `temperature=1.0`, `top_p=0.95`, `top_k=40`.
@@ -54,7 +94,7 @@ To achieve optimal performance, we recommend the following sampling parameters:
54
 
55
  If you find our work helpful, feel free to give us a cite.
56
 
57
- ```
58
  @techreport{qwen_qwen3_coder_next_tech_report,
59
  title = {Qwen3-Coder-Next Technical Report},
60
  author = {{Qwen Team}},
 
7
 
8
  # Qwen3-Coder-Next-Base
9
 
10
+ ## Introduction
11
+ **Qwen3-Coder-Next-Base** is an open-weight language model designed specifically for coding agents and local development. It is the base version of the 80B parameter model that activates only 3B parameters during inference, as described in the [Qwen3-Coder-Next Technical Report](https://huggingface.co/papers/2603.00729).
12
+
13
  ## Highlights
14
 
15
  Today, we're announcing **Qwen3-Coder-Next-Base**, an open-weight language model designed specifically for coding agents and local development. It features the following key enhancements:
 
48
 
49
  For more details, including benchmark evaluation, hardware requirements, and inference performance, please refer to our [blog](https://qwen.ai/blog?id=qwen3-coder-next), [GitHub](https://github.com/QwenLM/Qwen3-Coder), and [Documentation](https://qwen.readthedocs.io/en/latest/).
50
 
51
+ ## Sample Usage
52
+
53
+ ### Fill in the middle with Qwen3-Coder
54
+
55
+ The code insertion task, also referred to as the "fill-in-the-middle" challenge, requires the insertion of code segments in a manner that bridges the gaps within a given code context.
56
+
57
+ ```python
58
+ from transformers import AutoTokenizer, AutoModelForCausalLM
59
+
60
+ model_id = "Qwen/Qwen3-Coder-Next-Base"
61
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
62
+ model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto").eval()
63
+
64
+ input_text = """<|fim_prefix|>def quicksort(arr):
65
+ if len(arr) <= 1:
66
+ return arr
67
+ pivot = arr[len(arr) // 2]
68
+ <|fim_suffix|>
69
+ middle = [x for x in arr if x == pivot]
70
+ right = [x for x in arr if x > pivot]
71
+ return quicksort(left) + middle + quicksort(right)<|fim_middle|>"""
72
+
73
+ model_inputs = tokenizer([input_text], return_tensors="pt").to(model.device)
74
+
75
+ # Use `max_new_tokens` to control the maximum output length.
76
+ # FIM specific special tokens:
77
+ eos_token_ids = [151659, 151661, 151662, 151663, 151664, 151643, 151645]
78
+ generated_ids = model.generate(model_inputs.input_ids, max_new_tokens=512, do_sample=False, eos_token_id=eos_token_ids)[0]
79
+
80
+ # The generated_ids include prompt_ids, we only need to decode the tokens after prompt_ids.
81
+ output_text = tokenizer.decode(generated_ids[len(model_inputs.input_ids[0]):], skip_special_tokens=True)
82
+
83
+ print(f"Prompt: {input_text}
84
+
85
+ Generated text: {output_text}")
86
+ ```
87
+
88
  ## Best Practices
89
 
90
  To achieve optimal performance, we recommend the following sampling parameters: `temperature=1.0`, `top_p=0.95`, `top_k=40`.
 
94
 
95
  If you find our work helpful, feel free to give us a cite.
96
 
97
+ ```bibtex
98
  @techreport{qwen_qwen3_coder_next_tech_report,
99
  title = {Qwen3-Coder-Next Technical Report},
100
  author = {{Qwen Team}},