nielsr HF Staff commited on
Commit
75fcf7b
·
verified ·
1 Parent(s): 369800e

Improve model card: Add pipeline tag, library name, paper link, and GitHub code link

Browse files

This PR enhances the model card for DeepSeek-V3.1-Nex-N1 by:
- Adding `pipeline_tag: text-generation` to categorize the model accurately, improving discoverability on the Hugging Face Hub.
- Adding `library_name: transformers` as the model architecture (`DeepseekV3ForCausalLM` in `config.json`) indicates compatibility with the Transformers library, enabling the automated code snippet on the Hub.
- Updating the "Tech Report" link in the top navigation to point to the official Hugging Face paper page: [Nex-N1: Agentic Models Trained via a Unified Ecosystem for Large-Scale Environment Construction](https://huggingface.co/papers/2512.04987).
- Adding a direct link to the GitHub code repository ([https://github.com/nex-agi/Nex-N1](https://github.com/nex-agi/Nex-N1)) in the top navigation for improved access to the codebase.

Files changed (1) hide show
  1. README.md +101 -98
README.md CHANGED
@@ -1,98 +1,101 @@
1
- ---
2
- license: apache-2.0
3
- ---
4
-
5
- <div align="center">
6
- <img src="./figures/NEX_logo.svg" width="20%"/>
7
- </div>
8
-
9
- ---
10
-
11
- <div align="center">
12
- 🏠 <a href="https://nex.sii.edu.cn"><b>Home&nbspPage</b></a>&nbsp&nbsp | &nbsp&nbsp
13
- 🤗 <a href="https://hf.co/collections/nex-agi/nex-n1"><b>Model</b></a>&nbsp&nbsp | &nbsp&nbsp
14
- 🤗 <a href="https://huggingface.co/datasets/nex-agi/agent-sft"><b>Data</b></a>&nbsp&nbsp | &nbsp&nbsp
15
- 📑 <a href="https://github.com/nex-agi/Nex-N1/blob/main/Nex-N1-TechReport.pdf"><b>Tech&nbspReport</b></a>&nbsp&nbsp
16
- </div>
17
-
18
- # Nex-N1
19
-
20
- Nex is a next-generation, full-stack agentic platform that brings foundation models, synthetic data pipelines, RL training, agent frameworks, and deployment tools together in one unified ecosystem.
21
- DeepSeek-V3.1-Nex-N1 is the flagship release of the Nex-N1 series — a post-trained model designed to highlight agent autonomy, tool use, and real-world productivity.
22
- We are committed to making it easier than ever to build and deploy AI agents by offering researchers and entrepreneurs a high-performance, reliable, and cost-effective "out-of-the-box" agent system.
23
-
24
- ## Highlights
25
-
26
- - **Full spectrum model matrix:** From 8B to 671B parameters, the Nex series covers everything from edge-friendly setups to frontier-scale deployments.
27
- - **Agent-focused performance:** Demonstrates industry-leading results on programming, tool-use, web-search, and other multi-hop reasoning tasks.
28
- - **Production-ready utility:** Excels at mini-app development, website authoring, slide creation, and immersive role-play—delivering immediate productivity
29
- gains.
30
- - **End-to-end control:** Developers can build the entire data-to-deployment loop on top of Nex, ensuring sovereignty while keeping costs predictable.
31
- - **Open ecosystem:** Turnkey synthetic data pipelines, curated datasets, Nex-N1 checkpoints, the NexAU Agent framework, the EaaS MoE inference stack, and NexRL
32
- training services are all openly available.
33
-
34
- ## Performance
35
-
36
- Nex-N1 is evaluated on six representative agentic benchmarks (general + professional). The model consistently ranks at or near the top across tool-using, web-search, and coding-heavy evaluations, showing strong readiness for real-world agent workflows.
37
-
38
- ![Nex-N1 Benchmark Overview](./figures/Nex-N1-Benchamrk-white.png)
39
-
40
- <ul align="left" style="font-size:12px; color:#6c757d;">
41
- <li>Data points are sourced by default from the model’s official technical report or blog, as well as the benchmark’s official results. All other metrics were tested in strict compliance with the official standard evaluation framework.</li>
42
- <li>Results for Tau2-bench are derived via a weighted average.</li>
43
- <li>For SWE-verified-bench, test results are based on an internal scaffold built with OpenHands—using a 128k context length and 150 maximum steps—and represent the average of four runs.</li>
44
- <li>Terminal-Bench2 is evaluated using the official Terminus2 agent.</li>
45
- </ul>
46
-
47
- Nex-N1 provides various size models from 8B to 671B for different usage scenarios.
48
-
49
- | Model | GAIA2 | τ2-Bench | SWE-bench Verified | Terminal-Bench2 | BaxBench | BFCL v4 |
50
- | --- | --- | --- | --- | --- | --- | --- |
51
- | [DeepSeek-V3.1-Nex-N1](https://huggingface.co/nex-agi/DeepSeek-V3.1-Nex-N1) | 29.5 | 80.2 | 70.6 | 31.8 | 59.7 | 65.3 |
52
- | [Qwen3-32B-Nex-N1](https://huggingface.co/nex-agi/Qwen3-32B-Nex-N1) | 16.7 | 72.1 | 50.5 | 16.7 | 34.8 | 60.5 |
53
- | [Qwen3-30B-A3B-Nex-N1](https://huggingface.co/nex-agi/Qwen3-30B-A3B-Nex-N1) | 11.3 | 65.3 | 29.7 | 8.3 | 13.6 | 51.9 |
54
- | [internlm3-8B-Nex-N1](https://huggingface.co/nex-agi/internlm3-8B-Nex-N1) | 8.6 | 63.0 | 20.3 | - | - | 44.5 |
55
-
56
- Nex-N1 demonstrates competitive performance across all evaluation scenarios, showing particularly strong results in practical coding and HTML generation tasks.
57
-
58
- <div align="center">
59
- <img src="./figures/coding-eval.png" width="80%"/>
60
- <div>Practical Coding Evaluation</div>
61
- </div>
62
-
63
- <div align="center">
64
- <img src="./figures/html-eval.png" width="80%"/>
65
- <div>HTML Generation Evaluation</div>
66
- </div>
67
-
68
- Refer to <https://huggingface.co/datasets/nex-agi/coding-eval> and <https://huggingface.co/datasets/nex-agi/html-eval> for more details.
69
-
70
- ## Usage
71
-
72
- ### Local Deployment
73
-
74
- We recommend `sglang` for serving Nex-series models locally:
75
-
76
- ```bash
77
- python -m sglang.launch_server --model-path /path/to/your/model
78
- ```
79
-
80
- ### Function Calling
81
-
82
- Nex-series models support robust function-calling capabilities. To maximize the function-calling capabilities of the Nex-series models, we modified the tool parser of `qwen3_coder`, see: <https://github.com/sgl-project/sglang/pull/13411>. To enable this feature, simply add the `--tool-call-parser qwen3_coder` flag when launching the server:
83
-
84
- ```bash
85
- python -m sglang.launch_server --model-path /path/to/your/model --tool-call-parser qwen3_coder
86
- ```
87
-
88
- ### Mini Program Development
89
-
90
- Nex-N1 is optimized for mini program development. For optimal performance, we recommend using Claude Code configured with both `context7` and a search MCP.
91
-
92
- ```shell
93
- claude mcp add --transport http context7 https://mcp.context7.com/mcp --header "CONTEXT7_API_KEY: [CONTEXT7_API_KEY]"
94
-
95
- claude mcp add --transport stdio serper-search --env SERPER_API_KEY=[SERPER_API_KEY] -- npx -y serper-search-scrape-mcp-server
96
- ```
97
-
98
- Refer to <https://github.com/upstash/context7> for more details on setting up `context7`.
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: text-generation
4
+ library_name: transformers
5
+ ---
6
+
7
+ <div align="center">
8
+ <img src="./figures/NEX_logo.svg" width="20%"/>
9
+ </div>
10
+
11
+ ---
12
+
13
+ <div align="center">
14
+ 🏠 <a href="https://nex.sii.edu.cn"><b>Home&nbspPage</b></a>&nbsp&nbsp | &nbsp&nbsp
15
+ 🤗 <a href="https://hf.co/collections/nex-agi/nex-n1"><b>Model</b></a>&nbsp&nbsp | &nbsp&nbsp
16
+ 🤗 <a href="https://huggingface.co/datasets/nex-agi/agent-sft"><b>Data</b></a>&nbsp&nbsp | &nbsp&nbsp
17
+ 📚 <a href="https://huggingface.co/papers/2512.04987"><b>Paper</b></a>&nbsp&nbsp | &nbsp&nbsp
18
+ 💻 <a href="https://github.com/nex-agi/Nex-N1"><b>Code</b></a>
19
+ </div>
20
+
21
+ # Nex-N1
22
+
23
+ Nex is a next-generation, full-stack agentic platform that brings foundation models, synthetic data pipelines, RL training, agent frameworks, and deployment tools together in one unified ecosystem.
24
+ DeepSeek-V3.1-Nex-N1 is the flagship release of the Nex-N1 series — a post-trained model designed to highlight agent autonomy, tool use, and real-world productivity.
25
+ We are committed to making it easier than ever to build and deploy AI agents by offering researchers and entrepreneurs a high-performance, reliable, and cost-effective "out-of-the-box" agent system.
26
+
27
+ ## Highlights
28
+
29
+ - **Full spectrum model matrix:** From 8B to 671B parameters, the Nex series covers everything from edge-friendly setups to frontier-scale deployments.
30
+ - **Agent-focused performance:** Demonstrates industry-leading results on programming, tool-use, web-search, and other multi-hop reasoning tasks.
31
+ - **Production-ready utility:** Excels at mini-app development, website authoring, slide creation, and immersive role-play—delivering immediate productivity
32
+ gains.
33
+ - **End-to-end control:** Developers can build the entire data-to-deployment loop on top of Nex, ensuring sovereignty while keeping costs predictable.
34
+ - **Open ecosystem:** Turnkey synthetic data pipelines, curated datasets, Nex-N1 checkpoints, the NexAU Agent framework, the EaaS MoE inference stack, and NexRL
35
+ training services are all openly available.
36
+
37
+ ## Performance
38
+
39
+ Nex-N1 is evaluated on six representative agentic benchmarks (general + professional). The model consistently ranks at or near the top across tool-using, web-search, and coding-heavy evaluations, showing strong readiness for real-world agent workflows.
40
+
41
+ ![Nex-N1 Benchmark Overview](./figures/Nex-N1-Benchamrk-white.png)
42
+
43
+ <ul align="left" style="font-size:12px; color:#6c757d;">
44
+ <li>Data points are sourced by default from the model’s official technical report or blog, as well as the benchmark’s official results. All other metrics were tested in strict compliance with the official standard evaluation framework.</li>
45
+ <li>Results for Tau2-bench are derived via a weighted average.</li>
46
+ <li>For SWE-verified-bench, test results are based on an internal scaffold built with OpenHands—using a 128k context length and 150 maximum steps—and represent the average of four runs.</li>
47
+ <li>Terminal-Bench2 is evaluated using the official Terminus2 agent.</li>
48
+ </ul>
49
+
50
+ Nex-N1 provides various size models from 8B to 671B for different usage scenarios.
51
+
52
+ | Model | GAIA2 | τ2-Bench | SWE-bench Verified | Terminal-Bench2 | BaxBench | BFCL v4 |
53
+ | --- | --- | --- | --- | --- | --- | --- |
54
+ | [DeepSeek-V3.1-Nex-N1](https://huggingface.co/nex-agi/DeepSeek-V3.1-Nex-N1) | 29.5 | 80.2 | 70.6 | 31.8 | 59.7 | 65.3 |
55
+ | [Qwen3-32B-Nex-N1](https://huggingface.co/nex-agi/Qwen3-32B-Nex-N1) | 16.7 | 72.1 | 50.5 | 16.7 | 34.8 | 60.5 |
56
+ | [Qwen3-30B-A3B-Nex-N1](https://huggingface.co/nex-agi/Qwen3-30B-A3B-Nex-N1) | 11.3 | 65.3 | 29.7 | 8.3 | 13.6 | 51.9 |
57
+ | [internlm3-8B-Nex-N1](https://huggingface.co/nex-agi/internlm3-8B-Nex-N1) | 8.6 | 63.0 | 20.3 | - | - | 44.5 |
58
+
59
+ Nex-N1 demonstrates competitive performance across all evaluation scenarios, showing particularly strong results in practical coding and HTML generation tasks.
60
+
61
+ <div align="center">
62
+ <img src="./figures/coding-eval.png" width="80%"/>
63
+ <div>Practical Coding Evaluation</div>
64
+ </div>
65
+
66
+ <div align="center">
67
+ <img src="./figures/html-eval.png" width="80%"/>
68
+ <div>HTML Generation Evaluation</div>
69
+ </div>
70
+
71
+ Refer to <https://huggingface.co/datasets/nex-agi/coding-eval> and <https://huggingface.co/datasets/nex-agi/html-eval> for more details.
72
+
73
+ ## Usage
74
+
75
+ ### Local Deployment
76
+
77
+ We recommend `sglang` for serving Nex-series models locally:
78
+
79
+ ```bash
80
+ python -m sglang.launch_server --model-path /path/to/your/model
81
+ ```
82
+
83
+ ### Function Calling
84
+
85
+ Nex-series models support robust function-calling capabilities. To maximize the function-calling capabilities of the Nex-series models, we modified the tool parser of `qwen3_coder`, see: <https://github.com/sgl-project/sglang/pull/13411>. To enable this feature, simply add the `--tool-call-parser qwen3_coder` flag when launching the server:
86
+
87
+ ```bash
88
+ python -m sglang.launch_server --model-path /path/to/your/model --tool-call-parser qwen3_coder
89
+ ```
90
+
91
+ ### Mini Program Development
92
+
93
+ Nex-N1 is optimized for mini program development. For optimal performance, we recommend using Claude Code configured with both `context7` and a search MCP.
94
+
95
+ ```shell
96
+ claude mcp add --transport http context7 https://mcp.context7.com/mcp --header "CONTEXT7_API_KEY: [CONTEXT7_API_KEY]"
97
+
98
+ claude mcp add --transport stdio serper-search --env SERPER_API_KEY=[SERPER_API_KEY] -- npx -y serper-search-scrape-mcp-server
99
+ ```
100
+
101
+ Refer to <https://github.com/upstash/context7> for more details on setting up `context7`.