ZhouAlen commited on
Commit
ae12b3a
·
verified ·
1 Parent(s): deac1b8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -9
README.md CHANGED
@@ -10,10 +10,10 @@ library_name: transformers
10
  ## Introduction
11
 
12
  The Ming large language model (Ming‑LLM) is a domain‑specialized LLM for the energy sector.
13
- We release both the base model and the supervised fine‑tuned (SFT) variant.
14
- The Ming base model is initialized from the Qwen2.5‑72B base model and is subsequently adapted via continued pretraining on a high‑quality energy‑domain corpus.
15
- The SFT variant is initialized from the Ming base model and is trained on instruction‑tuning datasets, including conversational QA, sentiment analysis, and information extraction, among others.
16
- Both models demonstrate improved performance across the C‑Eval, CMMLU, MMLU, GSM8K, and IFEval benchmarks.
17
 
18
  ## Model Parameters
19
  Base model:
@@ -87,13 +87,13 @@ gen_ids = output_ids[0, inputs["input_ids"].shape[1]:]
87
  text = tokenizer.decode(gen_ids, skip_special_tokens=False)
88
  ```
89
  ## Bias, Risks, and Limitations
90
- Like any base language model or fine-tuned model without safety filtering, these models can easily be prompted by users to generate harmful and sensitive content.
91
- Such content may also be produced unintentionally, especially in cases involving bias, so we recommend that users consider the risks when applying this technology.
92
- Additionally, many statements from Ming Model or any LLM are often inaccurate, so facts should be verified.
93
 
94
  ## License and use
95
- Ming1.0 is built with Qwen-2.5-72B. Qwen-2.5-72B is licensed under the Qwen LICENSE AGREEMENT, Copyright (c) Alibaba Cloud. All Rights Reserved.
96
- Subject to the Qwen LICENSE AGREEMENT, Ming1.0 is under MIT license.
97
 
98
  ## Citation
99
  @
 
10
  ## Introduction
11
 
12
  The Ming large language model (Ming‑LLM) is a domain‑specialized LLM for the energy sector.
13
+ - We release both the base model and the supervised fine‑tuned (SFT) variant.
14
+ - The Ming base model is initialized from the Qwen2.5‑72B base model and is subsequently adapted via continued pretraining on a high‑quality energy‑domain corpus.
15
+ - The SFT variant is initialized from the Ming base model and is trained on instruction‑tuning datasets, including conversational QA, sentiment analysis, and information extraction, among others.
16
+ - Both models demonstrate improved performance across the C‑Eval, CMMLU, MMLU, GSM8K, and IFEval benchmarks.
17
 
18
  ## Model Parameters
19
  Base model:
 
87
  text = tokenizer.decode(gen_ids, skip_special_tokens=False)
88
  ```
89
  ## Bias, Risks, and Limitations
90
+ - Like any base language model or fine-tuned model without safety filtering, these models can easily be prompted by users to generate harmful and sensitive content.
91
+ - Such content may also be produced unintentionally, especially in cases involving bias, so we recommend that users consider the risks when applying this technology.
92
+ - Additionally, many statements from Ming Model or any LLM are often inaccurate, so facts should be verified.
93
 
94
  ## License and use
95
+ - Ming1.0 is built with Qwen-2.5-72B. Qwen-2.5-72B is licensed under the Qwen LICENSE AGREEMENT, Copyright (c) Alibaba Cloud. All Rights Reserved.
96
+ - Subject to the Qwen LICENSE AGREEMENT, Ming1.0 is under MIT license.
97
 
98
  ## Citation
99
  @