|
|
--- |
|
|
license: other |
|
|
datasets: |
|
|
- thu-coai/cdconv |
|
|
language: |
|
|
- en |
|
|
- zh |
|
|
metrics: |
|
|
- accuracy |
|
|
- precision |
|
|
- recall |
|
|
- f1 |
|
|
base_model: |
|
|
- google-bert/bert-base-multilingual-cased |
|
|
pipeline_tag: text-classification |
|
|
library_name: transformers |
|
|
tags: |
|
|
- manipulative-language |
|
|
- social-psychology |
|
|
--- |
|
|
|
|
|
# Model Card for mBERT Manipulative Language Detector |
|
|
|
|
|
本模型用于检测中文和英文文本中的**操纵性语言**(Manipulative Language),例如隐性控制、情感勒索、语言操控等,广泛应用于社交心理、文本筛查和内容审核等场景。 |
|
|
|
|
|
## 🧠 Model Details |
|
|
|
|
|
* **Developed by:** LilithHu |
|
|
* **Finetuned from:** google-bert/bert-base-multilingual-cased |
|
|
* **Languages:** 中文、英文 |
|
|
* **License:** other |
|
|
* **Model type:** 文本分类模型(binary classifier: manipulative / non-manipulative) |
|
|
|
|
|
## 🔧 Uses |
|
|
|
|
|
### ✅ Direct Use |
|
|
|
|
|
* 输入一段文本,模型将返回该文本是否包含操纵性语言。 |
|
|
* 可通过 Hugging Face Inference API 或 Web UI(Streamlit)直接调用。 |
|
|
|
|
|
### 👥 Intended Users |
|
|
|
|
|
* NLP 研究者 |
|
|
* 内容审核从业者 |
|
|
* 心理学研究人员 |
|
|
* 社交平台或对话系统开发者 |
|
|
|
|
|
### 🚫 Out-of-Scope Use |
|
|
|
|
|
* 本模型**不适合**用于: |
|
|
|
|
|
* 法律审判 |
|
|
* 医疗诊断 |
|
|
* 精准营销等高风险商业行为 |
|
|
* 判定他人动机、人格或情感 |
|
|
|
|
|
## ⚠️ Bias, Risks and Limitations |
|
|
|
|
|
请注意: |
|
|
|
|
|
* 模型输出不等于事实,仅基于训练数据的模式进行分类 |
|
|
* 操纵性语言的判断带有一定主观性与文化偏差 |
|
|
* 不应被用于评判具体个人、情感或行为正当性 |
|
|
|
|
|
### ✅ 建议 |
|
|
|
|
|
使用者应结合人工判断,多模态、多渠道地理解文本含义。对于模型预测结果不可盲信,应视为辅助工具。 |
|
|
|
|
|
## 🚀 How to Use |
|
|
|
|
|
```python |
|
|
from transformers import pipeline |
|
|
classifier = pipeline("text-classification", model="LilithHu/mbert-manipulative-detector") |
|
|
result = classifier("我爱你") |
|
|
print(result) |
|
|
``` |
|
|
|
|
|
也可通过终端调用: |
|
|
|
|
|
```bash |
|
|
curl -X POST https://api-inference.huggingface.co/models/LilithHu/mbert-manipulative-detector \ |
|
|
-H "Authorization: Bearer <your_hf_token>" \ |
|
|
-H "Content-Type: application/json" \ |
|
|
-d '{"inputs": "我爱你"}' |
|
|
``` |
|
|
|
|
|
## 🏋️ Training Details |
|
|
|
|
|
### 📚 Training Data |
|
|
|
|
|
* [CDial-GPT/toy\_valid](https://github.com/thu-coai/CDial-GPT/blob/master/data/toy_valid.txt) |
|
|
* thu-coai/esconv、cdconv 数据集 |
|
|
* 自建中文操纵性语言语料(未公开) |
|
|
|
|
|
### ⚙️ Training Procedure |
|
|
|
|
|
* 训练平台:Google Colab,GPU:T4 |
|
|
* Epochs: 3 |
|
|
* Batch size: 32 |
|
|
* Optimizer: AdamW |
|
|
* LR: 2e-5 |
|
|
|
|
|
## 📊 Evaluation |
|
|
|
|
|
| Metric | Score | |
|
|
| --------- | ----- | |
|
|
| Accuracy | 0.** | |
|
|
| Precision | 0.** | |
|
|
| Recall | 0.** | |
|
|
| F1-score | 0.** | |
|
|
|
|
|
## 🌍 Environmental Impact |
|
|
|
|
|
* 训练时间约 3 小时,使用 Google Colab GPU(T4) |
|
|
* 估算碳排放 < 2kg CO2eq |
|
|
|
|
|
## 🔒 Disclaimer |
|
|
|
|
|
* 本模型用于**研究与教育用途**,不得作为法律、道德、医疗或商业判断依据。 |
|
|
* 预测结果仅为参考,使用者需自行承担风险。 |
|
|
* 请勿用于恶意攻击、舆情操纵或误导他人行为。 |
|
|
|
|
|
## 📌 Model Card Authors |
|
|
|
|
|
LilithHu |
|
|
|
|
|
## 📬 Contact |
|
|
|
|
|
如需反馈建议,请通过 Hugging Face 留言联系作者。 |
|
|
|
|
|
## 📚 Citation |
|
|
|
|
|
```bibtex |
|
|
@misc{LilithHu2025, |
|
|
title={mBERT Manipulative Language Detector}, |
|
|
author={LilithHu}, |
|
|
year={2025}, |
|
|
url={https://huggingface.co/LilithHu/mbert-manipulative-detector} |
|
|
} |
|
|
``` |