|
|
--- |
|
|
license: cc-by-nc-4.0 |
|
|
tags: |
|
|
- bert |
|
|
- text-classification |
|
|
- disability |
|
|
- inclusive-language |
|
|
- academic-writing |
|
|
datasets: |
|
|
- assets |
|
|
library_name: transformers |
|
|
language: |
|
|
- en |
|
|
--- |
|
|
# Identifying Disability-Insensitive Language in Scholarly Works |
|
|
|
|
|
Refer to the code repository and paper here: [GitHub - Insensitive-Lang-Detection](https://github.com/RobyRoshna/Insensitive-Lang-Detection/tree/main) |
|
|
|
|
|
--- |
|
|
|
|
|
## Overview |
|
|
This is a fine-tuned BERT model designed to detect potentially insensitive or non-inclusive language relating to disability, specifically in academic and scholarly writing. |
|
|
|
|
|
The model helps promote more inclusive and respectful communication, aligning with social models of disability and various international guidelines. |
|
|
|
|
|
--- |
|
|
|
|
|
## Intended Use |
|
|
- Academic editors and reviewers who want to check abstracts and papers for disability-insensitive language. |
|
|
- Researchers studying accessibility, inclusive design, or language bias. |
|
|
- Automated writing support tools focused on scholarly communication. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Details |
|
|
- **Architecture**: BERT-base (uncased) |
|
|
- **Fine-tuned on**: Sentences from ASSETS conference papers (1994–2024) and organizational documents (ADA National Network, UN guidelines). |
|
|
- **Labels**: |
|
|
- `0`: Not insensitive |
|
|
- `1`: Insensitive |
|
|
|
|
|
--- |
|
|
|
|
|
## Training Data |
|
|
- Extracted and manually annotated sentences referencing disability-related terms. |
|
|
- Supported with data augmentation using OpenAI GPT-4o to balance underrepresented phrases. |
|
|
|
|
|
--- |
|
|
|
|
|
## License |
|
|
This model is licensed under the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) license. |
|
|
|
|
|
This means you are free to share and adapt the model for non-commercial purposes, as long as appropriate credit is given. Commercial use is not permitted without explicit permission. |
|
|
|
|
|
For details, see [CC BY-NC 4.0](https://creativecommons.org/licenses/by-nc/4.0/). |
|
|
|
|
|
--- |
|
|
## How to Use |
|
|
|
|
|
```python |
|
|
from transformers import BertForSequenceClassification, BertTokenizer |
|
|
|
|
|
model = BertForSequenceClassification.from_pretrained("rrroby/insensitive-language-bert") |
|
|
tokenizer = BertTokenizer.from_pretrained("rrroby/insensitive-language-bert") |
|
|
|
|
|
text = "This participant was wheelchair-bound and..." |
|
|
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=512) |
|
|
outputs = model(**inputs) |
|
|
logits = outputs.logits |
|
|
predicted_class = logits.argmax(-1).item() |
|
|
|
|
|
print("Predicted class:", predicted_class) |
|
|
|
|
|
|
|
|
|