| base_model: answerdotai/ModernBERT-base | |
| library_name: transformers | |
| pipeline_tag: text-classification | |
| tags: | |
| - text-classification | |
| - regression | |
| - legal | |
| - locus | |
| - modernbert | |
| license: apache-2.0 | |
| datasets: | |
| - LocalLaws/LOCUS-v1.0 | |
| # LocalLaws/LOCUS-Problem-Salience | |
| A ModernBERT regression model that scores local-ordinance text along the | |
| **Problem Salience** axis of the LOCUS (Local Ordinances Corpus, United States) dataset. | |
| Fine-tuned from [answerdotai/ModernBERT-base](https://huggingface.co/answerdotai/ModernBERT-base). The | |
| target is a TrueSkill `mu` distilled from pairwise LLM comparisons on the | |
| problem salience axis, then z-score normalized across the training corpus. | |
| ## Usage | |
| ```python | |
| from transformers import AutoTokenizer, AutoModelForSequenceClassification | |
| import torch | |
| tok = AutoTokenizer.from_pretrained("LocalLaws/LOCUS-Problem-Salience") | |
| model = AutoModelForSequenceClassification.from_pretrained("LocalLaws/LOCUS-Problem-Salience") | |
| model.eval() | |
| text = "No person shall keep any swine within the city limits." | |
| enc = tok(text, return_tensors="pt", truncation=True, max_length=2048) | |
| with torch.no_grad(): | |
| score = model(**enc).logits.squeeze(-1).item() | |
| print(score) | |
| ``` | |