Assignment_3 / README.md
Sristtee's picture
Upload README.md
500f1bd verified
---
tags:
- patents
- climate
- green-technology
- text-classification
- patent-classification
- human-in-the-loop
- multi-agent
- patentsberta
language:
- en
pipeline_tag: text-classification
library_name: transformers
---
# Green Patent Detection: Multi-Agent HITL + PatentSBERTa
This repository contains an advanced green patent detection workflow built for **binary classification of patent claims** into:
- **1 = Green / climate mitigation related**
- **0 = Non-green**
The project extends a baseline PatentSBERTa workflow by adding a **Human-in-the-Loop (HITL)** review stage and a **multi-agent debate system** before final fine-tuning.
## Project overview
The goal of this project is to improve green patent detection by combining:
1. **High-risk sample selection** from uncertainty sampling
2. **Multi-agent LLM review** of difficult claims
3. **Human verification** of the AI suggestions
4. **Final fine-tuning of PatentSBERTa** using silver labels + gold HITL labels
This workflow was designed to test whether a more advanced labeling pipeline produces stronger training data than a simple single-LLM labeling approach.
## Base model
The final classifier is built from:
- **Base encoder:** `AI-Growth-Lab/PatentSBERTa`
- **Task:** Binary text classification
- **Domain:** Patent claim classification for climate mitigation / green technology
## Data used in the notebook
The notebook uses the following files:
- `patents_50k_green.parquet`
- `train_meta.csv`
- `y_train.npy`
- `eval_silver.parquet`
- `hitl_green_100.csv`
- `hitl_review_progress_with_llm.csv`
- `hitl_green_gold.csv`
- `hitl_three_agents.csv`
## Methodology
### 1. High-risk claim selection
A set of **100 high-risk patent claims** was selected from earlier uncertainty sampling outputs. These were the most difficult / ambiguous examples for the model.
### 2. Multi-agent debate system
Three agents were created using `CrewAI` and an Ollama-hosted model (`qwen2.5:3b-instruct`):
- **Advocate Agent** – argues why the claim should be classified as green under Y02 climate mitigation logic
- **Skeptic Agent** – argues why the claim may not qualify and checks for weak evidence or greenwashing
- **Judge Agent** – reviews both sides and returns a structured final output with:
- predicted label
- confidence
- rationale
This produces an AI suggestion for each difficult claim.
### 3. Human-in-the-Loop review
The AI-generated suggestion was then manually reviewed.
The final human label was stored as:
- `is_green_human`
These human-reviewed labels form the **gold dataset** for the difficult claims.
### 4. Gold-enhanced training
The final training set combines:
- **Silver labels** from the earlier training data
- **100 gold human-reviewed claims** from the multi-agent workflow
This combined dataset was used to fine-tune PatentSBERTa.
## Training configuration
The notebook fine-tunes the model with the following setup:
- **Model:** `AI-Growth-Lab/PatentSBERTa`
- **Max sequence length:** `256`
- **Epochs:** `1`
- **Learning rate:** `2e-5`
- **Train batch size:** `8`
- **Eval batch size:** `8`
- **Weight decay:** `0.01`
- **Framework:** Hugging Face Transformers Trainer
## Dataset splits used during fine-tuning
From the notebook:
- **Training data:** silver training set + gold HITL labels
- **Evaluation data:** `eval_silver`
- **Additional check:** `gold_100`
The notebook text states that the final training dataset contains **35,200 claims**.
## Human vs AI agreement
According to the notebook:
- **Simple LLM from Assignment 2:** `94%` agreement with human labels
- **Agentic system from Assignment 3:** `87%` agreement with human labels
This suggests that the multi-agent system used stricter reasoning criteria, which created more disagreement on borderline cases.
## Repository contents
Depending on what you upload, this repository may include:
- the processed HITL dataset
- the final trained model
- tokenizer files
- training notebook
- prediction / rationale outputs for the 100 reviewed claims
## Expected columns in the HITL dataset
The notebook shows or creates columns such as:
- `id`
- `text`
- `p_green`
- `u`
- `llm_green_suggested`
- `llm_confidence`
- `llm_rationale`
- `is_green_human`
## Example use
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model_name = "YOUR_HF_REPO_NAME"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)
text = "A company develops a carbon capture system that reduces CO2 emissions from cement factories."
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=256)
with torch.no_grad():
logits = model(**inputs).logits
pred = torch.argmax(logits, dim=-1).item()
print("Predicted label:", pred)
```
## Intended use
This project is intended for:
- research and coursework on green patent detection
- experimentation with HITL labeling pipelines
- comparison of simple vs advanced AI-assisted annotation workflows
- climate-tech related document classification
## Limitations
- The gold set is relatively small (**100 reviewed claims**)
- The multi-agent workflow depends on LLM reasoning quality
- Agreement with humans does not automatically guarantee better downstream model performance
- Final performance metrics should be reported from the actual training run in this repository
## Notes
This README was prepared from the notebook workflow and code structure. If you are uploading the **model repo**, add the final evaluation metrics from your training output. If you are uploading the **dataset repo**, you can keep the methodology sections and remove the model inference example if not needed.
## Citation
If you use this work, please cite the repository and the base model:
- `AI-Growth-Lab/PatentSBERTa`
You may also describe the workflow as:
> Multi-Agent Human-in-the-Loop green patent detection using PatentSBERTa with gold-enhanced fine-tuning.