| --- |
| tags: |
| - patents |
| - climate |
| - green-technology |
| - text-classification |
| - patent-classification |
| - human-in-the-loop |
| - multi-agent |
| - patentsberta |
| language: |
| - en |
| pipeline_tag: text-classification |
| library_name: transformers |
| --- |
| |
| # Green Patent Detection: Multi-Agent HITL + PatentSBERTa |
|
|
| This repository contains an advanced green patent detection workflow built for **binary classification of patent claims** into: |
|
|
| - **1 = Green / climate mitigation related** |
| - **0 = Non-green** |
|
|
| The project extends a baseline PatentSBERTa workflow by adding a **Human-in-the-Loop (HITL)** review stage and a **multi-agent debate system** before final fine-tuning. |
|
|
| ## Project overview |
|
|
| The goal of this project is to improve green patent detection by combining: |
|
|
| 1. **High-risk sample selection** from uncertainty sampling |
| 2. **Multi-agent LLM review** of difficult claims |
| 3. **Human verification** of the AI suggestions |
| 4. **Final fine-tuning of PatentSBERTa** using silver labels + gold HITL labels |
|
|
| This workflow was designed to test whether a more advanced labeling pipeline produces stronger training data than a simple single-LLM labeling approach. |
|
|
| ## Base model |
|
|
| The final classifier is built from: |
|
|
| - **Base encoder:** `AI-Growth-Lab/PatentSBERTa` |
| - **Task:** Binary text classification |
| - **Domain:** Patent claim classification for climate mitigation / green technology |
|
|
| ## Data used in the notebook |
|
|
| The notebook uses the following files: |
|
|
| - `patents_50k_green.parquet` |
| - `train_meta.csv` |
| - `y_train.npy` |
| - `eval_silver.parquet` |
| - `hitl_green_100.csv` |
| - `hitl_review_progress_with_llm.csv` |
| - `hitl_green_gold.csv` |
| - `hitl_three_agents.csv` |
|
|
| ## Methodology |
|
|
| ### 1. High-risk claim selection |
|
|
| A set of **100 high-risk patent claims** was selected from earlier uncertainty sampling outputs. These were the most difficult / ambiguous examples for the model. |
|
|
| ### 2. Multi-agent debate system |
|
|
| Three agents were created using `CrewAI` and an Ollama-hosted model (`qwen2.5:3b-instruct`): |
|
|
| - **Advocate Agent** – argues why the claim should be classified as green under Y02 climate mitigation logic |
| - **Skeptic Agent** – argues why the claim may not qualify and checks for weak evidence or greenwashing |
| - **Judge Agent** – reviews both sides and returns a structured final output with: |
| - predicted label |
| - confidence |
| - rationale |
|
|
| This produces an AI suggestion for each difficult claim. |
|
|
| ### 3. Human-in-the-Loop review |
|
|
| The AI-generated suggestion was then manually reviewed. |
|
|
| The final human label was stored as: |
|
|
| - `is_green_human` |
|
|
| These human-reviewed labels form the **gold dataset** for the difficult claims. |
|
|
| ### 4. Gold-enhanced training |
|
|
| The final training set combines: |
|
|
| - **Silver labels** from the earlier training data |
| - **100 gold human-reviewed claims** from the multi-agent workflow |
|
|
| This combined dataset was used to fine-tune PatentSBERTa. |
|
|
| ## Training configuration |
|
|
| The notebook fine-tunes the model with the following setup: |
|
|
| - **Model:** `AI-Growth-Lab/PatentSBERTa` |
| - **Max sequence length:** `256` |
| - **Epochs:** `1` |
| - **Learning rate:** `2e-5` |
| - **Train batch size:** `8` |
| - **Eval batch size:** `8` |
| - **Weight decay:** `0.01` |
| - **Framework:** Hugging Face Transformers Trainer |
|
|
| ## Dataset splits used during fine-tuning |
|
|
| From the notebook: |
|
|
| - **Training data:** silver training set + gold HITL labels |
| - **Evaluation data:** `eval_silver` |
| - **Additional check:** `gold_100` |
|
|
| The notebook text states that the final training dataset contains **35,200 claims**. |
|
|
| ## Human vs AI agreement |
|
|
| According to the notebook: |
|
|
| - **Simple LLM from Assignment 2:** `94%` agreement with human labels |
| - **Agentic system from Assignment 3:** `87%` agreement with human labels |
|
|
| This suggests that the multi-agent system used stricter reasoning criteria, which created more disagreement on borderline cases. |
|
|
| ## Repository contents |
|
|
| Depending on what you upload, this repository may include: |
|
|
| - the processed HITL dataset |
| - the final trained model |
| - tokenizer files |
| - training notebook |
| - prediction / rationale outputs for the 100 reviewed claims |
|
|
| ## Expected columns in the HITL dataset |
|
|
| The notebook shows or creates columns such as: |
|
|
| - `id` |
| - `text` |
| - `p_green` |
| - `u` |
| - `llm_green_suggested` |
| - `llm_confidence` |
| - `llm_rationale` |
| - `is_green_human` |
|
|
| ## Example use |
|
|
| ```python |
| from transformers import AutoTokenizer, AutoModelForSequenceClassification |
| import torch |
| |
| model_name = "YOUR_HF_REPO_NAME" |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
| model = AutoModelForSequenceClassification.from_pretrained(model_name) |
| |
| text = "A company develops a carbon capture system that reduces CO2 emissions from cement factories." |
| inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=256) |
| |
| with torch.no_grad(): |
| logits = model(**inputs).logits |
| pred = torch.argmax(logits, dim=-1).item() |
| |
| print("Predicted label:", pred) |
| ``` |
|
|
| ## Intended use |
|
|
| This project is intended for: |
|
|
| - research and coursework on green patent detection |
| - experimentation with HITL labeling pipelines |
| - comparison of simple vs advanced AI-assisted annotation workflows |
| - climate-tech related document classification |
|
|
| ## Limitations |
|
|
| - The gold set is relatively small (**100 reviewed claims**) |
| - The multi-agent workflow depends on LLM reasoning quality |
| - Agreement with humans does not automatically guarantee better downstream model performance |
| - Final performance metrics should be reported from the actual training run in this repository |
|
|
| ## Notes |
|
|
| This README was prepared from the notebook workflow and code structure. If you are uploading the **model repo**, add the final evaluation metrics from your training output. If you are uploading the **dataset repo**, you can keep the methodology sections and remove the model inference example if not needed. |
|
|
| ## Citation |
|
|
| If you use this work, please cite the repository and the base model: |
|
|
| - `AI-Growth-Lab/PatentSBERTa` |
|
|
| You may also describe the workflow as: |
|
|
| > Multi-Agent Human-in-the-Loop green patent detection using PatentSBERTa with gold-enhanced fine-tuning. |
|
|