Assignment_3 / README.md
Sristtee's picture
Upload README.md
500f1bd verified
metadata
tags:
  - patents
  - climate
  - green-technology
  - text-classification
  - patent-classification
  - human-in-the-loop
  - multi-agent
  - patentsberta
language:
  - en
pipeline_tag: text-classification
library_name: transformers

Green Patent Detection: Multi-Agent HITL + PatentSBERTa

This repository contains an advanced green patent detection workflow built for binary classification of patent claims into:

  • 1 = Green / climate mitigation related
  • 0 = Non-green

The project extends a baseline PatentSBERTa workflow by adding a Human-in-the-Loop (HITL) review stage and a multi-agent debate system before final fine-tuning.

Project overview

The goal of this project is to improve green patent detection by combining:

  1. High-risk sample selection from uncertainty sampling
  2. Multi-agent LLM review of difficult claims
  3. Human verification of the AI suggestions
  4. Final fine-tuning of PatentSBERTa using silver labels + gold HITL labels

This workflow was designed to test whether a more advanced labeling pipeline produces stronger training data than a simple single-LLM labeling approach.

Base model

The final classifier is built from:

  • Base encoder: AI-Growth-Lab/PatentSBERTa
  • Task: Binary text classification
  • Domain: Patent claim classification for climate mitigation / green technology

Data used in the notebook

The notebook uses the following files:

  • patents_50k_green.parquet
  • train_meta.csv
  • y_train.npy
  • eval_silver.parquet
  • hitl_green_100.csv
  • hitl_review_progress_with_llm.csv
  • hitl_green_gold.csv
  • hitl_three_agents.csv

Methodology

1. High-risk claim selection

A set of 100 high-risk patent claims was selected from earlier uncertainty sampling outputs. These were the most difficult / ambiguous examples for the model.

2. Multi-agent debate system

Three agents were created using CrewAI and an Ollama-hosted model (qwen2.5:3b-instruct):

  • Advocate Agent – argues why the claim should be classified as green under Y02 climate mitigation logic
  • Skeptic Agent – argues why the claim may not qualify and checks for weak evidence or greenwashing
  • Judge Agent – reviews both sides and returns a structured final output with:
    • predicted label
    • confidence
    • rationale

This produces an AI suggestion for each difficult claim.

3. Human-in-the-Loop review

The AI-generated suggestion was then manually reviewed.

The final human label was stored as:

  • is_green_human

These human-reviewed labels form the gold dataset for the difficult claims.

4. Gold-enhanced training

The final training set combines:

  • Silver labels from the earlier training data
  • 100 gold human-reviewed claims from the multi-agent workflow

This combined dataset was used to fine-tune PatentSBERTa.

Training configuration

The notebook fine-tunes the model with the following setup:

  • Model: AI-Growth-Lab/PatentSBERTa
  • Max sequence length: 256
  • Epochs: 1
  • Learning rate: 2e-5
  • Train batch size: 8
  • Eval batch size: 8
  • Weight decay: 0.01
  • Framework: Hugging Face Transformers Trainer

Dataset splits used during fine-tuning

From the notebook:

  • Training data: silver training set + gold HITL labels
  • Evaluation data: eval_silver
  • Additional check: gold_100

The notebook text states that the final training dataset contains 35,200 claims.

Human vs AI agreement

According to the notebook:

  • Simple LLM from Assignment 2: 94% agreement with human labels
  • Agentic system from Assignment 3: 87% agreement with human labels

This suggests that the multi-agent system used stricter reasoning criteria, which created more disagreement on borderline cases.

Repository contents

Depending on what you upload, this repository may include:

  • the processed HITL dataset
  • the final trained model
  • tokenizer files
  • training notebook
  • prediction / rationale outputs for the 100 reviewed claims

Expected columns in the HITL dataset

The notebook shows or creates columns such as:

  • id
  • text
  • p_green
  • u
  • llm_green_suggested
  • llm_confidence
  • llm_rationale
  • is_green_human

Example use

from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch

model_name = "YOUR_HF_REPO_NAME"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)

text = "A company develops a carbon capture system that reduces CO2 emissions from cement factories."
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=256)

with torch.no_grad():
    logits = model(**inputs).logits
    pred = torch.argmax(logits, dim=-1).item()

print("Predicted label:", pred)

Intended use

This project is intended for:

  • research and coursework on green patent detection
  • experimentation with HITL labeling pipelines
  • comparison of simple vs advanced AI-assisted annotation workflows
  • climate-tech related document classification

Limitations

  • The gold set is relatively small (100 reviewed claims)
  • The multi-agent workflow depends on LLM reasoning quality
  • Agreement with humans does not automatically guarantee better downstream model performance
  • Final performance metrics should be reported from the actual training run in this repository

Notes

This README was prepared from the notebook workflow and code structure. If you are uploading the model repo, add the final evaluation metrics from your training output. If you are uploading the dataset repo, you can keep the methodology sections and remove the model inference example if not needed.

Citation

If you use this work, please cite the repository and the base model:

  • AI-Growth-Lab/PatentSBERTa

You may also describe the workflow as:

Multi-Agent Human-in-the-Loop green patent detection using PatentSBERTa with gold-enhanced fine-tuning.