absa-dataset / README.md
efainman's picture
Upload README.md
e9329d0 verified
# TRABL: Travel-Domain Aspect-Based Sentiment Analysis Dataset
This repository contains the **TRABL dataset**, released in support of our paper accepted to **The ACM Web Conference 2026 (WWW 2026)**:
> **TRABL: A Unified Framework for Travel Domain Aspect-Based Sentiment Analysis Applications of Large Language Models**
The dataset is designed to support research on **Aspect-Based Sentiment Analysis (ABSA)** in the travel domain, with a particular focus on *joint extraction* of structured sentiment tuples and *supporting evidence snippets*.
---
## 1. Overview
TRABL extends the standard ABSA *quad prediction* task by additionally requiring the extraction of **textual evidence snippets** that justify each extracted sentiment tuple. The dataset enables training and evaluation of models for both fine-grained opinion mining and downstream travel applications such as recommendation, review summarization, and property type detection.
The data consists of **English user reviews** from three travel-related domains:
* **Hotels**
* **Attractions**
* **Destinations**
Each review is annotated by **two annotators**, following carefully designed guidelines and quality control procedures.
---
## 2. Task Definition
Given a review text (T), the task is to extract a set of tuples of the form:
```
(aspect_term, aspect_category, opinion_span, sentiment, snippet)
```
Where:
* **Aspect Term**: Explicit mention of the target entity (e.g., *"room"*)
* **Aspect Category**: Semantic category from a **closed set of 112 travel-domain categories**
* **Opinion Span**: Evaluative expression (e.g., *"spacious"*)
* **Sentiment**: One of `{positive, negative, neutral}`
* **Snippet**: A *contiguous span* of text from the review that provides sufficient evidence for the other fields
Some fields may be `null` if not explicitly mentioned (e.g., implicit aspects).
This formulation generalizes multiple ABSA subtasks, including:
* Category extraction `(c)`
* Category–sentiment extraction `(c, p)`
* Category–sentiment with evidence `(c, p, s)`
* Standard quad prediction `(a, o, c, p)`
* Quad prediction with snippet extraction `(a, o, c, p, s)`
---
## 3. Dataset Splits
The dataset is split into **train**, **validation**, and **test** sets as follows:
| Split | #Rows |
| ---------- | ----- |
| Train | 5,768 |
| Validation | 898 |
| Test | 900 |
> Note: A *row* corresponds to a single annotated review instance (including annotator-specific annotations).
---
## 4. Data Format
The data is released in **JSONL** format (one JSON object per line).
Each example contains the following top-level fields:
* **review_id**: Unique identifier for the review
* **annotator**: Identifier of the annotator (`1` or `2`)
* **text**: Full review text
* Hotel reviews include title, positive, and negative sections
* Attraction and destination reviews contain free text only
* **labels**: List of labeled tuples
Each entry in `annotations` has the structure:
```json
{
"aspect_term": "value for money",
"aspect_category": "value for money",
"opinion_span": "good",
"sentiment": "positive",
"snippet": "The good value for money"
}
```
---
## 5. Example
**Input Review (excerpt):**
```
The good value for money, the helpful & polite staff, the level of cleanliness.
```
**Annotations (excerpt):**
```json
[
{
"aspect_term": "value for money",
"aspect_category": "value for money",
"opinion_span": "good",
"sentiment": "positive",
"snippet": "The good value for money"
},
{
"aspect_term": "staff",
"aspect_category": "property staff and service support",
"opinion_span": "helpful & polite",
"sentiment": "positive",
"snippet": "the helpful & polite staff"
}
]
```
---
## 6. Annotation Process and Quality
* Each review was annotated by **two annotators**
* Annotation was supported by LLM-generated initial labels, followed by human correction
* Disagreements were normalized using post-processing rules
* **Overall annotator agreement** (F1): ~67%, consistent with prior ABSA benchmarks
* Agreement exceeds **80%** when evaluated on partial field subsets (e.g., category or sentiment only)
---
## 7. Statistics
* **Domains**: Hotels, Attractions, Destinations
* **Total number of reviews**: 3,783
* **Average review length**: ~42 words
* **Average number of extracted tuples per review**:
* Annotator 1: ~6.96
* Annotator 2: ~6.77
* **Sentiment distribution**: Skewed positive, reflecting real-world travel review behavior
---
## 8. Intended Use
This dataset is intended for **research purposes only**, including but not limited to:
* Aspect-Based Sentiment Analysis
* Opinion mining and evidence extraction
* Travel-domain NLP
* Evaluation of LLMs and lightweight fine-tuned models
* Structured information extraction with explainability
---
## 9. Citation
If you use this dataset, please cite:
```
@inproceedings{madmon2026trabl,
title = {TRABL: A Unified Framework for Travel Domain Aspect-Based Sentiment Analysis Applications of Large Language Models},
author = {Madmon, Omer and Golan, Shiran and Fainman, Eran and Kleinfeld, Ofri and Beladev, Moran},
booktitle = {Proceedings of The ACM Web Conference},
year = {2026}
}
```
---
## 10. License and Privacy
* All reviews were processed under strict internal privacy and PII guidelines
* The dataset is released for **non-commercial research use**, cc-by-sa-4.0 license
* No personally identifiable information is included
---
For questions or issues related to the dataset, please refer to the paper or open an issue on the dataset repository.