| | --- |
| | language: |
| | - ar |
| | license: mit |
| | task_categories: |
| | - text-classification |
| | tags: |
| | - arabic |
| | - sentiment |
| | - logistics |
| | - customer-feedback |
| | - delivery-service |
| | pretty_name: Arabic Logistics Feedback Corpus |
| | size_categories: |
| | - 1K<n<10K |
| | dataset_info: |
| | features: |
| | - name: text |
| | dtype: string |
| | description: Customer feedback text in Arabic about logistics/delivery service |
| | - name: label |
| | dtype: string |
| | description: Sentiment label (positive/negative) |
| | - name: score |
| | dtype: float32 |
| | description: Confidence score for the label |
| | - name: domain |
| | dtype: string |
| | description: Domain of the feedback (currently all logistics) |
| | - name: is_conflict |
| | dtype: bool |
| | description: Indicates if there's a conflict between label and actual sentiment |
| | splits: |
| | - name: train |
| | num_bytes: 178122 |
| | num_examples: 1504 |
| | download_size: 75168 |
| | dataset_size: 178122 |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - split: train |
| | path: data/train-* |
| | --- |
| | |
| | # Arabic Feedback Corpus |
| |
|
| | ## Dataset Description |
| |
|
| | This dataset contains **1,504 Arabic customer feedback entries** for sentiment analysis and quality assessment in the logistics domain. The data consists of real customer reviews about delivery services, courier performance, and order fulfillment experiences in Egyptian Arabic and Modern Standard Arabic. |
| |
|
| | ### Languages |
| |
|
| | * **Primary**: Egyptian Arabic (العامية المصرية) |
| | * **Secondary**: Modern Standard Arabic (MSA) |
| | * **Code**: ar, ar_EG |
| | |
| | ## Dataset Summary |
| | |
| | Customer feedback is crucial for service improvement and quality assurance. This dataset provides: |
| | |
| | * Authentic customer reviews from logistics services |
| | * Binary sentiment labels (positive/negative) |
| | * Quality scores (1-5 scale) |
| | * Conflict detection flags for quality control |
| | * Real-world colloquial Egyptian Arabic expressions |
| | |
| | ## Dataset Structure |
| | |
| | ### Data Format |
| | |
| | Each entry contains: |
| | |
| | ```json |
| | { |
| | "text": "المندوب محترم جدا وسريع في التوصيل", |
| | "label": "positive", |
| | "score": 5.0, |
| | "domain": "logistics", |
| | "is_conflict": false |
| | } |
| | ``` |
| | |
| | ### Data Fields |
| | |
| | | Field | Type | Description | |
| | | --- | --- | --- | |
| | | `text` | string | Customer feedback text in Arabic | |
| | | `label` | string | Sentiment label ("positive" or "negative") | |
| | | `score` | float | Quality rating (1.0 to 5.0) | |
| | | `domain` | string | Content domain (always "logistics") | |
| | | `is_conflict` | bool | Flag for label-score conflicts | |
| | |
| | ### Field Details |
| | |
| | #### `text` |
| | Customer feedback ranging from 3 to 75 characters, containing: |
| | - Delivery experience descriptions |
| | - Courier behavior comments |
| | - Service quality assessments |
| | - Product condition feedback |
| | - Timing and professionalism complaints/praise |
| | |
| | #### `label` |
| | Binary sentiment classification: |
| | - **positive**: Satisfied customers, good experiences |
| | - **negative**: Complaints, dissatisfaction, problems |
| | |
| | #### `score` |
| | Numerical rating on 1-5 scale: |
| | - **5.0**: Excellent service |
| | - **4.0**: Good service |
| | - **3.0**: Average service |
| | - **2.0**: Below average |
| | - **1.0**: Poor service |
| | |
| | #### `is_conflict` |
| | Quality control flag indicating mismatch between label and score: |
| | - **false**: Label and score are consistent |
| | - **true**: Conflict detected (e.g., positive label with score 1.0) |
| | |
| | ## Dataset Statistics |
| | |
| | ### Overview |
| | |
| | * **Total Entries**: 1,504 |
| | * **Positive Reviews**: ~35% |
| | * **Negative Reviews**: ~65% |
| | * **Conflicted Labels**: ~2% |
| | * **Average Text Length**: 38.5 characters |
| | * **Domain**: Logistics only |
| | |
| | ### Label Distribution |
| | |
| | | Label | Count | Percentage | |
| | | --- | --- | --- | |
| | | negative | ~978 | 65% | |
| | | positive | ~526 | 35% | |
| | |
| | ### Score Distribution |
| | |
| | | Score | Count | Typical Label | |
| | | --- | --- | --- | |
| | | 1.0 | ~1,450 | negative | |
| | | 5.0 | ~50 | positive | |
| | | 2.0-4.0 | ~4 | varies | |
| | |
| | ### Conflict Examples |
| | |
| | Conflicted entries (where label contradicts score): |
| | |
| | ```python |
| | { |
| | "text": "ممتاز وسرعة في الاداء", |
| | "label": "positive", |
| | "score": 1.0, # ← Conflict! |
| | "is_conflict": true |
| | } |
| | |
| | { |
| | "text": "المندوب بيبلغ بوقت وبيجي بعديها ب ٧ ساعات", |
| | "label": "positive", # ← Conflict! |
| | "score": 1.0, |
| | "is_conflict": true |
| | } |
| | ``` |
| | |
| | ## Common Feedback Themes |
| | |
| | ### Positive Feedback Topics |
| | - ✅ Professional and respectful couriers |
| | - ✅ Fast delivery |
| | - ✅ Good communication |
| | - ✅ Helpful service |
| | - ✅ On-time arrival |
| | |
| | ### Negative Feedback Topics |
| | - ❌ Rude or unprofessional behavior |
| | - ❌ Delivery delays |
| | - ❌ Courier refusing to come upstairs |
| | - ❌ Extra charges/tips demanded |
| | - ❌ Not answering calls |
| | - ❌ Poor product condition |
| | - ❌ Wrong items delivered |
| | - ❌ Courier attitude problems |
| | |
| | ## Use Cases |
| | |
| | ### ✅ Recommended Use Cases |
| | |
| | * **Sentiment Analysis**: Train Arabic sentiment classifiers |
| | * **Quality Assessment**: Predict service quality scores |
| | * **Conflict Detection**: Identify inconsistent reviews |
| | * **Egyptian Arabic NLP**: Understand colloquial expressions |
| | * **Customer Service AI**: Build chatbots understanding complaints |
| | * **Logistics Analytics**: Analyze delivery service quality |
| | * **Multi-Task Learning**: Joint sentiment + score prediction |
| | * **Data Quality Models**: Detect annotation inconsistencies |
| | |
| | ### ⚠️ Limitations |
| | |
| | * **Domain Specificity**: Limited to logistics/delivery domain |
| | * **Geographic Scope**: Primarily Egyptian context |
| | * **Label Noise**: Contains ~2% conflicted labels |
| | * **Imbalanced Data**: 65% negative vs 35% positive |
| | * **Size**: 1,504 entries (medium-sized dataset) |
| | * **Score Distribution**: Heavily skewed toward 1.0 and 5.0 |
| | |
| | ## Loading the Dataset |
| | |
| | ### Using Hugging Face Datasets |
| | |
| | ```python |
| | from datasets import load_dataset |
| | |
| | # Load the dataset |
| | dataset = load_dataset("fr3on/arabic-feedback-corpus") |
| |
|
| | # Access the data |
| | print(dataset['train'][0]) |
| |
|
| | # Filter by sentiment |
| | positive_reviews = dataset['train'].filter(lambda x: x['label'] == 'positive') |
| | negative_reviews = dataset['train'].filter(lambda x: x['label'] == 'negative') |
| |
|
| | # Filter clean data (no conflicts) |
| | clean_data = dataset['train'].filter(lambda x: x['is_conflict'] == False) |
| |
|
| | # Filter by score |
| | excellent_service = dataset['train'].filter(lambda x: x['score'] == 5.0) |
| | poor_service = dataset['train'].filter(lambda x: x['score'] == 1.0) |
| | ``` |
| | |
| | ### Using Pandas |
| | |
| | ```python |
| | import pandas as pd |
| |
|
| | # Load Parquet file directly |
| | df = pd.read_parquet("hf://datasets/fr3on/arabic-feedback-corpus/data/train-00000-of-00001.parquet") |
| | |
| | # Analyze sentiment distribution |
| | print(df['label'].value_counts()) |
| |
|
| | # Check for conflicts |
| | conflicts = df[df['is_conflict'] == True] |
| | print(f"Conflicted entries: {len(conflicts)}") |
| | |
| | # Score statistics |
| | print(df['score'].describe()) |
| | |
| | # Export filtered data |
| | positive_df = df[df['label'] == 'positive'] |
| | positive_df.to_csv('positive_feedback.csv', index=False) |
| | ``` |
| | |
| | ## Training Examples |
| | |
| | ### Sentiment Classification |
| | |
| | ```python |
| | from datasets import load_dataset |
| | from transformers import AutoTokenizer, AutoModelForSequenceClassification, Trainer |
| |
|
| | # Load dataset |
| | dataset = load_dataset("fr3on/arabic-feedback-corpus") |
| | |
| | # Remove conflicted samples for clean training |
| | clean_dataset = dataset['train'].filter(lambda x: not x['is_conflict']) |
| | |
| | # Load Arabic BERT model |
| | model_name = "asafaya/bert-base-arabic" |
| | tokenizer = AutoTokenizer.from_pretrained(model_name) |
| | model = AutoModelForSequenceClassification.from_pretrained( |
| | model_name, |
| | num_labels=2 # positive/negative |
| | ) |
| | |
| | # Tokenize |
| | def preprocess(examples): |
| | return tokenizer( |
| | examples['text'], |
| | truncation=True, |
| | max_length=128, |
| | padding='max_length' |
| | ) |
| | |
| | tokenized = clean_dataset.map(preprocess, batched=True) |
| | |
| | # Convert labels to numbers |
| | label_map = {'negative': 0, 'positive': 1} |
| | tokenized = tokenized.map(lambda x: {'label': label_map[x['label']]}) |
| | |
| | # Train |
| | trainer = Trainer( |
| | model=model, |
| | train_dataset=tokenized, |
| | ) |
| | trainer.train() |
| | ``` |
| | |
| | ### Multi-Task Learning (Sentiment + Score) |
| | |
| | ```python |
| | from datasets import load_dataset |
| | import torch.nn as nn |
| | |
| | dataset = load_dataset("fr3on/arabic-feedback-corpus") |
| | clean_data = dataset['train'].filter(lambda x: not x['is_conflict']) |
| |
|
| | # Multi-task model architecture |
| | class MultiTaskModel(nn.Module): |
| | def __init__(self, base_model): |
| | super().__init__() |
| | self.base = base_model |
| | self.sentiment_head = nn.Linear(768, 2) # positive/negative |
| | self.score_head = nn.Linear(768, 1) # score prediction |
| | |
| | def forward(self, input_ids, attention_mask): |
| | outputs = self.base(input_ids, attention_mask) |
| | pooled = outputs.last_hidden_state[:, 0] # CLS token |
| | |
| | sentiment = self.sentiment_head(pooled) |
| | score = self.score_head(pooled) |
| | |
| | return sentiment, score |
| | |
| | # Train with both objectives |
| | # sentiment_loss = CrossEntropyLoss() |
| | # score_loss = MSELoss() |
| | # total_loss = sentiment_loss + score_loss |
| | ``` |
| | |
| | ### Conflict Detection |
| | |
| | ```python |
| | from datasets import load_dataset |
| |
|
| | dataset = load_dataset("fr3on/arabic-feedback-corpus") |
| | |
| | # Train a model to detect annotation conflicts |
| | # Features: text + predicted_label + predicted_score |
| | # Target: is_conflict flag |
| |
|
| | def extract_features(example): |
| | return { |
| | 'text': example['text'], |
| | 'label': example['label'], |
| | 'score': example['score'], |
| | 'target': example['is_conflict'] |
| | } |
| | |
| | conflict_dataset = dataset['train'].map(extract_features) |
| |
|
| | # This can help identify: |
| | # - Annotation errors |
| | # - Sarcastic comments |
| | # - Ambiguous feedback |
| | ``` |
| | |
| | ## Data Collection & Processing |
| | |
| | ### Source |
| | |
| | * **Origin**: Real customer feedback from logistics services |
| | * **Language**: Primarily Egyptian Arabic (colloquial) |
| | * **Quality**: Authentic user-generated content |
| | |
| | ### Annotation Process |
| | |
| | 1. **Text Collection**: Customer reviews and feedback |
| | 2. **Labeling**: Binary sentiment annotation (positive/negative) |
| | 3. **Scoring**: Quality ratings on 1-5 scale |
| | 4. **Conflict Detection**: Automated flag for label-score mismatches |
| | 5. **Validation**: Quality checks and consistency reviews |
| | |
| | ### Data Quality |
| | |
| | * ✅ Real customer feedback (not synthetic) |
| | * ⚠️ Contains ~2% label-score conflicts |
| | * ✅ Text lengths validated (3-75 characters) |
| | * ✅ Domain consistency (all logistics) |
| | * ⚠️ Class imbalance (65% negative) |
| | |
| | ## Considerations for Using the Data |
| | |
| | ### Egyptian Arabic Characteristics |
| | |
| | This dataset contains colloquial Egyptian expressions: |
| | |
| | * **Informal spelling**: مش instead of ليس |
| | * **Egyptian vocabulary**: مندوب، اوردر، شحنة |
| | * **Mixed language**: Some English words (أوردر = order) |
| | * **Abbreviated words**: ج for جنيه (Egyptian pound) |
| | |
| | ### Handling Conflicts |
| | |
| | The `is_conflict` flag identifies potential issues: |
| | |
| | ```python |
| | # Option 1: Exclude conflicts |
| | clean_data = dataset.filter(lambda x: not x['is_conflict']) |
| |
|
| | # Option 2: Use conflicts for quality control training |
| | conflicts = dataset.filter(lambda x: x['is_conflict']) |
| | |
| | # Option 3: Manually review and correct |
| | for item in conflicts: |
| | # Review and fix annotations |
| | pass |
| | ``` |
| | |
| | ### Recommended Training Approaches |
| | |
| | 1. **Balance the dataset** using oversampling or class weights |
| | 2. **Remove conflicts** for cleaner training |
| | 3. **Use Arabic-specific models** (AraBERT, MARBERT) |
| | 4. **Consider dialectal variations** in preprocessing |
| | 5. **Apply data augmentation** to address class imbalance |
| | |
| | ### Ethical Considerations |
| | |
| | * **Privacy**: Customer names and personal info removed |
| | * **Bias**: Dataset reflects real customer experiences |
| | * **Negativity bias**: More complaints than praise (common in feedback data) |
| | * **Cultural context**: Egyptian service expectations and norms |
| | |
| | ## Applications |
| | |
| | ### Customer Service Automation |
| | |
| | ```python |
| | # Real-time sentiment analysis for support tickets |
| | def analyze_feedback(text): |
| | sentiment = model.predict(text) |
| | if sentiment == 'negative' and score < 3.0: |
| | # Escalate to human agent |
| | priority = "high" |
| | return sentiment, score, priority |
| | ``` |
| | |
| | ### Quality Monitoring |
| |
|
| | ```python |
| | # Track service quality trends |
| | import pandas as pd |
| | |
| | df = pd.read_parquet("data.parquet") |
| | daily_scores = df.groupby('date')['score'].mean() |
| | |
| | # Alert on quality drops |
| | if daily_scores.last() < 3.0: |
| | send_alert("Service quality declining") |
| | ``` |
| |
|
| | ### Training Data Annotation |
| |
|
| | ```python |
| | # Use model to pre-annotate new data |
| | new_feedback = ["المندوب كان ممتاز"] |
| | predicted_label = model.predict(new_feedback) |
| | # Human reviews and corrects predictions |
| | ``` |
| |
|
| | ## Common Arabic Tokens |
| |
|
| | **Positive indicators**: |
| | - ممتاز (excellent) |
| | - محترم (respectful) |
| | - سريع (fast) |
| | - كويس (good) |
| | - شكرا (thanks) |
| |
|
| | **Negative indicators**: |
| | - سيء (bad) |
| | - اتأخر (delayed) |
| | - قليل الذوق (rude, lit. "little taste") |
| | - وحش (bad/ugly) |
| | - مش (not) |
| | - رفض (refused) |
| |
|
| | **Neutral/Context-dependent**: |
| | - المندوب (the courier) |
| | - الاوردر (the order) |
| | - الشحنة (the shipment) |
| | - وصل (arrived) |
| |
|
| | ## License |
| |
|
| | This dataset is released under the **Apache 2.0 License**. |
| |
|
| | ## Citation |
| |
|
| | If you use this dataset in your research, please cite: |
| |
|
| | ```bibtex |
| | @dataset{arabic_feedback_corpus, |
| | title={Arabic Feedback Corpus: Logistics Domain Sentiment Analysis}, |
| | author={fr3on}, |
| | year={2026}, |
| | publisher={Hugging Face}, |
| | url={https://huggingface.co/datasets/fr3on/arabic-feedback-corpus} |
| | } |
| | ``` |
| |
|
| | ## Acknowledgments |
| |
|
| | * Source: Customer feedback from logistics services |
| | * Annotation: Sentiment labels and quality scores |
| | * Format: Parquet for efficient storage and loading |
| |
|
| | ## Version History |
| |
|
| | * **v1.0.0** (2026-01-06): Initial release |
| | + 1,504 entries |
| | + Binary sentiment labels |
| | + 1-5 quality scores |
| | + Conflict detection flags |
| | + Parquet format |
| |
|
| | --- |
| |
|
| | **Keywords**: Arabic NLP, sentiment analysis, customer feedback, logistics, Egyptian Arabic, colloquial Arabic, quality assessment, conflict detection, delivery services |
| |
|
| | **Dataset Size**: 1,504 examples | **Format**: Parquet | **License**: Apache 2.0 |
| |
|