tom-010's picture
Update README.md
4b35a45 verified
---
dataset_info:
features:
- name: question_id
dtype: int64
- name: question
dtype: string
- name: text
dtype: string
- name: label
dtype: int64
splits:
- name: validation
num_bytes: 7885383.303442969
num_examples: 20000
- name: train
num_bytes: 3571579491.696557
num_examples: 9058734
download_size: 2330139568
dataset_size: 3579464875.0
configs:
- config_name: default
data_files:
- split: validation
path: data/validation-*
- split: train
path: data/train-*
---
**Dataset Description**
This dataset is crafted for training a binary classification model to determine whether a given text passage answers a specific user query. Its primary purpose is to enhance our search engine by filtering out irrelevant passages, ensuring that users receive accurate and helpful responses to their questions.
**Background and Motivation**
In our search engine, users submit queries and receive multiple passages as results. Not all retrieved passages effectively answer the user's question, leading to a suboptimal user experience. To address this, we need a model capable of assessing each passage's relevance to the query, allowing us to present only the most pertinent information.
**Source Dataset**
The dataset is based on the [MS MARCO V2.1](https://github.com/zhouyonglong/MSMARCOV2) dataset from Microsoft, accessed via [Hugging Face Datasets](https://huggingface.co/datasets/microsoft/ms_marco) (version 2.1). MS MARCO is a large-scale corpus designed for machine reading comprehension and question answering tasks, containing real anonymized user queries and corresponding passages from web documents.
**Dataset Construction**
- **Original Format**: Each sample in MS MARCO V2.1 consists of:
- A **query** (user's question).
- A set of **10 passages** retrieved for that query.
- **Labels** indicating whether each passage was selected as an answer.
- **Transformation Process**:
- **Reshaping**: We transformed the dataset to suit a binary classification task by iterating over each passage in every sample.
- **Sample Creation**: For each query-passage pair, we created a new sample with:
- `question_id`: Unique identifier for the query.
- `question`: The user's query.
- `text`: The passage text.
- `label`: Binary label (`1` if the passage answers the question, `0` otherwise).
- **Dataset Splitting**:
- Combined the original **train** and **validation** splits, excluding the **test** split.
- Shuffled the combined dataset to ensure randomness.
- **Validation Set Size**: Determined based on statistical calculations to ensure a sufficient sample size for reliable validation metrics:
- **Accuracy Assumption**: 90%
- **Margin of Error**: 0.5%
- **Confidence Level**: 98% (z-score of 2.326)
- **Calculated Validation Size**: Approximately 20,000 samples
- Split the dataset into:
- **Training Set**: Remaining samples after allocating 20,000 to validation.
- **Validation Set**: 20,000 samples for model evaluation.
**Dataset Features**
- `question_id` (string): Unique identifier for each query.
- `question` (string): The user's query.
- `text` (string): The text of the passage.
- `label` (int): Binary indicator (`1` if the passage answers the question, `0` otherwise).
**Intended Use**
- **Primary Task**: Train a binary classification model to predict whether a passage answers a given query.
- **Application**: Integrate the model into our search engine pipeline to filter out non-relevant passages, improving the overall quality and relevance of search results.
**Considerations**
- **Class Balance**: The dataset may be imbalanced due to the nature of the original labels. It's important to consider this during model training (e.g., using class weights or resampling techniques).
- **Data Quality**: Passages are sourced from real-world search data and may contain noise or irrelevant information typical of web content.
- **Licensing**: Ensure compliance with MS MARCO's licensing terms when using this dataset for development or distribution.
**Conclusion**
This dataset aligns closely with our goal of improving answerability assessment in search results. By leveraging real user queries and associated passages, the trained model will be well-suited to judge the relevance of passages retrieved by our search engine, ultimately enhancing user satisfaction by providing accurate and relevant answers.
**Additional Notes**
- **Dataset Location**: Stored internally within our dataset repository for easy access and version control.
- **Reproducibility**: The notebook used to create this dataset is available and contains all steps for dataset generation, allowing for updates or modifications as needed.
- **Future Work**: Consider exploring sequence classification tasks or more complex models to further improve answerability predictions.