Datasets:
Duplicate from Dingdong-Inc/FreshRetailNet-50K
Browse filesCo-authored-by: Jarvis Gu <JarvisGu@users.noreply.huggingface.co>
- .gitattributes +61 -0
- .gitignore +10 -0
- README.md +123 -0
- data/eval.parquet +3 -0
- data/train.parquet +3 -0
.gitattributes
ADDED
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.lz4 filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.mds filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
# Audio files - uncompressed
|
| 39 |
+
*.pcm filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
*.sam filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
*.raw filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
# Audio files - compressed
|
| 43 |
+
*.aac filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
*.flac filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
*.mp3 filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
*.ogg filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
*.wav filter=lfs diff=lfs merge=lfs -text
|
| 48 |
+
# Image files - uncompressed
|
| 49 |
+
*.bmp filter=lfs diff=lfs merge=lfs -text
|
| 50 |
+
*.gif filter=lfs diff=lfs merge=lfs -text
|
| 51 |
+
*.png filter=lfs diff=lfs merge=lfs -text
|
| 52 |
+
*.tiff filter=lfs diff=lfs merge=lfs -text
|
| 53 |
+
# Image files - compressed
|
| 54 |
+
*.jpg filter=lfs diff=lfs merge=lfs -text
|
| 55 |
+
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
| 56 |
+
*.webp filter=lfs diff=lfs merge=lfs -text
|
| 57 |
+
# Video files - compressed
|
| 58 |
+
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 59 |
+
*.webm filter=lfs diff=lfs merge=lfs -text
|
| 60 |
+
data/train.parquet filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
data/eval.parquet filter=lfs diff=lfs merge=lfs -text
|
.gitignore
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#system ignore
|
| 2 |
+
.DS_Store
|
| 3 |
+
Thumbs.db
|
| 4 |
+
.idea
|
| 5 |
+
|
| 6 |
+
# files by the editor
|
| 7 |
+
*.swp
|
| 8 |
+
*.vscode
|
| 9 |
+
.ipynb_checkpoints
|
| 10 |
+
|
README.md
ADDED
|
@@ -0,0 +1,123 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
language:
|
| 3 |
+
- en
|
| 4 |
+
license: cc-by-4.0
|
| 5 |
+
task_categories:
|
| 6 |
+
- time-series-forecasting
|
| 7 |
+
tags:
|
| 8 |
+
- fresh-retail
|
| 9 |
+
- censored-demand
|
| 10 |
+
- hourly-stock-status
|
| 11 |
+
size_categories:
|
| 12 |
+
- 1M<n<10M
|
| 13 |
+
pretty_name: FreshRetailNet-50K
|
| 14 |
+
configs:
|
| 15 |
+
- config_name: default
|
| 16 |
+
data_files:
|
| 17 |
+
- split: train
|
| 18 |
+
path: data/train.parquet
|
| 19 |
+
- split: eval
|
| 20 |
+
path: data/eval.parquet
|
| 21 |
+
---
|
| 22 |
+
|
| 23 |
+
# FreshRetailNet-50K
|
| 24 |
+
|
| 25 |
+
## Dataset Overview
|
| 26 |
+
FreshRetailNet-50K is the first large-scale benchmark for censored demand estimation in the fresh retail domain, **incorporating approximately 20% organically occurring stockout data**. It comprises 50,000 store-product 90-day time series of detailed hourly sales data from 898 stores in 18 major cities, encompassing 863 perishable SKUs with meticulous stockout event annotations. The hourly stock status records unique to this dataset, combined with rich contextual covariates including promotional discounts, precipitation, and other temporal features, enable innovative research beyond existing solutions.
|
| 27 |
+
|
| 28 |
+
- [Technical Report](https://arxiv.org/abs/2505.16319) - Discover the methodology and technical details behind FreshRetailNet-50K.
|
| 29 |
+
- [Github Repo](https://github.com/Dingdong-Inc/frn-50k-baseline) - Access the complete pipeline used to train and evaluate.
|
| 30 |
+
|
| 31 |
+
This dataset is ready for commercial/non-commercial use.
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
## Data Fields
|
| 35 |
+
|Field|Type|Description|
|
| 36 |
+
|:---|:---|:---|
|
| 37 |
+
|city_id|int64|The encoded city id|
|
| 38 |
+
|store_id|int64|The encoded store id|
|
| 39 |
+
|management_group_id|int64|The encoded management group id|
|
| 40 |
+
|first_category_id|int64|The encoded first category id|
|
| 41 |
+
|second_category_id|int64|The encoded second category id|
|
| 42 |
+
|third_category_id|int64|The encoded third category id|
|
| 43 |
+
|product_id|int64|The encoded product id|
|
| 44 |
+
|dt|string|The date|
|
| 45 |
+
|sale_amount|float64|The daily sales amount after global normalization (Multiplied by a specific coefficient)|
|
| 46 |
+
|hours_sale|Sequence(float64)|The hourly sales amount after global normalization (Multiplied by a specific coefficient)|
|
| 47 |
+
|stock_hour6_22_cnt|int32|The number of out-of-stock hours between 6:00 and 22:00|
|
| 48 |
+
|hours_stock_status|Sequence(int32)|The hourly out-of-stock status|
|
| 49 |
+
|discount|float64|The discount rate (1.0 means no discount, 0.9 means 10% off)|
|
| 50 |
+
|holiday_flag|int32|Holiday indicator|
|
| 51 |
+
|activity_flag|int32|Activity indicator|
|
| 52 |
+
|precpt|float64|The total precipitation|
|
| 53 |
+
|avg_temperature|float64|The average temperature|
|
| 54 |
+
|avg_humidity|float64|The average humidity|
|
| 55 |
+
|avg_wind_level|float64|The average wind force|
|
| 56 |
+
|
| 57 |
+
### Hierarchical structure
|
| 58 |
+
- **warehouse**: city_id > store_id
|
| 59 |
+
- **product category**: management_group_id > first_category_id > second_category_id > third_category_id > product_id
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
|
| 63 |
+
## How to use it
|
| 64 |
+
|
| 65 |
+
You can load the dataset with the following lines of code.
|
| 66 |
+
|
| 67 |
+
```python
|
| 68 |
+
from datasets import load_dataset
|
| 69 |
+
dataset = load_dataset("Dingdong-Inc/FreshRetailNet-50K")
|
| 70 |
+
print(dataset)
|
| 71 |
+
```
|
| 72 |
+
```text
|
| 73 |
+
DatasetDict({
|
| 74 |
+
train: Dataset({
|
| 75 |
+
features: ['city_id', 'store_id', 'management_group_id', 'first_category_id', 'second_category_id', 'third_category_id', 'product_id', 'dt', 'sale_amount', 'hours_sale', 'stock_hour6_22_cnt', 'hours_stock_status', 'discount', 'holiday_flag', 'activity_flag', 'precpt', 'avg_temperature', 'avg_humidity', 'avg_wind_level'],
|
| 76 |
+
num_rows: 4500000
|
| 77 |
+
})
|
| 78 |
+
eval: Dataset({
|
| 79 |
+
features: ['city_id', 'store_id', 'management_group_id', 'first_category_id', 'second_category_id', 'third_category_id', 'product_id', 'dt', 'sale_amount', 'hours_sale', 'stock_hour6_22_cnt', 'hours_stock_status', 'discount', 'holiday_flag', 'activity_flag', 'precpt', 'avg_temperature', 'avg_humidity', 'avg_wind_level'],
|
| 80 |
+
num_rows: 350000
|
| 81 |
+
})
|
| 82 |
+
})
|
| 83 |
+
```
|
| 84 |
+
|
| 85 |
+
|
| 86 |
+
## License/Terms of Use
|
| 87 |
+
|
| 88 |
+
This dataset is licensed under the Creative Commons Attribution 4.0 International License (CC BY 4.0) available at https://creativecommons.org/licenses/by/4.0/legalcode.
|
| 89 |
+
|
| 90 |
+
**Data Developer:** Dingdong-Inc
|
| 91 |
+
|
| 92 |
+
|
| 93 |
+
### Use Case: <br>
|
| 94 |
+
Developers researching latent demand recovery and demand forecasting techniques. <br>
|
| 95 |
+
|
| 96 |
+
### Release Date: <br>
|
| 97 |
+
05/08/2025 <br>
|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
## Data Version
|
| 101 |
+
1.0 (05/08/2025)
|
| 102 |
+
|
| 103 |
+
|
| 104 |
+
## Intended use
|
| 105 |
+
|
| 106 |
+
The FreshRetailNet-50K Dataset is intended to be freely used by the community to continue to improve latent demand recovery and demand forecasting techniques.
|
| 107 |
+
**However, for each dataset an user elects to use, the user is responsible for checking if the dataset license is fit for the intended purpose**.
|
| 108 |
+
|
| 109 |
+
|
| 110 |
+
## Citation
|
| 111 |
+
|
| 112 |
+
If you find the data useful, please cite:
|
| 113 |
+
```
|
| 114 |
+
@article{2025freshretailnet-50k,
|
| 115 |
+
title={FreshRetailNet-50K: A Stockout-Annotated Censored Demand Dataset for Latent Demand Recovery and Forecasting in Fresh Retail},
|
| 116 |
+
author={Yangyang Wang, Jiawei Gu, Li Long, Xin Li, Li Shen, Zhouyu Fu, Xiangjun Zhou, Xu Jiang},
|
| 117 |
+
year={2025},
|
| 118 |
+
eprint={2505.16319},
|
| 119 |
+
archivePrefix={arXiv},
|
| 120 |
+
primaryClass={cs.LG},
|
| 121 |
+
url={https://arxiv.org/abs/2505.16319},
|
| 122 |
+
}
|
| 123 |
+
```
|
data/eval.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1b118840664280c6b88bffc84c80ee1f54c05d911e354b7599e5da10995e960e
|
| 3 |
+
size 8440124
|
data/train.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6706832db892bbae4969c19d87e07975d2543d2ba7d7d4756360654785de5a3d
|
| 3 |
+
size 106436287
|