Datasets:

Modalities:
Time-series
Formats:
parquet
ArXiv:
License:
fabiencasenave commited on
Commit
48655bd
·
verified ·
1 Parent(s): ada6d9a

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +114 -0
README.md CHANGED
@@ -1,4 +1,11 @@
1
  ---
 
 
 
 
 
 
 
2
  dataset_info:
3
  features:
4
  - name: Base_2_3/Zone/CellData/Density
@@ -58,3 +65,110 @@ configs:
58
  - split: test
59
  path: data/test-*
60
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-sa-4.0
3
+ task_categories:
4
+ - graph-ml
5
+ pretty_name: 3D RANS simulations of the rotor37
6
+ tags:
7
+ - physics learning
8
+ - geometry learning
9
  dataset_info:
10
  features:
11
  - name: Base_2_3/Zone/CellData/Density
 
65
  - split: test
66
  path: data/test-*
67
  ---
68
+ <p align='center'>
69
+ <img src='https://i.ibb.co/zThPK7B8/Logo-Rotor37-2-consolas-100.png' alt='https://i.ibb.co/zThPK7B8/Logo-Rotor37-2-consolas-100.png' width='1000'/>
70
+ <img src='https://i.ibb.co/DKP161M/rotor37-preview.png' alt='https://i.ibb.co/DKP161M/rotor37-preview.png' width='1000'/>
71
+ </p>
72
+
73
+ ```yaml
74
+ legal:
75
+ owner: Safran
76
+ license: cc-by-sa-4.0
77
+ data_production:
78
+ type: simulation
79
+ physics: 3D CFD RANS compressor blade
80
+ num_samples:
81
+ train: 1000
82
+ test: 200
83
+ storage_backend: hf_datasets
84
+ plaid:
85
+ version: 0.1.11.dev21+g94f13b9c8
86
+
87
+ ```
88
+ This dataset was generated with [`plaid`](https://plaid-lib.readthedocs.io/), we refer to this documentation for additional details on how to extract data from `plaid_sample` objects.
89
+
90
+ The simplest way to use this dataset is to first download it:
91
+ ```python
92
+ from plaid.storage import download_from_hub
93
+
94
+ repo_id = "channel/dataset"
95
+ local_folder = "downloaded_dataset"
96
+
97
+ download_from_hub(repo_id, local_folder)
98
+ ```
99
+
100
+ Then, to iterate over the dataset and instantiate samples:
101
+ ```python
102
+ from plaid.storage import init_from_disk
103
+
104
+ local_folder = "downloaded_dataset"
105
+ split_name = "train"
106
+
107
+ datasetdict, converterdict = init_from_disk(local_folder)
108
+
109
+ dataset = datasetdict[split]
110
+ converter = converterdict[split]
111
+
112
+ for i in range(len(dataset)):
113
+ raw_sample = dataset[i]
114
+ plaid_sample = converter.to_plaid(dataset, i)
115
+ ```
116
+
117
+ It is possible to stream the data directly:
118
+ ```python
119
+ from plaid.storage import init_streaming_from_hub
120
+
121
+ repo_id = "channel/dataset"
122
+
123
+ datasetdict, converterdict = init_streaming_from_hub(repo_id)
124
+
125
+ dataset = datasetdict[split]
126
+ converter = converterdict[split]
127
+
128
+ for sample_raw in dataset:
129
+ plaid_sample = converter.sample_to_plaid(sample_raw)
130
+ ```
131
+
132
+ Plaid samples' features can be retrieved like the following:
133
+ ```python
134
+ from plaid.storage import load_problem_definitions_from_disk
135
+ local_folder = "downloaded_dataset"
136
+ pb_defs = load_problem_definitions_from_disk(local_folder)
137
+
138
+ # or
139
+ from plaid.storage import load_problem_definitions_from_hub
140
+ repo_id = "channel/dataset"
141
+ pb_defs = load_problem_definitions_from_hub(repo_id)
142
+
143
+
144
+ pb_def = pb_defs[0]
145
+
146
+ plaid_sample = ... # use a method from above to instantiate a plaid sample
147
+
148
+ for t in plaid_sample.get_all_time_values():
149
+ for path in pb_def.get_in_features_identifiers():
150
+ plaid_sample.get_feature_by_path(path=path, time=t)
151
+ for path in pb_def.get_out_features_identifiers():
152
+ plaid_sample.get_feature_by_path(path=path, time=t)
153
+ ```
154
+
155
+ For those familiar with HF's `datasets` library, raw data can be retrieved without using the `plaid` library:
156
+ ```python
157
+ from datasets import load_dataset
158
+
159
+ repo_id = "channel/dataset"
160
+
161
+ datasetdict = load_dataset(repo_id)
162
+
163
+ for split_name, dataset in datasetdict.items():
164
+ for raw_sample in dataset:
165
+ for feat_name in dataset.column_names:
166
+ feature = raw_sample[feat_name]
167
+ ```
168
+ Notice that raw data refers to the variable features only, with a specific encoding for time variable features.
169
+
170
+ ### Dataset Sources
171
+
172
+ - **Papers:**
173
+ - [arxiv](https://arxiv.org/pdf/2305.12871)
174
+ - [arxiv](https://arxiv.org/abs/2505.02974)