Datasets:
File size: 5,098 Bytes
05d9a2b 00f78dd 05d9a2b 007de7e 05d9a2b 007de7e 05d9a2b 00f78dd 2c7b9ee 00f78dd |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 |
---
license: cc-by-4.0
task_categories:
- graph-ml
pretty_name: ForceASR dataset
tags:
- physics learning
- geometry learning
dataset_info:
features:
- name: Base_2_2/Zone
list:
list: int64
- name: Base_2_2/Zone/Elements_QUAD_4/ElementConnectivity
list: int64
- name: Base_2_2/Zone/Elements_QUAD_4/ElementConnectivity_times
list: float64
- name: Base_2_2/Zone/Elements_QUAD_4/ElementRange
list: int64
- name: Base_2_2/Zone/Elements_QUAD_4/ElementRange_times
list: float64
- name: Base_2_2/Zone/GridCoordinates/CoordinateX
list: float32
- name: Base_2_2/Zone/GridCoordinates/CoordinateX_times
list: float64
- name: Base_2_2/Zone/GridCoordinates/CoordinateY
list: float32
- name: Base_2_2/Zone/GridCoordinates/CoordinateY_times
list: float64
- name: Base_2_2/Zone/VertexFields/Displacement_X
list: float32
- name: Base_2_2/Zone/VertexFields/Displacement_X_times
list: float64
- name: Base_2_2/Zone/VertexFields/Displacement_Y
list: float32
- name: Base_2_2/Zone/VertexFields/Displacement_Y_times
list: float64
- name: Base_2_2/Zone/VertexFields/PhaseField
list: float32
- name: Base_2_2/Zone/VertexFields/PhaseField_times
list: float64
- name: Base_2_2/Zone/VertexFields/materialID
list: float32
- name: Base_2_2/Zone/VertexFields/materialID_times
list: float64
- name: Base_2_2/Zone_times
list: float64
- name: Global/config
list: string
- name: Global/fracture energy
list: float32
- name: Global/fref
list: float32
- name: Global/fref_times
list: float64
- name: Global/pfThres
list: float32
- name: Global/pfThres_times
list: float64
- name: Global/strain energy
list: float32
- name: Global/total energy
list: float32
- name: Global/x-force
list: float32
- name: Global/y-force
list: float32
splits:
- name: res_SENS
num_bytes: 3436764335
num_examples: 28
download_size: 1916893989
dataset_size: 3436764335
configs:
- config_name: default
data_files:
- split: res_SENS
path: data/res_SENS-*
---
<p align='center'>
<img src='https://i.ibb.co/gZtL8VrY/force-ASR-samples.gif' alt='https://i.ibb.co/gZtL8VrY/force-ASR-samples.gif' width='1000'/>
</p>
```yaml
legal:
owner: RK 2423 FRASCAL (https://zenodo.org/records/7445749)
license: cc-by-4.0
data_production:
physics: phase-field fracture models for brittle fracture
type: simulation
script: Subset 'res-SENS' of the initial dataset, 1/5th time steps, converted to
PLAID format for standardized access; no changes to data content.
num_samples:
res_SENS: 28
storage_backend: hf_datasets
plaid:
version: 0.1.13.dev1+gb350f274a
```
This dataset was generated with [`plaid`](https://plaid-lib.readthedocs.io/), we refer to this documentation for additional details on how to extract data from `plaid_sample` objects.
The simplest way to use this dataset is to first download it:
```python
from plaid.storage import download_from_hub
repo_id = "channel/dataset"
local_folder = "downloaded_dataset"
download_from_hub(repo_id, local_folder)
```
Then, to iterate over the dataset and instantiate samples:
```python
from plaid.storage import init_from_disk
local_folder = "downloaded_dataset"
split_name = "train"
datasetdict, converterdict = init_from_disk(local_folder)
dataset = datasetdict[split]
converter = converterdict[split]
for i in range(len(dataset)):
plaid_sample = converter.to_plaid(dataset, i)
```
It is possible to stream the data directly:
```python
from plaid.storage import init_streaming_from_hub
repo_id = "channel/dataset"
datasetdict, converterdict = init_streaming_from_hub(repo_id)
dataset = datasetdict[split]
converter = converterdict[split]
for sample_raw in dataset:
plaid_sample = converter.sample_to_plaid(sample_raw)
```
Plaid samples' features can be retrieved like the following:
```python
from plaid.storage import load_problem_definitions_from_disk
local_folder = "downloaded_dataset"
pb_defs = load_problem_definitions_from_disk(local_folder)
# or
from plaid.storage import load_problem_definitions_from_hub
repo_id = "channel/dataset"
pb_defs = load_problem_definitions_from_hub(repo_id)
pb_def = pb_defs[0]
plaid_sample = ... # use a method from above to instantiate a plaid sample
for t in plaid_sample.get_all_time_values():
for path in pb_def.get_in_features_identifiers():
plaid_sample.get_feature_by_path(path=path, time=t)
for path in pb_def.get_out_features_identifiers():
plaid_sample.get_feature_by_path(path=path, time=t)
```
For those familiar with HF's `datasets` library, raw data can be retrieved without using the `plaid` library:
```python
from datasets import load_dataset
repo_id = "channel/dataset"
datasetdict = load_dataset(repo_id)
for split_name, dataset in datasetdict.items():
for raw_sample in dataset:
for feat_name in dataset.column_names:
feature = raw_sample[feat_name]
```
Notice that raw data refers to the variable features only, with a specific encoding for time variable features.
|