File size: 3,265 Bytes
93cf065
 
36a30ba
 
 
 
666b630
36a30ba
4b9f728
07a3468
85f20da
f6200be
50ed0c8
 
 
 
 
221bf26
50ed0c8
 
 
 
 
 
6dcea4e
324cda7
4b9f728
 
6dcea4e
6642934
 
 
cac47c0
7be2261
cac47c0
fe9b7a0
6642934
cac47c0
6642934
 
 
6dcea4e
 
6642934
4b9f728
 
 
 
 
 
 
 
6dcea4e
6642934
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50ed0c8
6642934
 
 
 
 
 
 
 
 
 
 
 
 
 
7d9d02d
6642934
e2a2c69
78c2e74
 
 
 
 
 
 
6fae9bb
78c2e74
 
 
 
 
50ed0c8
 
 
 
 
78c2e74
50ed0c8
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107

---
language:
- en
tags:
- bioassay
pretty_name: ChAFF
size_categories:
- 100K<n<1M
dataset_info:
  - config_name: ChAFF
    features:
    - name: Type
      dtype: string
    - name: DatasetName
      dtype: string
    - name: AID
      dtype: string
    - name: ID
      dtype: string
    - name: IDType
      dtype: string
    - name: SMILES
      dtype: string
---

# ChAFF datasets
This dataset collection contains ~200K curated Active compound lists from ~90 different BioAssay datasets, focusing on known assay interference artifacts. We applied SMILES standardization using RDKit and MolVS, including molecule sanitization and fragment removal. The final dataset is suitable for training and evaluating machine learning models.

## Types and Number of Active Compounds
| Type                 | NumActiveCompounds |
-----------------------|--------------------| 
| Absorbance           | 1486               |
| Artifact             | 10952              |
| Autofluoresence      | 32054              |
| ColloidalAggregators | 19553              |
| HeavyHitters         | 71981              |
| LuciferaseInhibition | 32831              |
| Misannotation        | 39                 |
| Reactivity           | 3107               |
| REDOX                | 217                |


## Dataset Columns
| Column     | Description                           |
|------------|-------------------------------------- |
| Type       | Task domain (e.g. Absorbance)         |
| DatasetName| Source dataset name                   |
| AID        | Pubchem Assay ID                      |
| ID         | Identifier for the compound           |
| IDType     | Type of identifier (e.g. CID)         |
| SMILES     | Curated SMILES                        |

Datasets can be found in the [data folder](https://huggingface.co/datasets/maomlab/ChAFF/tree/main/data).


## Dataset summary
A summary file is uploaded, which lists:
- Type
- DatasetName
- AID
- NumActiveCompounds
- PaperTitle
- Reference
- URL
- AssayName
- Description

Dataset summary file can be found: ChAFF_dataset_summary.json


# License
Each dataset comes from different sources (i.e., PubChem, Papers). Please check our dataset summary file if you are looking for references.


## Usage
Load a dataset in python

Each subset can be loaded into python using the Huggingface [datasets](https://huggingface.co/docs/datasets/index) library.
First, from the command line install the `datasets` library

    $ pip install datasets

then, from within python load the datasets library.

    >>> import datasets
    >>> from datasets import load_dataset, Features, Value

Specifiy column types to prevent pyarrow error.
```python
features = Features({
    "Type": Value("string"),
    "DatasetName": Value("string"),
    "AID": Value("string"), # Treat int as string
    "ID": Value("string"),
    "IDType": Value("string"),
    "SMILES": Value("string")
})
```

Now load one of the 'ChAFF' datasets, e.g.,

    >>> dataset = datasets.load_dataset("maomlab/ChAFF", name = "default", data_files = "data/Absorbance.csv", split = "train", features = features)


You can modify "data/Absorbance.csv" based on your interest (e.g., "data/Reactivity.csv"). 
The default is split = "train" as we did not split the datasets.