introvoyz041 haneulpark commited on
Commit
8cbb1ab
·
verified ·
0 Parent(s):

Duplicate from maomlab/MolData

Browse files

Co-authored-by: Haneul Park <haneulpark@users.noreply.huggingface.co>

.gitattributes ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.lz4 filter=lfs diff=lfs merge=lfs -text
12
+ *.mds filter=lfs diff=lfs merge=lfs -text
13
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
14
+ *.model filter=lfs diff=lfs merge=lfs -text
15
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
16
+ *.npy filter=lfs diff=lfs merge=lfs -text
17
+ *.npz filter=lfs diff=lfs merge=lfs -text
18
+ *.onnx filter=lfs diff=lfs merge=lfs -text
19
+ *.ot filter=lfs diff=lfs merge=lfs -text
20
+ *.parquet filter=lfs diff=lfs merge=lfs -text
21
+ *.pb filter=lfs diff=lfs merge=lfs -text
22
+ *.pickle filter=lfs diff=lfs merge=lfs -text
23
+ *.pkl filter=lfs diff=lfs merge=lfs -text
24
+ *.pt filter=lfs diff=lfs merge=lfs -text
25
+ *.pth filter=lfs diff=lfs merge=lfs -text
26
+ *.rar filter=lfs diff=lfs merge=lfs -text
27
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
28
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
29
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
30
+ *.tar filter=lfs diff=lfs merge=lfs -text
31
+ *.tflite filter=lfs diff=lfs merge=lfs -text
32
+ *.tgz filter=lfs diff=lfs merge=lfs -text
33
+ *.wasm filter=lfs diff=lfs merge=lfs -text
34
+ *.xz filter=lfs diff=lfs merge=lfs -text
35
+ *.zip filter=lfs diff=lfs merge=lfs -text
36
+ *.zst filter=lfs diff=lfs merge=lfs -text
37
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
38
+ # Audio files - uncompressed
39
+ *.pcm filter=lfs diff=lfs merge=lfs -text
40
+ *.sam filter=lfs diff=lfs merge=lfs -text
41
+ *.raw filter=lfs diff=lfs merge=lfs -text
42
+ # Audio files - compressed
43
+ *.aac filter=lfs diff=lfs merge=lfs -text
44
+ *.flac filter=lfs diff=lfs merge=lfs -text
45
+ *.mp3 filter=lfs diff=lfs merge=lfs -text
46
+ *.ogg filter=lfs diff=lfs merge=lfs -text
47
+ *.wav filter=lfs diff=lfs merge=lfs -text
48
+ # Image files - uncompressed
49
+ *.bmp filter=lfs diff=lfs merge=lfs -text
50
+ *.gif filter=lfs diff=lfs merge=lfs -text
51
+ *.png filter=lfs diff=lfs merge=lfs -text
52
+ *.tiff filter=lfs diff=lfs merge=lfs -text
53
+ # Image files - compressed
54
+ *.jpg filter=lfs diff=lfs merge=lfs -text
55
+ *.jpeg filter=lfs diff=lfs merge=lfs -text
56
+ *.webp filter=lfs diff=lfs merge=lfs -text
57
+ # Video files - compressed
58
+ *.mp4 filter=lfs diff=lfs merge=lfs -text
59
+ *.webm filter=lfs diff=lfs merge=lfs -text
MolData/test-00000-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af2a847e62073dd1e6055e7ba9973631de1326d88c765125cf8348ca361fbe36
3
+ size 184187236
MolData/test-00001-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:212e7510bbf240ae82dd7ae40d422a0a7e7f37ecd571b5ff0ec07aacc8b8fd9a
3
+ size 184124411
MolData/test-00002-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30ba93f0845e880c42abbdd9883b02519da52d0ce58f65bb6de79b9735938357
3
+ size 188741798
MolData/train-00000-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4f010c610c33017521af6f6ba20e93ee20dec28ddcf2fac94ddc9cb663ef78e
3
+ size 202668267
MolData/train-00001-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d751742aaebdbd8ad0d9deafa0a59b7061718579c21a1fce6fa73fe79b97df2a
3
+ size 202394600
MolData/train-00002-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b46fa3dbc65e8037b96ca6f20483cb1d82f7bd6d9ed53936cd2d49ee6229c1d
3
+ size 202664327
MolData/train-00003-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b3b457d8a68f8b5db4ff210a05492f6b82bd2eb0724bf3aee928664d92a4923
3
+ size 202347680
MolData/train-00004-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:235269732481559388ab474e3240d0d576d75e4ce7a77b4ef37a24d90ecce650
3
+ size 202653252
MolData/train-00005-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:362fe3963319cda5f3974a560fab488011b0faa877092d7171bffd2d2959bfa9
3
+ size 202519765
MolData/train-00006-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f641301fd2cffcfaafdf18184f083fdc14cad69bf7d92bd4ae32890172676ba
3
+ size 202953150
MolData/train-00007-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3b21ab5a430107f52245d7e98a7909524244bb1a17d3284b823f3b971d95d0a
3
+ size 202809599
MolData/train-00008-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeddaa730acc7efbcc43d87fdbf35410edd76fa63854f768d716265d889c6cce
3
+ size 202713668
MolData/train-00009-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46fc737d1bee8a652a793e690944dae96c7a6c49f546cfd752f54f9f9a80acba
3
+ size 202653673
MolData/train-00010-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bddf0fc2917448b14b41325ccc87ca5c537be7fe331ae25fe14b7469413fa19b
3
+ size 202819162
MolData/train-00011-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01b749f237d3f4ad93710127263f49b92be464616640e8fa7f2d7ce9ecf18b2c
3
+ size 202896523
MolData/train-00012-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:528be663423c388c8687ed449dcc0217b3e13f3b3d5f8d848bdac92d6670203a
3
+ size 202881935
MolData/train-00013-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ac0fe1ebc0170e2537e232bb279e2c70458394094aec915ed4adcc1101f339e
3
+ size 202478843
MolData/train-00014-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3394e4a8a99911cfacbaa7cfa346f7102315fd27b9cf7df27c72bc9a338ff20a
3
+ size 203190431
MolData/train-00015-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:280522747b3d4629cb70564f38d5a23afa35dd2f0627be19512ca51ee50f6f49
3
+ size 202724096
MolData/train-00016-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d88731025743868e32380b1a0246887c0a3a94204066bdb9dfc063d0d4a5101
3
+ size 202911131
MolData/train-00017-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:956fa0829d6a58339ab5244338691896582f280e5206bad5850935b0c39d01ac
3
+ size 204766491
MolData/train-00018-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a078eb98dbdd6908343e91f48bfb5706e6c6b006d45023e5649608fcd58cc21
3
+ size 213013580
MolData/train-00019-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e16c39c93a1efebe5359b50245de54708e8074d2def24d2ba417636ed4db474d
3
+ size 212642076
MolData/train-00020-of-00021.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:047896ff291dcbc727a76098f9390c042dc7bf75fd4e4b00cd4752756b265473
3
+ size 215801462
MolData/validation-00000-of-00002.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:369a83f4f8098a1dd05f0a496374c40e3cfd3dd4d62dc62f37069c31a54cdcde
3
+ size 206879527
MolData/validation-00001-of-00002.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e66925bea34bb2c9b419b3c22761d33e0b2470a517fd27869d5598f788292e36
3
+ size 205720624
MolData_preprocessing.py ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # This is a script for MolData dataset preprocessing
2
+
3
+ # 1. Load modules
4
+ import pandas as pd
5
+ import numpy as np
6
+ import urllib.request
7
+ import rdkit
8
+ from rdkit import Chem
9
+ import os
10
+ import molvs
11
+ import csv
12
+ import json
13
+ import tqdm
14
+
15
+ standardizer = molvs.Standardizer()
16
+ fragment_remover = molvs.fragment.FragmentRemover()
17
+
18
+
19
+ # 2. Download the original dataset
20
+ # https://github.com/LumosBio/MolData
21
+ # Suppose that 'all_molecular_data.csv' has been downloaded from GitHub
22
+
23
+
24
+ # 3. Check if any SMILES is missing in the dataset (first column)
25
+ df = pd.read_csv('all_molecular_data.csv')
26
+
27
+ missing_SMILES = df[df.iloc[:, 0].isna()]
28
+
29
+ print(f'There are {len(missing_SMILES)} rows with missing SMILES.') # This prints 'There are 0 rows with missing SMILES.'
30
+
31
+
32
+ # 4. Sanitize SMILES with MolVS and print problems
33
+ # Since the dataset is large, we divided it into four portions to sanitize
34
+ quarter_df_1 = df.iloc[:len(df)//4]
35
+
36
+ quarter_df_1['X'] = [ \
37
+ rdkit.Chem.MolToSmiles(
38
+ fragment_remover.remove(
39
+ standardizer.standardize(
40
+ rdkit.Chem.MolFromSmiles(
41
+ smiles))))
42
+ for smiles in quarter_df_1['smiles']]
43
+
44
+ problems = []
45
+ for index, row in tqdm.tqdm(quarter_df_1.iterrows()):
46
+ result = molvs.validate_smiles(row['X'])
47
+ if len(result) == 0:
48
+ continue
49
+ problems.append((row['X'], result))
50
+
51
+ # Most are because it includes the salt form and/or it is not neutralized
52
+ for result, alert in problems:
53
+ print(f"SMILES: {result}, problem: {alert[0]}")
54
+
55
+ quarter_df_1.to_csv('MolData_sanitized_0.25.csv')
56
+
57
+
58
+
59
+ quarter_df_2 = df.iloc[len(df)//4 : len(df)//2]
60
+
61
+ quarter_df_2['X'] = [ \
62
+ rdkit.Chem.MolToSmiles(
63
+ fragment_remover.remove(
64
+ standardizer.standardize(
65
+ rdkit.Chem.MolFromSmiles(
66
+ smiles))))
67
+ for smiles in quarter_df_2['smiles']]
68
+
69
+ problems = []
70
+ for index, row in tqdm.tqdm(quarter_df_2.iterrows()):
71
+ result = molvs.validate_smiles(row['X'])
72
+ if len(result) == 0:
73
+ continue
74
+ problems.append((row['X'], result))
75
+
76
+ # Most are because it includes the salt form and/or it is not neutralized
77
+ for result, alert in problems:
78
+ print(f"SMILES: {result}, problem: {alert[0]}")
79
+
80
+ quarter_df_2.to_csv('MolData_sanitized_0.5.csv')
81
+
82
+
83
+ quarter_df_3 = df.iloc[len(df)//2 : 3 *len(df)//4]
84
+
85
+ quarter_df_3['X'] = [ \
86
+ rdkit.Chem.MolToSmiles(
87
+ fragment_remover.remove(
88
+ standardizer.standardize(
89
+ rdkit.Chem.MolFromSmiles(
90
+ smiles))))
91
+ for smiles in quarter_df_3['smiles']]
92
+
93
+ problems = []
94
+ for index, row in tqdm.tqdm(quarter_df_3.iterrows()):
95
+ result = molvs.validate_smiles(row['X'])
96
+ if len(result) == 0:
97
+ continue
98
+ problems.append((row['X'], result))
99
+
100
+ # Most are because it includes the salt form and/or it is not neutralized
101
+ for result, alert in problems:
102
+ print(f"SMILES: {result}, problem: {alert[0]}")
103
+
104
+ quarter_df_3.to_csv('MolData_sanitized_0.75.csv')
105
+
106
+
107
+
108
+ quarter_df_4 = df.iloc[3 *len(df)//4 :len(df)]
109
+
110
+ quarter_df_4['X'] = [ \
111
+ rdkit.Chem.MolToSmiles(
112
+ fragment_remover.remove(
113
+ standardizer.standardize(
114
+ rdkit.Chem.MolFromSmiles(
115
+ smiles))))
116
+ for smiles in quarter_df_4['smiles']]
117
+
118
+ problems = []
119
+ for index, row in tqdm.tqdm(quarter_df_4.iterrows()):
120
+ result = molvs.validate_smiles(row['X'])
121
+ if len(result) == 0:
122
+ continue
123
+ problems.append((row['X'], result))
124
+
125
+ # Most are because it includes the salt form and/or it is not neutralized
126
+ for result, alert in problems:
127
+ print(f"SMILES: {result}, problem: {alert[0]}")
128
+
129
+ quarter_df_4.to_csv('MolData_sanitized_1.0.csv')
130
+
131
+
132
+ # 4. Concatenate
133
+ sanitized1 = pd.read_csv('MolData_sanitized_0.25.csv')
134
+ sanitized2 = pd.read_csv('MolData_sanitized_0.5.csv')
135
+ sanitized3 = pd.read_csv('MolData_sanitized_0.75.csv')
136
+ sanitized4 = pd.read_csv('MolData_sanitized_1.0.csv')
137
+
138
+ smiles_concatenated = pd.concat([sanitized1, sanitized2, sanitized3, sanitized4], ignore_index=True)
139
+
140
+ smiles_concatenated.to_csv('MolData_sanitized_concatenated.csv', index = False)
141
+
142
+
143
+
144
+
145
+ # 5. Formatting and naming (wide form to long form, & column naming)
146
+ # Due to the large size of the dataset, we processed it using chunks to efficiently handle the data.
147
+ chunk_size = 10**5
148
+ input_file = 'MolData_sanitized_concatenated.csv'
149
+ output_prefix = 'MolData_long_form_'
150
+
151
+ column_names = pd.read_csv(input_file, nrows=1).columns
152
+ column_names = column_names.tolist()
153
+
154
+ column_names = ['SMILES' if col == 'X' else col for col in column_names]
155
+
156
+ var_name_list = [col for col in column_names if col.startswith('activity_')]
157
+
158
+ with pd.read_csv(input_file, chunksize=chunk_size) as reader:
159
+ for i, chunk in enumerate(reader):
160
+ chunk.columns = column_names
161
+
162
+ long_df = pd.melt(chunk, id_vars=['SMILES', 'PUBCHEM_CID', 'split'],
163
+ value_vars=var_name_list, var_name='AID', value_name='score')
164
+
165
+ long_df = long_df.dropna(subset=['score'])
166
+ long_df['score'] = long_df['score'].astype('Int64')
167
+
168
+ output_file = f"{output_prefix}{i+1}.csv"
169
+ long_df.to_csv(output_file, index=False)
170
+
171
+ print(f"Saved: {output_file}")
172
+
173
+
174
+
175
+ # 6. Split into train, test, and validation
176
+ chunk_size = 10**5
177
+ input_files = [f'MolData_long_form_{i+1}.csv' for i in range(15)]
178
+
179
+ output_train_file = 'MolData_train.csv'
180
+ output_test_file = 'MolData_test.csv'
181
+ output_valid_file = 'MolData_validation.csv'
182
+
183
+ train_data = []
184
+ test_data = []
185
+ valid_data = []
186
+
187
+ for input_file in input_files:
188
+ with pd.read_csv(input_file, chunksize=chunk_size) as reader:
189
+ for chunk in reader:
190
+ train_chunk = chunk[chunk['split'] == 'train']
191
+ test_chunk = chunk[chunk['split'] == 'test']
192
+ valid_chunk = chunk[chunk['split'] == 'validation']
193
+
194
+ train_data.append(train_chunk)
195
+ test_data.append(test_chunk)
196
+ valid_data.append(valid_chunk)
197
+
198
+ train_df = pd.concat(train_data, ignore_index=True)
199
+ test_df = pd.concat(test_data, ignore_index=True)
200
+ valid_df = pd.concat(valid_data, ignore_index=True)
201
+
202
+ train_df.to_csv(output_train_file, index=False)
203
+ test_df.to_csv(output_test_file, index=False)
204
+ valid_df.to_csv(output_valid_file, index=False)
205
+
206
+
207
+ def fix_cid_column(df):
208
+ df['PUBCHEM_CID'] = df['PUBCHEM_CID'].astype(str).apply(lambda x: x.split(',')[0]) # Because some molecule have two CIDs
209
+ df['PUBCHEM_CID'] = df['PUBCHEM_CID'].astype('Int64')
210
+ df = df.rename(columns = {'score' : 'Y'}) # This is for column renaming
211
+ return df
212
+
213
+ train_csv = fix_cid_column(pd.read_csv('MolData_train.csv'))
214
+ test_csv = fix_cid_column(pd.read_csv('MolData_test.csv'))
215
+ valid_csv = fix_cid_column(pd.read_csv('MolData_validation.csv'))
216
+
217
+ train_csv.to_parquet('MolData_train.parquet', index=False)
218
+ test_csv.to_parquet('MolData_test.parquet', index=False)
219
+ valid_csv.to_parquet('MolData_validation.parquet', index=False)
220
+
README.md ADDED
@@ -0,0 +1,178 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ version: 1.0.0
3
+ license: cc-by-sa-4.0
4
+ task_categories:
5
+ - tabular-classification
6
+ language:
7
+ - en
8
+ pretty_name: MolData
9
+ size_categories:
10
+ - 1M<n<10M
11
+ tags:
12
+ - drug discovery
13
+ - bioassay
14
+ dataset_summary: A comprehensive disease and target-based dataset with roughly 170
15
+ million drug screening results from 1.4 million unique molecules and 600 assays
16
+ which are collected from PubChem to accelerate molecular machine learning for better
17
+ drug discovery.
18
+ citation: "@article{KeshavarziArshadi2022,\n title = {MolData, a molecular benchmark\
19
+ \ for disease and target based machine learning},\n volume = {14},\n ISSN = {1758-2946},\n\
20
+ \ url = {http://dx.doi.org/10.1186/s13321-022-00590-y},\n DOI = {10.1186/s13321-022-00590-y},\n\
21
+ \ number = {1},\n journal = {Journal of Cheminformatics},\n publisher = {Springer\
22
+ \ Science and Business Media LLC},\n author = {Keshavarzi Arshadi, Arash and Salem,\
23
+ \ Milad and Firouzbakht, Arash and Yuan, Jiann Shiun},\n year = {2022},\n month\
24
+ \ = mar \n}"
25
+ dataset_info:
26
+ config_name: MolData
27
+ features:
28
+ - name: SMILES
29
+ dtype: string
30
+ - name: PUBCHEM_CID
31
+ dtype: int64
32
+ - name: PUBCHEM_AID
33
+ dtype: string
34
+ - name: Y
35
+ dtype: int64
36
+ splits:
37
+ - name: train
38
+ num_bytes: 10140424890
39
+ num_examples: 138547273
40
+ - name: test
41
+ num_bytes: 1288513312
42
+ num_examples: 17069726
43
+ - name: validation
44
+ num_bytes: 961758159
45
+ num_examples: 12728449
46
+ download_size: 5262157307
47
+ dataset_size: 12390696361
48
+ configs:
49
+ - config_name: MolData
50
+ data_files:
51
+ - split: train
52
+ path: MolData/train-*
53
+ - split: test
54
+ path: MolData/test-*
55
+ - split: validation
56
+ path: MolData/validation-*
57
+ - config_name: default
58
+ data_files:
59
+ - split: train
60
+ path: data/train-*
61
+ - split: test
62
+ path: data/test-*
63
+ - split: validation
64
+ path: data/validation-*
65
+ ---
66
+
67
+ # MolData
68
+
69
+ [MolData](https://jcheminf.biomedcentral.com/articles/10.1186/s13321-022-00590-y) is a comprehensive disease and target-based dataset collected from PubChem.
70
+ The dataset contains 1.4 million unique molecules, and it is one the largest efforts to date for democratizing the molecular machine learning.
71
+ This is a mirror of the [Official Github repo](https://github.com/LumosBio/MolData/tree/main/Data) where the dataset was uploaded in 2021.
72
+
73
+
74
+ ## Preprocessing
75
+
76
+ We utilized the raw data uploaded on [Github](https://github.com/LumosBio/MolData) and performed several preprocessing:
77
+ 1. Sanitize the molecules using RDKit and MolVS (standardize SMILES format)
78
+ 2. Formatting (from wide form to long form)
79
+ 3. Rename the columns
80
+ 4. Split the dataset (train, test, validation)
81
+
82
+ If you would like to try these processes with the original dataset,
83
+ please follow the instructions in the [preprocessing script](https://huggingface.co/datasets/maomlab/MolData/blob/main/MolData_preprocessing.py) file located in our MolData repository.
84
+
85
+
86
+
87
+ ## Quickstart Usage
88
+
89
+ ### Load a dataset in python
90
+ Each subset can be loaded into python using the Huggingface [datasets](https://huggingface.co/docs/datasets/index) library.
91
+ First, from the command line install the `datasets` library
92
+
93
+ $ pip install datasets
94
+
95
+ then, from within python load the datasets library
96
+
97
+ >>> import datasets
98
+
99
+ and load the `MolData` datasets, e.g.,
100
+
101
+ >>> MolData = datasets.load_dataset("maomlab/MolData", name = "MolData")
102
+ Generating train split: 100%|███████████████████████████████████████████████████████████████████████████████████████████| 138547273/138547273 [02:07<00:00, 1088043.12 examples/s]
103
+ Generating test split: 100%|██████████████████████████████████████████████████████████████████████████████████████████████| 17069726/17069726 [00:16<00:00, 1037407.67 examples/s]
104
+ Generating validation split: 100%|████████████████████████████████████████████████████████████████████████████████████████| 12728449/12728449 [00:11<00:00, 1093675.24 examples/s]
105
+
106
+ and inspecting the loaded dataset
107
+
108
+ >>> MolData
109
+ DatasetDict({
110
+ train: Dataset({
111
+ features: ['SMILES', 'PUBCHEM_CID', 'PUBCHEM_AID', 'Y'],
112
+ num_rows: 138547273
113
+ })
114
+ test: Dataset({
115
+ features: ['SMILES', 'PUBCHEM_CID', 'PUBCHEM_AID', 'Y'],
116
+ num_rows: 17069726
117
+ })
118
+ validation: Dataset({
119
+ features: ['SMILES', 'PUBCHEM_CID', 'PUBCHEM_AID', 'Y'],
120
+ num_rows: 12728449
121
+ })
122
+ })
123
+
124
+ ### Use a dataset to train a model
125
+ One way to use the dataset is through the [MolFlux](https://exscientia.github.io/molflux/) package developed by Exscientia.
126
+ First, from the command line, install `MolFlux` library with `catboost` and `rdkit` support
127
+
128
+ pip install 'molflux[catboost,rdkit]'
129
+
130
+ then load, featurize, split, fit, and evaluate the catboost model
131
+
132
+ import json
133
+ from datasets import load_dataset
134
+ from molflux.datasets import featurise_dataset
135
+ from molflux.features import load_from_dicts as load_representations_from_dicts
136
+ from molflux.splits import load_from_dict as load_split_from_dict
137
+ from molflux.modelzoo import load_from_dict as load_model_from_dict
138
+ from molflux.metrics import load_suite
139
+
140
+ Split and evaluate the catboost model
141
+
142
+ split_dataset = load_dataset('maomlab/MolData', name = 'MolData')
143
+
144
+ split_featurised_dataset = featurise_dataset(
145
+ split_dataset,
146
+ column = "SMILES",
147
+ representations = load_representations_from_dicts([{"name": "morgan"}, {"name": "maccs_rdkit"}]))
148
+
149
+ model = load_model_from_dict({
150
+ "name": "cat_boost_classifier",
151
+ "config": {
152
+ "x_features": ['SMILES::morgan', 'SMILES::maccs_rdkit'],
153
+ "y_features": ['Y']}})
154
+
155
+ model.train(split_featurised_dataset["train"])
156
+ preds = model.predict(split_featurised_dataset["test"])
157
+
158
+ classification_suite = load_suite("classification")
159
+
160
+ scores = classification_suite.compute(
161
+ references=split_featurised_dataset["test"]['Y'],
162
+ predictions=preds["cat_boost_classifier::Y"])
163
+
164
+
165
+ ### Citation
166
+ @article{KeshavarziArshadi2022,
167
+ title = {MolData, a molecular benchmark for disease and target based machine learning},
168
+ volume = {14},
169
+ ISSN = {1758-2946},
170
+ url = {http://dx.doi.org/10.1186/s13321-022-00590-y},
171
+ DOI = {10.1186/s13321-022-00590-y},
172
+ number = {1},
173
+ journal = {Journal of Cheminformatics},
174
+ publisher = {Springer Science and Business Media LLC},
175
+ author = {Keshavarzi Arshadi, Arash and Salem, Milad and Firouzbakht, Arash and Yuan, Jiann Shiun},
176
+ year = {2022},
177
+ month = mar
178
+ }