Spaces:
Running
Running
Upload 45 files
Browse files- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3.md +458 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_cleaned.md +205 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_content_list.json +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_layout.pdf +3 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_middle.json +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_model.json +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_origin.pdf +3 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_span.pdf +3 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/dag.json +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/032759948e20868c031881dac87a152765d78ea757955c32f34b83ca8e975d1d.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/07aae76295c011d4cdd34a6c00be2fe8427447017185173c5e25e9468ccf833d.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/084cd722defc01e058c1747c1103ac680c0cd8217a93077cab2a30ae06000a37.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/16005489fb642f13912d85a6d90523ef0a5e1d56b5c30d9d2ee35336fdb04f4f.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/1695d160ca54e8992fed88c81378f03bd31abccebafebc3c6a3d5dcadc6a747f.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/1ee1e5bd4e35d7acf3b80e7404b113daf11b6925490369aaa637e60964fb744d.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/27c884c07b2f80cded2a18e590a81c38cc2c409bbb73f478c4432366bc9ede6d.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/35c4ba087f1ae7299aa554638aef5647f9efa2de01450fec172beb01bf64538e.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/43ac77c0fc35ecea17e2a91075a4f1c272643487659bc83241629c4a80c6ae86.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/5237f9010e1b7bb79b84bdc91c83cb5152b351c1f8b05fe25b9f0b961f759e2d.jpg +3 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/55d0ac7d29bb938c1a28dc8f66121322e7c07f72ae0f7cfa10cb5cd784b67c21.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/6b6a48b7abea8c9dc0c3c72345c9c94cbb0ac1d1ef8824d485ff0ace9b5a0a1e.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/6c50a187dd8276b7372af9e7e00b99521d3741fa210eee7a1af570b334053570.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/71b0bba244aef7a34cb53a9293a46808b73589fc9a53aa7b50884d5e277c2b36.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/7ad84d22cd3c911b24b3de89b1077c3b9f08c6b762840ecd38160beb882f3a16.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/7c18349955cc82bf082d353ddb8f1ff323edb09a34c2ad852dba320d3a0a3faa.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/88fc24d2ba2e08617faed6c58a55f1e92d5803ea4ae7a49eaaf118fe4ad56429.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/977b41496865477e7652249bec51630a1a037097fa74f17c10e8a72851cd7ce3.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/b1279274391a8ee385cd510483726424c1b7bb542dd3dfcaa664dd7784591ce0.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/b5f78fc1f86d0c7c6f996497a567b0462a6e0529447787d81ffe3fdbb71b3ccb.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/bdae202d54ee161317731819baf6967d189260b12ff333dc514a8cc065c47f97.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/d1786526fd710a309dcca5721e53bf82d19ca2622ac3314d3bfc46284f668f1d.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/dcbdd7f1637fddc90c9a9616e1e732fa41edaf190414db6aba1fd62b99a32716.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/e78ea26b8e8328ff5cf96a82a322d391d8879301b032bec5c51b78065f043e34.jpg +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/success_dag.txt +0 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/auto/visual_dag.json +130 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/1 INTRODUCTION.md_dag.json +146 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/2 PRELIMINARIES.md_dag.json +104 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/3 THE PROPOSED MOLSPECTRA METHOD.md_dag.json +187 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/4 EXPERIMENTS.md_dag.json +187 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/6 CONCLUSION.md_dag.json +45 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/1 INTRODUCTION.md +19 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/2 PRELIMINARIES.md +40 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/3 THE PROPOSED MOLSPECTRA METHOD.md +56 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/4 EXPERIMENTS.md +73 -0
- mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/6 CONCLUSION.md +3 -0
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3.md
ADDED
|
@@ -0,0 +1,458 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# MOLSPECTRA: PRE-TRAINING 3D MOLECULAR REP-RESENTATION WITH MULTI-MODAL ENERGY SPECTRA
|
| 2 |
+
|
| 3 |
+
Liang Wang1,2∗Shaozhen Liu1 Yu Rong3† Deli Zhao3 Qiang Liu1,2† Shu Wu1,2 Liang Wang1,2
|
| 4 |
+
|
| 5 |
+
1New Laboratory of Pattern Recognition (NLPR),
|
| 6 |
+
State Key Laboratory of Multimodal Artificial Intelligence Systems (MAIS),
|
| 7 |
+
Institute of Automation, Chinese Academy of Sciences (CASIA)
|
| 8 |
+
2School of Artificial Intelligence, University of Chinese Academy of Sciences
|
| 9 |
+
3DAMO Academy, Alibaba Group
|
| 10 |
+
|
| 11 |
+
# ABSTRACT
|
| 12 |
+
|
| 13 |
+
Establishing the relationship between 3D structures and the energy states of molecular systems has proven to be a promising approach for learning 3D molecular representations. However, existing methods are limited to modeling the molecular energy states from classical mechanics. This limitation results in a significant oversight of quantum mechanical effects, such as quantized (discrete) energy level structures, which offer a more accurate estimation of molecular energy and can be experimentally measured through energy spectra. In this paper, we propose to utilize the energy spectra to enhance the pre-training of 3D molecular representations (MolSpectra), thereby infusing the knowledge of quantum mechanics into the molecular representations. Specifically, we propose SpecFormer, a multispectrum encoder for encoding molecular spectra via masked patch reconstruction. By further aligning outputs from the 3D encoder and spectrum encoder using a contrastive objective, we enhance the 3D encoder’s understanding of molecules. Evaluations on public benchmarks reveal that our pre-trained representations surpass existing methods in predicting molecular properties and modeling dynamics.
|
| 14 |
+
|
| 15 |
+
# 1 INTRODUCTION
|
| 16 |
+
|
| 17 |
+
Learning 3D molecular representations from geometric conformations offers a promising approach for understanding molecular geometry and predicting quantum properties and interactions, which is significant in drug discovery and materials science (Musaelian et al., 2023; Batatia et al., 2022; Liao & Smidt, 2023; Wang et al., 2023b; Du et al., 2023b). Given the scarcity of molecular property labels, self-supervised representation pre-training has been proposed and utilized to provide generalizable representations (Hu et al., 2020; Rong et al., 2020; Ma et al., 2024).
|
| 18 |
+
|
| 19 |
+
In contrast to contrastive learning (Wang et al., 2022; Kim et al., 2022) and masked modeling (Hou et al., 2022; Liu et al., 2023c; Wang et al., 2024b) on 2D molecular graphs and molecular languages (e.g., SMILES), the design of pre-training strategies on 3D molecular geometries is more closely aligned with physical principles. Previous studies (Zaidi et al., 2023; Jiao et al., 2023) have guided representation learning through denoising processes on 3D molecular geometries, theoretically demonstrating that denoising 3D geometries is equivalent to learning molecular force fields, specifically the negative gradient of molecular potential energy with respect to position. Essentially, these studies reveal that establishing the relationship between 3D geometries and the energy states of molecular systems is an effective pathway to learn 3D molecular representations.
|
| 20 |
+
|
| 21 |
+
However, existing methods are limited to the continuous description (i.e., the potential energy function) of the molecular energy states within the classical mechanics, overlooking the quantized (discrete) energy level structures from the quantum mechanical perspective. From the quantum perspective, molecular systems exhibit quantized energy level structures, meaning that energy states can only assume specific discrete values. Specifically, different types of molecular motion, such as electronic, vibrational, and rotational motion, correspond to different energy level structures. Knowledge of these energy levels is crucial in molecular physics and quantum chemistry, as they determine the spectroscopic characteristics, chemical reactivity, and many other important molecular properties. Fortunately, experimental measurements of molecular energy spectra can reflect these structures. Meanwhile, there are many molecular spectra data obtained through experimental measurements or simulations (Zou et al., 2023; Alberts et al., 2024). Therefore, incorporating the knowledge of energy levels into molecular representation learning is expected to facilitate the development of more informative molecular representations.
|
| 22 |
+
|
| 23 |
+

|
| 24 |
+
Figure 1: The conceptual view of MolSpectra, which leverages both molecular conformation and spectra for pre-training. Prior works only model classical mechanics by denoising on conformations.
|
| 25 |
+
|
| 26 |
+
In this paper, we propose MolSpectra, a framework that incorporates molecular spectra into the pre-training of 3D molecular representations, thereby infusing the knowledge of quantized energy level structures into the representations, as shown in Figure 1. In MolSpectra, we introduce a multispectrum encoder, SpecFormer, to capture both intra-spectrum and inter-spectrum peak correlations by training with a masked patches reconstruction (MPR) objective. Additionally, we employ a contrastive objective to distills the spectral features and its inherent knowledge into the learning of 3D representations. After pre-training, the resulting 3D encoder can be fine-tuned for downstream tasks, providing expressive 3D molecular representations without the need for associated spectral data. Extensive experiments over different downstream molecular property prediction benchmarks shows the superiority of MolSpectra.
|
| 27 |
+
|
| 28 |
+
In summary, our contributions are as follows:
|
| 29 |
+
|
| 30 |
+
• We introduce quantized energy level structures and molecular spectra into 3D molecular representation pre-training for the first time, surpassing previous work that relied solely on physical knowledge within the scope of classical mechanics.
|
| 31 |
+
• We propose SpecFormer as an expressive multi-spectrum encoder, along with the masked patches reconstruction objective for spectral representation learning.
|
| 32 |
+
• We propose a contrastive objective to align molecular representations in the 3D modality and spectral modalities, enabling the pre-trained 3D encoder to infer molecular spectral features in downstream tasks without relying on spectral data.
|
| 33 |
+
• Experiments across different downstream benchmarks demonstrate that our method effectively enhances the expressiveness of the pre-trained 3D molecular representations.
|
| 34 |
+
|
| 35 |
+
# 2 PRELIMINARIES
|
| 36 |
+
|
| 37 |
+
# 2.1 NOTATIONS
|
| 38 |
+
|
| 39 |
+
Consider a molecule characterized by its 3D structure and spectra, represented as M = (a, x, S). Here, a ∈ {1, 2, . . . , 118}N specifies the atomic numbers, indicating the types of atoms within the molecule. The vector x ∈ R3N describes the conformation of the molecule, while S represents its spectra. The parameter N denotes the number of atoms in the molecule. Note that the atoms are arranged in the same order in both a and x, ensuring consistency between the atomic numbers and their corresponding spatial coordinates.
|
| 40 |
+
|
| 41 |
+
S = (s1, . . . , s|S|) represents the set of spectra for a molecule, where |S| denotes the number of spectrum types considered. In our study, we focus on three types, so |S| = 3. The first spectrum, s1 ∈ R601, is the UV-Vis spectrum, which spans from 1.5 to 13.5 eV with 601 data points at intervals of 0.02 eV. The second spectrum, s2 ∈ R3501, is the IR spectrum, covering a range from 500 to 4000 cm−1 with 3501 data points at intervals of 1 cm−1. The third spectrum, s3 ∈ R3501, is the Raman spectrum, with the same range and intervals as the IR spectrum. Together, these spectra provide a comprehensive description of the molecular characteristics across different spectral modalities.
|
| 42 |
+
|
| 43 |
+
# 2.2 PRE-TRAINING 3D MOLECULAR REPRESENTATION VIA DENOISING
|
| 44 |
+
|
| 45 |
+
Denoising has emerged as a prominent pre-training objective in 3D molecular representation learning, excelling in various downstream tasks. This method involves training models to predict and remove noise introduced deliberately into molecular structures. This approach is physically interpretable due to its proven equivalence to learning the molecular force field.
|
| 46 |
+
|
| 47 |
+
Equivalence between denoising and learning molecular force fields. The equivalence between coordinate denoising and force field learning is established by Zaidi et al. (2023). For a given molecule M, perturb its equilibrium structure x0 according to the distribution p(x|x0), where x is the noisy conformation. Assuming the molecular distribution adheres to the energy-based Boltzmann distribution with respect to the energy function E(·), then
|
| 48 |
+
|
| 49 |
+

|
| 50 |
+
|
| 51 |
+
where GNNθ(x) denotes a graph neural network parameterized by θ, which processes the conformation x to produce node-level predictions. The notation ≃ signifies the equivalence of different objectives. The proof of this equivalence is provided in the Appendix A. In prior research, the energy function E(·) has been defined in several forms. Below are three representative studies.
|
| 52 |
+
|
| 53 |
+
Energy function I: mixture of isotropic Gaussians. In Coord (Zaidi et al., 2023), the energy function is approximated using a mixture of isotropic Gaussians centered at the known equilibrium structures to replace the Boltzmann distribution, since these structures are local maxima of the Boltzman distribution. Leveraging the equivalence between the score-matching objective and denoising autoencoders (Vincent, 2011), the following denoising-based energy function ECoord(·) is derived:
|
| 54 |
+
|
| 55 |
+

|
| 56 |
+
|
| 57 |
+
Note that this objective is derived under the assumption of isotropic Gaussian noise, i.e., p(x|x0) ∼ N (x0, τ 2c I3N ), where I3N represents the identity matrix of size 3N , and the subscript c indicates the coordinate denoising approach.
|
| 58 |
+
|
| 59 |
+
Energy function II: mixture of anisotropic Gaussians. Considering rigid and flexible components in molecular structures, isotropic Gaussian can lead to significant approximation errors. To address the anisotropic distribution, Frad (Feng et al., 2023) introduces hybrid noise on dihedral angles of rotatable bonds and atomic coordinates, incorporating fractional denoising of the coordinate noise. The equilibrium structure x0 is initially perturbed by dihedral angle noise p(ψa|ψ0) ∼ N (ψ0, σ2f Im), followed by coordinate noise p(x|xa) ∼ N (xa, τ 2f I3N ). Here, ψa, ψ0 ∈ [0, 2π)m represent to the dihedral angles of rotatable bonds in structures xa and x0, respectively, with m denoting the number of rotatable bonds. The subscript f indicates the fractional denoising approach. Subsequently, the energy function is induced:
|
| 60 |
+
|
| 61 |
+

|
| 62 |
+
|
| 63 |
+
where Στf ,σf = τ 2f I3N + σ2f CC⊤, and C ∈ R3N×m is a matrix used to linearly transform the dihedral angle noise into coordinate change, expressed as ∆x ≈ C∆ψ.
|
| 64 |
+
|
| 65 |
+
Energy function III: classical potential energy theory. SliDe (Ni et al., 2024) derives energy function from classical molecular potential energy theory (Alavi, 2020; Zhou & Liu, 2022). In this
|
| 66 |
+
|
| 67 |
+

|
| 68 |
+
Figure 2: Overview of the MolSpectra pre-training framework. Our pre-training framework comprises three sub-objectives: the denoising objective and the MPR objective, which respectively guide the representation learning of the 3D and spectral modalities, and the contrastive objective, which aligns the representations of both modalities.
|
| 69 |
+
|
| 70 |
+
form, the total intramolecular potential energy is mainly attributed to three types of interactions: bond stretching, bond angle bending, and bond torsion. The following energy function is derived:
|
| 71 |
+
|
| 72 |
+

|
| 73 |
+
|
| 74 |
+
where r ∈ (R≥0)m1 , θ ∈ [0, 2π)m2 , ϕ ∈ [0, 2π)m3 represent vectors of the bond lengths, bond angles, and bond torsion angles of the molecule, respectively. r0, θ0, ϕ0 correspond to the respective equilibrium values. The parameter vectors kB, kA, kT determine the interaction strength.
|
| 75 |
+
|
| 76 |
+
# 3 THE PROPOSED MOLSPECTRA METHOD
|
| 77 |
+
|
| 78 |
+
Considering the complementarity of different spectra, we introduce multiple spectra into molecular representation learning. To effectively comprehend molecular spectra, we designed a Transformerbased multi-spectrum encoder, SpecFormer, along with a masked reconstruction objective to guide its training. Finally, a contrastive objective is employed to align the 3D encoding guided by the denoising objective with the spectra encoding guided by the reconstruction objective, endowing the 3D encoding with the capability to understand spectra and the knowledge they encompass.
|
| 79 |
+
|
| 80 |
+
# 3.1 SPECFORMER: A SINGLE-STREAM ENCODER FOR MULTI-MODAL ENERGY SPECTRA
|
| 81 |
+
|
| 82 |
+
For different types of spectra, each spectrum is independently patched and initially encoded. Then, all the resulting patch embeddings are concatenated and encoded using a Transformer-based encoder.
|
| 83 |
+
|
| 84 |
+
Patching. Compared to directly encoding individual frequency points, we divided each spectrum into multiple patches. This approach offers two distinct advantages: (i) By forming patches from adjacent frequency points, local semantic features, such as absorption peaks, can be captured more effectively. (ii) It reduces the computational overhead of subsequent Transformer layers. Technically, each spectrum si ∈ RLi where i = 1, · · · , |S| is first divided into patches according to the patch length Pi and the stride Di. When 0 < Di < Pi, the consecutive patches will be overlapped with overlapping region length Pi − Di. When Di = Pi, the consecutive patches will be nonoverlapped. Li denotes the length of si. The patching process on each spectrum will generate a sequence of patches pi ∈ RNi×Pi, where Ni = Li−PiDi + 1 is the number of patches.
|
| 85 |
+
|
| 86 |
+
Patch encoding and position encoding. Prior to be fed into the encoder, the patches of the i-th spectrum are mapped to the latent space of dimension d via a trainable linear projection Wi ∈ RPi×d. A learnable additive position encoding W posi ∈ RNi×d is applied to maintain the order of the patches: p′i = piWi + W posi , where p′i ∈ RNi×d denotes the latent representation of the spectrum si that will be fed into the subsequent SpecFormer encoder.
|
| 87 |
+
|
| 88 |
+
SpecFormer: multi-spectrum Transformer encoder. Although several encoders have been proposed to map molecular spectrum into implicit representations, such as the CNN-AM (Tao et al., 2024) based on one-dimensional convolution, these encoders are designed to encode only a single type of spectrum. In our approach, multiple molecular spectra (UV-Vis, IR, Raman) are jointly considered. When encoding multiple spectra of a molecule simultaneously, an observation caught our attention and led us to adopt a Transformer-based encoder with multiple spectra as input, similar to the single-stream Transformer in multi-modal learning (Shin et al., 2021).
|
| 89 |
+
|
| 90 |
+
The observation refers to the fact that the same functional group not only causes multiple peaks within a single spectrum, but also generates peaks across different spectra. As shown on the left of Figure 3, the different vibrational modes of the methyl group (-CH3) in methanol (CH3OH) result in three peaks in the IR spectrum, indicating intra-spectrum dependencies among these peaks. A similar phenomenon occurs with the hydroxyl group (-OH) in methanol. Additionally, the aromatic ring in phenol (C6H5OH), shown on the right of
|
| 91 |
+
|
| 92 |
+

|
| 93 |
+
Figure 3: Illustration of intra-spectrum (left) and interspectrum (right) dependencies.
|
| 94 |
+
|
| 95 |
+
Figure 3, not only produces multiple peaks in the IR spectrum due to different vibrational modes but also causes an absorption peak near 270 nm in the UV-Vis spectrum due to the π → π∗ transition in the aromatic ring, demonstrating the existence of inter-spectrum dependencies. Such dependencies have been theoretically studied, for example, in the context of vibronic coupling (Kong et al., 2021).
|
| 96 |
+
|
| 97 |
+
To capture intra-spectrum and inter-spectrum dependencies, we concatenate the embeddings obtained from patch encoding and position encoding of different spectra: pˆ = p′1∥ · · · ∥p′|S| ∈ R(P|S|i=1 Ni)×d, and then input them into the Transformer encoder as depicted in Figure 2. Then each head h = 1, . . . , H in multi-head attention will transform them into query matrices Qh = pWˆ Qh , key matrices Kh = pWˆ Kh and value matrices Vh = pWˆ Vh , where W Qh , W Kh ∈ Rd×dk and WVh ∈ Rd× dH . Afterward, a scaled product is utilized to obtain the attention output Oh ∈ R(P|S|i=1 Ni)× dH :
|
| 98 |
+
|
| 99 |
+

|
| 100 |
+
|
| 101 |
+
The multi-head attention block also includes BatchNorm layers and a feed forward network with residual connections as shown in Figure 2. After combining the outputs of all heads, it generates the representation denoted as z ∈ R(P|S|i=1 Ni)×d. Finally, a flatten layer with representation projection head is used to obtain the molecular spectra representation zs ∈ Rd.
|
| 102 |
+
|
| 103 |
+
# 3.2 MASKED PATCHES RECONSTRUCTION PRE-TRAINING FOR SPECTRA
|
| 104 |
+
|
| 105 |
+
Before distilling the spectra information into 3D molecular representation learning, we need first ensure that the spectrum encoder can effectively comprehend molecular spectra and generate spectral representations. Considering the success of masking modeling across various domains (Devlin et al., 2019; He et al., 2022; Hou et al., 2022; Xia et al., 2023; Wang et al., 2024b; Nie et al., 2023), we propose a masked patches reconstruction (MPR) objective to guide the training of SpecFormer.
|
| 106 |
+
|
| 107 |
+
After the patching step, we randomly select a portion of patches according to the mask ratio α and replace them with zero vectors to implement the masking. Subsequently, the masked patches undergo patch encoding and position encoding. In this way, the semantics of the masked patches (the absorption intensity at specific wavelengths) are obscured during patch encoding, while the positional information is retained to facilitate the reconstruction of the original semantics.
|
| 108 |
+
|
| 109 |
+
After encoding by SpecFormer, the encoded results corresponding to the masked patches are input into a spectrum-specific reconstruction head to reconstruct the original spectral values that were masked. The mean squared error (MSE) between the reconstruction results and the original masked spectra serves as the loss function for the MPR task, guiding the training of SpecFormer:
|
| 110 |
+
|
| 111 |
+

|
| 112 |
+
|
| 113 |
+
where Pi denotes the set of masked patches in the i-th type of molecular spectra, and pˆi,j denotes ethe reconstructed patch corresponding to the masked patch pi,j .
|
| 114 |
+
|
| 115 |
+
# 3.3 CONTRASTIVE LEARNING BETWEEN 3D STRUCTURES AND SPECTRA
|
| 116 |
+
|
| 117 |
+
Under the guidance of the denoising objective for 3D representation learning and the MPR objective for spectral representation learning, we further introduce a contrastive objective to align the representations across these two modalities. We treat the 3D representation zx ∈ Rd and spectral representation zs ∈ Rd of the same molecule as positive samples, and negative samples otherwise. Subsequently, the consistency between positive samples and the discrepancy between negative samples are maximized through the contrastive objective. Given the theoretical and empirical effectiveness, we employ InfoNCE (van den Oord et al., 2018) as the contrastive objective:
|
| 118 |
+
|
| 119 |
+

|
| 120 |
+
|
| 121 |
+
where zjx, zjs are randomly sampled 3D and spectra views regarding to the positive pair (zx, zs). fx(zx, zs) and fs(zs, zx) are scoring functions for the two corresponding views, with flexible formulations. Here we adopt fx(zx, zs) = fs(zs, zx) = ⟨zx, zs⟩.
|
| 122 |
+
|
| 123 |
+
Note that the denoising objective can utilize any form from existing 3D molecular representation pre-training studies, enabling seamless integration of our method into these frameworks.
|
| 124 |
+
|
| 125 |
+
# 3.4 TWO-STAGE PRE-TRAINING PIPELINE
|
| 126 |
+
|
| 127 |
+
Previous pre-training efforts for 3D molecular representation have been conducted on unlabeled datasets using denoising objective. These datasets typically provide only equilibrium 3D structures without offering spectra for all molecules. To enhance the pre-training effect by incorporating spectra while leveraging denoising pre-training, we employ a two-stage pre-training approach. The first stage involves training on a larger dataset (Nakata & Shimazaki, 2017) without spectra using only the denoising objective. Subsequently, the second stage involves training on a dataset that includes spectra using the complete objective as follows:
|
| 128 |
+
|
| 129 |
+

|
| 130 |
+
|
| 131 |
+
where βDenoising, βMPR, and βContrast denote the weights of each sub-objective.
|
| 132 |
+
|
| 133 |
+
# 4 EXPERIMENTS
|
| 134 |
+
|
| 135 |
+
To comprehensively evaluate the impact of molecular spectra on molecular tasks, we first verify the effectiveness of molecular spectra in the training-from-scratch method for the downstream task. Furthermore, we evaluate the effectiveness of our pre-training framework MolSpectra.
|
| 136 |
+
|
| 137 |
+
# 4.1 EFFECTIVENESS OF MOLECULAR SPECTRA IN TRAINING FROM SCRATCH
|
| 138 |
+
|
| 139 |
+
This pilot experiment aims to demonstrate the rationality for incorporating molecular spectra into pre-training. We introduce additional spectral features into a train-from-scratch molecular property prediction model to observe the impact of spectral information on prediction outcomes. We employ EGNN (Satorras et al., 2021), a representative 3D molecular encoder, equipped with an MLP-based prediction head as the baseline model. While EGNN encodes the 3D representations, the UV-Vis spectrum of each molecule provided by the QM9S (Zou et al., 2023) dataset is encoded into spectral representations by a spectrum encoder. Before making predictions with the final MLP, we concatenate the spectral and 3D representations for prediction. The results are presented in Table 1.
|
| 140 |
+
|
| 141 |
+
Table 1: Performance (MAE ↓) when training from scratch on QM9 dataset.
|
| 142 |
+

|
| 143 |
+
|
| 144 |
+
We observe that by directly concatenating spectral representations, the performance of molecular property prediction can be effectively enhanced. This indicates that the information from molecular spectra is beneficial for downstream molecular property prediction. Further incorporating molecular spectra into the pre-training phase of molecular representation has the potential to enhance the informativeness and generalization capability of the representations, thereby broadly improving the performance of downstream tasks.
|
| 145 |
+
|
| 146 |
+
# 4.2 EFFECTIVENESS OF MOLECULAR SPECTRA IN REPRESENTATION PRE-TRAINING
|
| 147 |
+
|
| 148 |
+
We conduct experiments to evaluate MolSpectra by first introducing spectral data into the pretraining of 3D representations, followed by evaluating the performance on downstream tasks. For a comprehensive comparison, two types of baselines are adopted: (1) training-from-scratch methods, including SchNet (Schutt et al. ¨ , 2017), EGNN, DimeNet (Klicpera et al., 2020b), DimeNet++ (Klicpera et al., 2020a), PaiNN (Schutt et al. ¨ , 2021), SphereNet (Liu et al., 2021), and TorchMD-Net (Tholke & Fabritiis ¨ , 2022); and (2) pre-training methods, including Transformer-M (Luo et al., 2023), SE(3)-DDM (Liu et al., 2023b), 3D-EMGP (Jiao et al., 2023), and Coord.
|
| 149 |
+
|
| 150 |
+
MolSpectra can be seamlessly plugged into any existing denoising method. To evaluate the enhancement provided by our method compared to denoising alone, we select the representative coordinate denoising (Coord) as our denoising sub-objective. This method also serves as our primary baseline.
|
| 151 |
+
|
| 152 |
+
# 4.2.1 PRE-TRAINING DATASET.
|
| 153 |
+
|
| 154 |
+
As described in Section 3.4, we first perform denoising pre-training on the PCQM4Mv2 (Nakata & Shimazaki, 2017) dataset, followed by a second stage of pre-training on the QM9Spectra (QM9S) (Zou et al., 2023) dataset, which includes multi-modal molecular energy spectra. In both stages, we adopt the denoising objective provided by Coord (Zaidi et al., 2023), as defined in Eq. 2.
|
| 155 |
+
|
| 156 |
+
The QM9S dataset comprises organic molecules from the QM9 (Ramakrishnan et al., 2014) dataset. The UV-Vis, IR, and Raman spectra of the molecules are calculated at the B3LYP/def-TZVP level of theory, through frequency analysis and time-dependent density functional theory (TD-DFT).
|
| 157 |
+
|
| 158 |
+
# 4.2.2 QM9
|
| 159 |
+
|
| 160 |
+
The QM9 dataset is a quantum chemistry dataset comprising over 134,000 small molecules, each consisting of up to 9 heavy atoms (C, N, O, F) and additional H atoms. This dataset provides an equilibrium geometric conformation for each molecule along with 12 property labels. The dataset is divided into a training set of 110k molecules, a validation set of 10k molecules, and a test set containing the remaining over 10k molecules. Prediction errors are measured using the mean absolute error (MAE). The experimental results are presented in Table 2.
|
| 161 |
+
|
| 162 |
+
The 3D molecular representations pre-trained using our method are fine-tuned and used for prediction across various properties, achieving state-of-the-art performance in 8 out of 12 properties and outperforms Coord in 10 out of 12 properties. In conjunction with the observations in Section 4.1, the performance improvement can be attributed to our incorporation of an understanding of molecular spectra and the knowledge they entail into the 3D molecular representations.
|
| 163 |
+
|
| 164 |
+
Table 2: Performance (MAE↓) on QM9 dataset. The compared methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are highlighted in bold.
|
| 165 |
+

|
| 166 |
+
|
| 167 |
+
Table 3: Performance (MAE↓) on MD17 force prediction (kcal/mol/ ˚A). The methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are in bold.
|
| 168 |
+

|
| 169 |
+
|
| 170 |
+
# 4.2.3 MD17
|
| 171 |
+
|
| 172 |
+
The MD17 dataset contains molecular dynamics trajectories for eight organic molecules, including aspirin, benzene, and ethanol. It offers 150k to nearly 1M conformations per molecule, with energy and force labels. Unlike QM9, MD17 emphasizes dynamic behavior in addition to static properties. We use a standard limited data split: models train on 1k samples, validate on 50, and test on the rest. Performance is evaluated using MAE, with results in Table 3.
|
| 173 |
+
|
| 174 |
+
Our approach also results in the expected performance improvement on MD17. MD17 is a dataset comprising a large number of non-equilibrium molecular structures and their corresponding force fields, which serves to evaluate a model’s understanding of molecular dynamics. However, previous pre-training methods based solely on denoising have only learned force field patterns at static equilibrium states, failing to adequately capture the dynamic evolution of molecular systems. In contrast, our MolSpectra learns the dynamic evolution of molecules by understanding energy level transition patterns, thereby outperforming denoising-based pre-training methods.
|
| 175 |
+
|
| 176 |
+
4.3 SENSITIVITY ANALYSIS OF PATCH LENGTH Pi, STRIDE Di, AND MASK RATIO α
|
| 177 |
+
|
| 178 |
+
We conduct experiments to evaluate the impact of patch length Pi, stride Di, and mask ratio α.
|
| 179 |
+
Results are summarized in Table 4 and Table 5.
|
| 180 |
+
|
| 181 |
+
From Table 4, we observe that when consecutive patches have overlap (Di < Pi), the performance of pre-training is superior compared to scenarios without overlap (Di = Pi). Specifically, the performance is optimal when the stride is half of the patch length. This is because appropriate overlap can better preserve and capture local features, particularly the information at the patch boundaries. Additionally, we find that choosing an appropriate patch length further enhances performance. In our experiments, the configuration of Pi = 20, Di = 10 yields the best results.
|
| 182 |
+
|
| 183 |
+
Table 4: Sensitivity of patch length and stride.
|
| 184 |
+

|
| 185 |
+
|
| 186 |
+
Table 5: Sensitivity of mask ratio.
|
| 187 |
+

|
| 188 |
+
|
| 189 |
+
Regarding the mask ratio, α = 0.10 is a preferable choice. A small mask ratio results in insufficient MPR optimization, hindering SpecFormer training. Conversely, a large mask ratio causes excessive spectral perturbation, degrading performance when aligning with the 3D representations with the contrastive objective. An appropriate mask ratio strikes a balance between these two aspects.
|
| 190 |
+
|
| 191 |
+
# 4.4 ABLATION STUDY
|
| 192 |
+
|
| 193 |
+
To rigorously demonstrate the contributions of masked patches reconstruction, the incorporation of molecular spectra, and each spectral modality, we conducted an ablation study on them.
|
| 194 |
+
|
| 195 |
+
Ablation study of masked patches reconstruction. We remove the MPR loss to analyze the impact of masked patches reconstruction, referred to as “w/o MPR” in Table 6. Removing the MPR objective leads to performance deterioration. This is consistent with the sensitivity analysis of the mask ratio α in Section 4.3, as removing MPR is an extreme case where α = 0. This decline is due to the lack of effective guidance in training SpecFormer. Using an undertrained SpecFormer for contrastive learning with 3D encoder outputs limits performance improvement.
|
| 196 |
+
|
| 197 |
+
Table 6: Ablation of optimization objectives.
|
| 198 |
+

|
| 199 |
+
|
| 200 |
+
Ablation study of molecular spectra. We retain only the denoising loss, removing both the MPR loss and contrastive loss, referred to as “w/o MPR, Contrast” in Table 6. The only difference between this variant and MolSpectra is the incorporation of molecular spectra into the pre-training. The ”w/o MPR, Contrast” results are inferior to those of MolSpectra, highlighting that incorporating molecular spectra effectively enhances the quality and generalizability of molecular 3D representations.
|
| 201 |
+
|
| 202 |
+
Ablation study of each spectral modality. To evaluate the contributions of each spectral modality to the performance, we conduct an ablation study for each modality. The results are presented in Table 7. It can be observed that each spectral modality contributes differently, with the UV-Vis spectrum having the smallest contribution and the IR spectrum the largest, likely due to the varying information content in each modality.
|
| 203 |
+
|
| 204 |
+
Table 7: Ablation of spectral modalities.
|
| 205 |
+

|
| 206 |
+
|
| 207 |
+
# 5 RELATED WORK
|
| 208 |
+
|
| 209 |
+
3D molecular pre-training. Molecular 2D structures are typically represented as graphs and modeled using graph learning methods (Gilmer et al., 2017; Li et al., 2023; Jiang et al., 2024; Yuan et al., 2025). However, 3D molecular structures provide critical geometric information that is essential for understanding physicochemical properties (Chen et al., 2023; 2024; Wang et al., 2024a; Sun et al., 2024; Han et al., 2024), which cannot be directly inferred from 2D graphs or SMILES representations (Gong et al., 2024). Designing effective strategies for pre-training 3D molecular representations remains challenging due to the geometric symmetries inherent in 3D structures and their strong connection to physical knowledge, such as potential energy functions.
|
| 210 |
+
|
| 211 |
+
Denoising the geometric structure has been demonstrated as an effective strategy for 3D representation pre-training (Liu et al., 2023b; Jiao et al., 2023; Kim et al., 2023; Zhou et al., 2023; Wang et al., 2025). Coordinate denoising (Coord) (Zaidi et al., 2023) first theoretically proves that the denoising objective is equivalent to learning the gradient of the potential energy with respect to positions, essentially the force field. Building on this work, fractional denoising (Frad) (Feng et al., 2023) introduces dihedral angle noise to optimize the sampling of low-energy structures. Further, SliDe (Ni et al., 2024) incorporates a more rigorous potential energy from classical mechanics. Another line of research simultaneously leverages both 2D and 3D structures for pre-training molecular representations, addressing the complementarity of the two modalities (Li et al., 2022; Zhu et al., 2022; Liu et al., 2023a; Du et al., 2023a; Yu et al., 2024) or the computational complexity of 3D structure determination (Liu et al., 2022; Stark et al.¨ , 2022; Wang et al., 2023a).
|
| 212 |
+
|
| 213 |
+
Although these studies elucidate the relationship between molecular 3D structures and their energy states, they remain limited to the description of molecular energy states within classical mechanics, without considering the quantized energy level structures as described by quantum mechanics.
|
| 214 |
+
|
| 215 |
+
Molecular spectroscopy. Molecular spectroscopy studies interactions between molecules and electromagnetic radiation. Analyzing spectra provides valuable insights into molecular structure, composition, and dynamics (Lancaster et al., 2024). When encountering unknown substances, researchers conduct spectroscopic measurements on samples and compare the observed spectra with libraries for identification. To expand library coverage, machine learning methods are widely used to predict molecules’ spectra (Zou et al., 2023; Wei et al., 2018; Zong et al., 2024).
|
| 216 |
+
|
| 217 |
+
Some studies incorporate physical principles into spectra prediction models as inductive biases, including molecular dynamics simulations via equivariant message passing (Schutt et al. ¨ , 2021), fragmentation (Duhrkop et al. ¨ , 2020; Cao et al., 2020; Goldman et al., 2023a), motifs (Park et al., 2023), and long-distance atomic interactions (Young et al., 2024). Another line of research approach bypasses spectral library comparison and directly performs de novo structure elucidation from spectra (Stravs et al., 2021; Goldman et al., 2023b; Tao et al., 2024).
|
| 218 |
+
|
| 219 |
+
Since different spectroscopic techniques offer complementary advantages, the joint analysis of multiple spectra can provide comprehensive information (Alberts et al., 2024). In this study, we encodes multiple spectra, and introduce them into molecular representation pre-training for the first time.
|
| 220 |
+
|
| 221 |
+
# 6 CONCLUSION
|
| 222 |
+
|
| 223 |
+
In this study, we explore pre-training molecular 3D representations beyond classical mechanics. By leveraging the correlation between molecular energy level structures and molecular spectra in quantum mechanics, we introduce molecular spectra for pre-training molecular 3D representations (MolSpectra). By aligning the 3D encoder trained with a denoising objective and the spectrum encoder trained with a masked patch reconstruction objective, we enhance the informativeness and transferability of the resulting 3D representations.
|
| 224 |
+
|
| 225 |
+
# ACKNOWLEDGMENTS
|
| 226 |
+
|
| 227 |
+
This work is jointly supported by National Science and Technology Major Project (2023ZD0120901) and National Natural Science Foundation of China (62372454, 62236010).
|
| 228 |
+
|
| 229 |
+
# REFERENCES
|
| 230 |
+
|
| 231 |
+
Saman Alavi. Intra- and intermolecular potentials in simulations. In Chapter 3, pp. 39–71. John Wiley & Sons, Ltd, 2020. ISBN 9783527699452. doi: 10.1002/9783527699452.ch3.
|
| 232 |
+
|
| 233 |
+
Marvin Alberts, Oliver Schilter, Federico Zipoli, Nina Hartrampf, and Teodoro Laino. Unraveling molecular structure: A multimodal spectroscopic dataset for chemistry. In NeurIPS Datasets and Benchmarks Track, 2024.
|
| 234 |
+
|
| 235 |
+
Ilyes Batatia, D’avid P’eter Kov’acs, Gregor N. C. Simm, Christoph Ortner, and Gabor Cs ´ anyi. ´ Mace: Higher order equivariant message passing neural networks for fast and accurate force fields. In NeurIPS, 2022.
|
| 236 |
+
|
| 237 |
+
Liu Cao, Mustafa Guler, Azat M. Tagirdzhanov, Yi-Yuan Lee, Alexey A. Gurevich, and Hosein Mohimani. Moldiscovery: learning mass spectrometry fragmentation of small molecules. Nature Communications, 12, 2020.
|
| 238 |
+
|
| 239 |
+
Dingshuo Chen, Yanqiao Zhu, Jieyu Zhang, Yuanqi Du, Zhixun Li, Qiang Liu, Shu Wu, and Liang Wang. Uncovering neural scaling laws in molecular representation learning. In NeurIPS, 2023.
|
| 240 |
+
|
| 241 |
+
Dingshuo Chen, Zhixun Li, Yuyan Ni, Guibin Zhang, Ding Wang, Qiang Liu, Shu Wu, Jeffrey Xu Yu, and Liang Wang. Beyond efficiency: Molecular data pruning for enhanced generalization. In NeurIPS, 2024.
|
| 242 |
+
|
| 243 |
+
Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey E. Hinton. A simple framework for contrastive learning of visual representations. In ICML, volume 119, pp. 1597–1607, 2020.
|
| 244 |
+
|
| 245 |
+
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: pre-training of deep bidirectional transformers for language understanding. In NAACL-HLT (1), pp. 4171–4186. Association for Computational Linguistics, 2019.
|
| 246 |
+
|
| 247 |
+
Weitao Du, Jiujiu Chen, Xuecang Zhang, Zhi-Ming Ma, and Shengchao Liu. Molecule joint autoencoding: Trajectory pretraining with 2d and 3d diffusion. In NeurIPS, 2023a.
|
| 248 |
+
|
| 249 |
+
Weitao Du, Yuanqi Du, Limei Wang, Dieqiao Feng, Guifeng Wang, Shuiwang Ji, Carla P. Gomes, and Zhi-Ming Ma. A new perspective on building efficient and expressive 3d equivariant graph neural networks. In NeurIPS, 2023b.
|
| 250 |
+
|
| 251 |
+
Kai Duhrkop, Louis-F ¨ elix Nothias, Markus Fleischauer, Raphael Reher, Marcus Ludwig, Martin A.´ Hoffmann, Daniel Petras, William H. Gerwick, Juho Rousu, Pieter C. Dorrestein, and Sebastian´ Bocker. Systematic classification of unknown metabolites using high-resolution fragmentation ¨ mass spectra. Nature Biotechnology, 39:462 – 471, 2020.
|
| 252 |
+
|
| 253 |
+
Shikun Feng, Yuyan Ni, Yanyan Lan, Zhi-Ming Ma, and Wei-Ying Ma. Fractional denoising for 3d molecular pre-training. In ICML, volume 202. PMLR, 2023.
|
| 254 |
+
|
| 255 |
+
Justin Gilmer, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals, and George E. Dahl. Neural message passing for quantum chemistry. In ICML, 2017.
|
| 256 |
+
|
| 257 |
+
Samuel Goldman, John Bradshaw, Jiayi Xin, and Connor W. Coley. Prefix-tree decoding for predicting mass spectra from molecules. In NeurIPS, 2023a.
|
| 258 |
+
|
| 259 |
+
Samuel Goldman, Jeremy Wohlwend, Martin, Strazar, Guy Haroush, Ramnik J. Xavier, W. Connor, and Coley. Annotating metabolite mass spectra with domain-inspired chemical formula transformers. Nature Machine Intelligence, 2023b.
|
| 260 |
+
|
| 261 |
+
Haisong Gong, Qiang Liu, Shu Wu, and Liang Wang. Text-guided molecule generation with diffusion language model. In AAAI, 2024.
|
| 262 |
+
|
| 263 |
+
Jiaqi Han, Jiacheng Cen, Liming Wu, Zongzhao Li, Xiangzhe Kong, Rui Jiao, Ziyang Yu, Tingyang Xu, Fandi Wu, Zihe Wang, Hongteng Xu, Zhewei Wei, Yang Liu, Yu Rong, and Wenbing Huang. A survey of geometric graph neural networks: Data structures, models and applications. arXiv, abs/2403.00485, 2024.
|
| 264 |
+
|
| 265 |
+
Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollar, and Ross B. Girshick. Masked ´ autoencoders are scalable vision learners. In CVPR, pp. 15979–15988. IEEE, 2022.
|
| 266 |
+
|
| 267 |
+
Zhenyu Hou, Xiao Liu, Yukuo Cen, Yuxiao Dong, Hongxia Yang, Chunjie Wang, and Jie Tang. Graphmae: Self-supervised masked graph autoencoders. In KDD, pp. 594–604. ACM, 2022.
|
| 268 |
+
|
| 269 |
+
Weihua Hu, Bowen Liu, Joseph Gomes, Marinka Zitnik, Percy Liang, Vijay S. Pande, and Jure Leskovec. Strategies for pre-training graph neural networks. In ICLR, 2020.
|
| 270 |
+
|
| 271 |
+
Xinke Jiang, Rihong Qiu, Yongxin Xu, Wentao Zhang, Yichen Zhu, Ruizhe Zhang, Yuchen Fang, Xu Chu, Junfeng Zhao, and Yasha Wang. Ragraph: A general retrieval-augmented graph learning framework. In NeurIPS, 2024.
|
| 272 |
+
|
| 273 |
+
Rui Jiao, Jiaqi Han, Wenbing Huang, Yu Rong, and Yang Liu. Energy-motivated equivariant pretraining for 3d molecular graphs. In AAAI, pp. 8096–8104. AAAI Press, 2023.
|
| 274 |
+
|
| 275 |
+
Dongki Kim, Jinheon Baek, and Sung Ju Hwang. Graph self-supervised learning with accurate discrepancy learning. In NeurIPS, 2022.
|
| 276 |
+
|
| 277 |
+
Hyeonsu Kim, Jeheon Woo, Seonghwan Kim, Seokhyun Moon, Jun Hyeong Kim, and Woo Youn Kim. Geotmi: Predicting quantum chemical property with easy-to-obtain geometry via positional denoising. In NeurIPS, 2023.
|
| 278 |
+
|
| 279 |
+
Johannes Klicpera, Shankari Giri, Johannes T. Margraf, and Stephan Gunnemann. Fast and uncertainty-aware directional message passing for non-equilibrium molecules. arXiv, abs/2011.14115, 2020a.
|
| 280 |
+
|
| 281 |
+
Johannes Klicpera, Janek Groß, and Stephan Gunnemann. Directional message passing for molec- ¨ ular graphs. In ICLR, 2020b.
|
| 282 |
+
|
| 283 |
+
Fan-Fang Kong, Xiao-Jun Tian, Yang Zhang, Yun-Jie Yu, Shi-Hao Jing, Yao Zhang, Guangjun Tian, Yi Luo, Jinlong Yang, Zhenchao Dong, and J. G. Hou. Probing intramolecular vibronic coupling through vibronic-state imaging. Nature Communications, 12, 2021.
|
| 284 |
+
|
| 285 |
+
Noah M. Lancaster, Pavel Sinitcyn, Patrick Forny, Trenton M. Peters-Clarke, Caroline Fecher, Andrew J. Smith, Evgenia Shishkova, Tabiwang N. Arrey, Anna Pashkova, Margaret Lea Robinson, Nicholas L. Arp, Jing Fan, Julia K. Hansen, Andrea Galmozzi, Lia R. Serrano, Julie Rojas, Audrey P. Gasch, Michael S. Westphall, Hamish I Stewart, Christian Hock, Eugen Damoc, David J. Pagliarini, Vlad Zabrouskov, and Joshua J. Coon. Fast and deep phosphoproteome analysis with the orbitrap astral mass spectrometer. Nature Communications, 15, 2024.
|
| 286 |
+
|
| 287 |
+
Shuangli Li, Jingbo Zhou, Tong Xu, Dejing Dou, and Hui Xiong. Geomgcl: Geometric graph contrastive learning for molecular property prediction. In AAAI, pp. 4541–4549. AAAI Press, 2022.
|
| 288 |
+
|
| 289 |
+
Zhixun Li, Liang Wang, Xin Sun, Yifan Luo, Yanqiao Zhu, Dingshuo Chen, Yingtao Luo, Xiangxin Zhou, Qiang Liu, Shu Wu, Liang Wang, and Jeffrey Xu Yu. GSLB: the graph structure learning benchmark. In NeurIPS, 2023.
|
| 290 |
+
|
| 291 |
+
Yi-Lun Liao and Tess E. Smidt. Equiformer: Equivariant graph attention transformer for 3d atomistic graphs. In ICLR, 2023.
|
| 292 |
+
|
| 293 |
+
Shengchao Liu, Hanchen Wang, Weiyang Liu, Joan Lasenby, Hongyu Guo, and Jian Tang. Pretraining molecular graph representation with 3d geometry. In ICLR, 2022.
|
| 294 |
+
|
| 295 |
+
Shengchao Liu, Weitao Du, Zhi-Ming Ma, Hongyu Guo, and Jian Tang. A group symmetric stochastic differential equation model for molecule multi-modal pretraining. In ICML, volume 202, pp. 21497–21526. PMLR, 2023a.
|
| 296 |
+
|
| 297 |
+
Shengchao Liu, Hongyu Guo, and Jian Tang. Molecular geometry pretraining with se(3)-invariant denoising distance matching. In ICLR, 2023b.
|
| 298 |
+
|
| 299 |
+
Yi Liu, Limei Wang, Meng Liu, Xuan Zhang, Bora Oztekin, and Shuiwang Ji. Spherical message passing for 3d molecular graphs. In International Conference on Learning Representations, 2021.
|
| 300 |
+
|
| 301 |
+
Zhiyuan Liu, Yaorui Shi, An Zhang, Enzhi Zhang, Kenji Kawaguchi, Xiang Wang, and Tat-Seng Chua. Rethinking tokenizer and decoder in masked graph modeling for molecules. In NeurIPS, 2023c.
|
| 302 |
+
|
| 303 |
+
Shengjie Luo, Tianlang Chen, Yixian Xu, Shuxin Zheng, Tie-Yan Liu, Liwei Wang, and Di He. One transformer can understand both 2d & 3d molecular data. In ICLR, 2023.
|
| 304 |
+
|
| 305 |
+
Hehuan Ma, Feng Jiang, Yu Rong, Yuzhi Guo, and Junzhou Huang. Toward robust self-training paradigm for molecular prediction tasks. Journal of Computational Biology, 31(3):213–228, 2024. doi: 10.1089/cmb.2023.0187.
|
| 306 |
+
|
| 307 |
+
Albert Musaelian, Simon L. Batzner, Anders Johansson, Lixin Sun, Cameron J. Owen, Mordechai Kornbluth, and Boris Kozinsky. Learning local equivariant representations for large-scale atomistic dynamics. Nature Communications, 14, 2023.
|
| 308 |
+
|
| 309 |
+
Maho Nakata and Tomomi Shimazaki. Pubchemqc project: A large-scale first-principles electronic structure database for data-driven chemistry. Journal of chemical information and modeling, 57 6:1300–1308, 2017.
|
| 310 |
+
|
| 311 |
+
Yuyan Ni, Shikun Feng, Wei-Ying Ma, Zhi-Ming Ma, and Yanyan Lan. Sliced denoising: A physicsinformed molecular pre-training method. In ICLR, 2024.
|
| 312 |
+
|
| 313 |
+
Yuqi Nie, Nam H. Nguyen, Phanwadee Sinthong, and Jayant Kalagnanam. A time series is worth 64 words: Long-term forecasting with transformers. In ICLR, 2023.
|
| 314 |
+
|
| 315 |
+
Jiwon Victoria Park, Jeonghee Jo, and Sungroh Yoon. Mass spectra prediction with structural motifbased graph neural networks. Scientific Reports, 14, 2023.
|
| 316 |
+
|
| 317 |
+
Raghunathan Ramakrishnan, Pavlo O. Dral, Pavlo O. Dral, Matthias Rupp, and O. Anatole von Lilienfeld. Quantum chemistry structures and properties of 134 kilo molecules. Scientific Data, 1, 2014.
|
| 318 |
+
|
| 319 |
+
Yu Rong, Yatao Bian, Tingyang Xu, Weiyang Xie, Ying Wei, Wenbing Huang, and Junzhou Huang. Self-supervised graph transformer on large-scale molecular data. In NeurIPS, 2020.
|
| 320 |
+
|
| 321 |
+
Victor Garcia Satorras, Emiel Hoogeboom, and Max Welling. E(n) equivariant graph neural networks. In ICML, volume 139, pp. 9323–9332. PMLR, 2021.
|
| 322 |
+
|
| 323 |
+
Kristof Schutt, Oliver T. Unke, and Michael Gastegger. Equivariant message passing for the pre- ¨ diction of tensorial properties and molecular spectra. In ICML, volume 139 of Proceedings of Machine Learning Research, pp. 9377–9388. PMLR, 2021.
|
| 324 |
+
|
| 325 |
+
Kristof T. Schutt, Huziel E. Sauceda, P J Kindermans, Alexandre Tkatchenko, and Klaus-Robert ¨ Muller. Schnet - a deep learning architecture for molecules and materials. ¨ The Journal of chemical physics, 148 24:241722, 2017.
|
| 326 |
+
|
| 327 |
+
Andrew Shin, Masato Ishii, and Takuya Narihira. Perspectives and prospects on transformer architecture for cross-modal tasks with language and vision. International Journal of Computer Vision, 130:435 – 454, 2021.
|
| 328 |
+
|
| 329 |
+
Hannes Stark, Dominique Beaini, Gabriele Corso, Prudencio Tossou, Christian Dallago, Stephan ¨ Gunnemann, and Pietro Li ¨ o. 3d infomax improves gnns for molecular property prediction. In ´ ICML, volume 162, pp. 20479–20502. PMLR, 2022.
|
| 330 |
+
|
| 331 |
+
Michael A. Stravs, Kai Duhrkop, Sebastian B ¨ ocker, and Nicola Zamboni. Msnovelist: de novo ¨ structure generation from mass spectra. Nature Methods, 19:865 – 870, 2021.
|
| 332 |
+
|
| 333 |
+
Xin Sun, Liang Wang, Qiang Liu, Shu Wu, Zilei Wang, and Liang Wang. DIVE: subgraph disagreement for graph out-of-distribution generalization. In KDD, 2024.
|
| 334 |
+
|
| 335 |
+
Shijie Tao, Yi Feng, Wenmin Wang, Tiantian Han, Pieter E S Smith, and Jun Jiang. A machine learning protocol for geometric information retrieval from molecular spectra. Artificial Intelligence Chemistry, 2024.
|
| 336 |
+
|
| 337 |
+
Philipp Tholke and Gianni De Fabritiis. Equivariant transformers for neural network based molecu- ¨ lar potentials. In ICLR, 2022.
|
| 338 |
+
|
| 339 |
+
Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predic-¨ tive coding. arXiv, abs/1807.03748, 2018.
|
| 340 |
+
|
| 341 |
+
Pascal Vincent. A connection between score matching and denoising autoencoders. Neural Computation, 23:1661–1674, 2011.
|
| 342 |
+
|
| 343 |
+
Liang Wang, Qiang Liu, Shaozhen Liu, Xin Sun, Shu Wu, and Liang Wang. Pin-tuning: Parameterefficient in-context tuning for few-shot molecular property prediction. In NeurIPS, 2024a.
|
| 344 |
+
|
| 345 |
+
Liang Wang, Xiang Tao, Qiang Liu, Shu Wu, and Liang Wang. Rethinking graph masked autoencoders through alignment and uniformity. In AAAI, 2024b.
|
| 346 |
+
|
| 347 |
+
Liang Wang, Chao Song, Zhiyuan Liu, Yu Rong, Q. Liu, Shu Wu, and Liang Wang. Diffusion models for molecules: A survey of methods and tasks. arXiv, abs/2502.09511, 2025.
|
| 348 |
+
|
| 349 |
+
Xu Wang, Huan Zhao, Wei-Wei Tu, and Quanming Yao. Automated 3d pre-training for molecular property prediction. In KDD, pp. 2419–2430. ACM, 2023a.
|
| 350 |
+
|
| 351 |
+
Yiqun Wang, Yuning Shen, Shi Chen, Lihao Wang, Fei Ye, and Hao Zhou. Learning harmonic molecular representations on riemannian manifold. In ICLR, 2023b.
|
| 352 |
+
|
| 353 |
+
Yuyang Wang, Jianren Wang, Zhonglin Cao, and Amir Barati Farimani. Molecular contrastive learning of representations via graph neural networks. Nature Machine Intelligence, 4(3):279– 287, 2022.
|
| 354 |
+
|
| 355 |
+
Jennifer N. Wei, Jennifer N. Wei, David Belanger, Ryan P. Adams, and D. Sculley. Rapid prediction of electron–ionization mass spectrometry using neural networks. ACS Central Science, 5:700 – 708, 2018.
|
| 356 |
+
|
| 357 |
+
Jun Xia, Chengshuai Zhao, Bozhen Hu, Zhangyang Gao, Cheng Tan, Yue Liu, Siyuan Li, and Stan Z. Li. Mole-bert: Rethinking pre-training graph neural networks for molecules. In ICLR, 2023.
|
| 358 |
+
|
| 359 |
+
Adamo Young, Bo Wang, and Hannes Rost. Tandem mass spectrum prediction for small molecules using graph transformers. Nature Machine Intelligence, 2024.
|
| 360 |
+
|
| 361 |
+
Qiying Yu, Yudi Zhang, Yuyan Ni, Shikun Feng, Yanyan Lan, Hao Zhou, and Jingjing Liu. Multimodal molecular pretraining via modality blending. In ICLR, 2024.
|
| 362 |
+
|
| 363 |
+
Chaohao Yuan, Kangfei Zhao, Ercan Engin Kuruoglu, Liang Wang, Tingyang Xu, Wenbing Huang, Deli Zhao, Hong Cheng, and Yu Rong. A survey of graph transformers: Architectures, theories and applications. arXiv, abs/2502.16533, 2025.
|
| 364 |
+
|
| 365 |
+
Sheheryar Zaidi, Michael Schaarschmidt, James Martens, Hyunjik Kim, Yee Whye Teh, Alvaro Sanchez-Gonzalez, Peter W. Battaglia, Razvan Pascanu, and Jonathan Godwin. Pre-training via denoising for molecular property prediction. In ICLR, 2023.
|
| 366 |
+
|
| 367 |
+
Gengmo Zhou, Zhifeng Gao, Qiankun Ding, Hang Zheng, Hongteng Xu, Zhewei Wei, Linfeng Zhang, and Guolin Ke. Uni-mol: A universal 3d molecular representation learning framework. In ICLR, 2023.
|
| 368 |
+
|
| 369 |
+
Kun Zhou and Bo Liu. Chapter 2 - potential energy functions. In Molecular Dynamics Simulation, pp. 41–65. Elsevier, 2022. ISBN 978-0-12-816419-8. doi: 10.1016/B978-0-12-816419-8. 00007-6.
|
| 370 |
+
|
| 371 |
+
Jinhua Zhu, Yingce Xia, Lijun Wu, Shufang Xie, Tao Qin, Wengang Zhou, Houqiang Li, and Tie-Yan Liu. Unified 2d and 3d pre-training of molecular representations. In KDD, pp. 2626–2636. ACM, 2022.
|
| 372 |
+
|
| 373 |
+
Yu Zong, Yuxin Wang, Xipeng Qiu, Xuanjing Huang, and Liang Qiao. Deep learning prediction of glycopeptide tandem mass spectra powers glycoproteomics. Nature Machine Intelligence, 2024.
|
| 374 |
+
|
| 375 |
+
Zihan Zou, Yujin Zhang, Lijun Liang, Mingzhi Wei, Jiancai Leng, Jun Jiang, Yi Luo, and Wei Hu. A deep learning model for predicting selected organic molecular spectra. Nature Computional Science, 3(11):957–964, 2023.
|
| 376 |
+
|
| 377 |
+
# Appendix
|
| 378 |
+
|
| 379 |
+
Contents of the appendix
|
| 380 |
+
|
| 381 |
+
A Proof of theoretical results 16
|
| 382 |
+
B Visualization and analysis of spectra 16
|
| 383 |
+
C Implementation details 17
|
| 384 |
+
C.1 Hardware and software 17
|
| 385 |
+
C.2 Model configuration . 17
|
| 386 |
+
D More experimental results and discussions 17
|
| 387 |
+
E Visualization of attention patterns and learned spectra representations in SpecFormer 18
|
| 388 |
+
F Limitations and potential future directions 19
|
| 389 |
+
|
| 390 |
+
A PROOF OF THEORETICAL RESULTS
|
| 391 |
+
|
| 392 |
+
Theorem A.1 (Equivalence between the denoising objective and the learning of molecular force fields (Zaidi et al., 2023)). Assume the conformation distribution is a mixture of Gaussian distribution centered at the equilibriums:
|
| 393 |
+
|
| 394 |
+

|
| 395 |
+
|
| 396 |
+
x0, x ∈ R3N are equilibrium and noisy conformation respectively, N is the number of atoms in the molecule. It relates to molecular energy by Boltzmann distribution p(x) ∝ exp(−E(x)).
|
| 397 |
+
|
| 398 |
+
Then given a sampled molecule M, the denoising loss on the conformation coordinates is an equivalent optimization target to force field prediction:
|
| 399 |
+
|
| 400 |
+

|
| 401 |
+
|
| 402 |
+
where GNNθ(x) denotes a graph neural network with parameters θ which takes conformation x as an input and returns node-level noise predictions, ≃ denotes equivalence.
|
| 403 |
+
|
| 404 |
+
Proof. According to Boltzmann distribution, Eq. A3 is equal to Ep(x)||GN Nθ(x)−∇x log p(x)||2. By using a conditional score matching lemma (Vincent, 2011), the equation above is further equal to Ep(x|x0)p(x0)||GN Nθ(x) − ∇x log p(x|x0)||2 + T1, where T1 is constant independent of θ. Then with the Gaussian assumption, it becomes Ep(x|x0)p(x0)||GN Nθ(x) − x0−xτ2c ||2 + T1. Finally,since coefficients − 1 do not rely on the input x, it can be absorbed into GNN , thus obtaining Eq. A2.
|
| 405 |
+
|
| 406 |
+
# B VISUALIZATION AND ANALYSIS OF SPECTRA
|
| 407 |
+
|
| 408 |
+
In this section, we visualize the three types of spectra we utilize (UV-Vis, IR, Raman) and standardize the initial spectral data based on data analysis. In Figure A1, we visualize 20 randomly sampled spectra from QM9S for each type of spectrum. A notable pattern observed is that, although each spectrum consists of numerous absorption peaks, there are significant differences in the heights (absorption intensities) of these peaks. For instance, in the IR spectra, the absorption intensity at most peaks is around 200, but a few peaks reach an intensity of 800. However, in qualitative analysis, the position and shape of the peaks are more critical than their heights. Therefore, the differences in peak absorption intensities can interfere with model training under the MSE loss metric. To address this issue, we pre-process the absorption intensities of the spectra by applying a log10 transformation to mitigate the interference caused by peak intensity differences.
|
| 409 |
+
|
| 410 |
+

|
| 411 |
+
Figure A1: Randomly sampled examples of molecular energy spectra.
|
| 412 |
+
|
| 413 |
+
# C IMPLEMENTATION DETAILS
|
| 414 |
+
|
| 415 |
+
# C.1 HARDWARE AND SOFTWARE
|
| 416 |
+
|
| 417 |
+
Our experiments are conducted on Linux servers equipped with 184 Intel Xeon Platinum 8469C CPUs, 920GB RAM, and 8 NVIDIA H20 96GB GPUs. Our model is implemented in PyTorch version 2.3.1, PyTorch Geometric version 2.6.1 (https://pyg.org/) with CUDA version 12.1, and Python 3.10.14.
|
| 418 |
+
|
| 419 |
+
# C.2 MODEL CONFIGURATION
|
| 420 |
+
|
| 421 |
+
The SpecFormer is implemented using a 3-layer Transformer with 16 attention heads. Following previous works, we set both d and dk as 256. TorchMD-Net (Tholke & Fabritiis ¨ , 2022) is adopted as the 3D molecular encoder. We tune the mask ratio (i.e., α) in {0.05, 0.10, 0.15, 0.20, 0.25, 0.30}, tune the “stride/patch length” pair (i.e., Di/Pi) in {5/20, 10/20, 15/20, 20/20, 8/16, 15/30}, and tune the weights of sub-objectives (i.e., βDenoising, βMPR, and βContrast ) in {0.01, 0.1, 1, 10}. Since our goal is to align the 3D representations and spectra representations of molecules during the pretraining phase, and not rely on molecular spectra data during downstream fine-tuning, these hyperparameters related to molecular spectra are tuned on the pre-training dataset. Based on the results of hyper-parameter tuning, we adopt α = 0.10, Di = 10, Pi = 20, βDenoising = 1.0, βMPR = 1.0, and βContrast = 1.0.
|
| 422 |
+
|
| 423 |
+
Following SimCLR (Chen et al., 2020), the contrastive loss in our Eq. 7 is implemented using in-batch contrastive loss, where positive and negative pairs are constructed within each data batch. Therefore, for each anchor representation in a batch, there is one positive sample and bs−1 negative samples, where bs is the batch size. In our method, bs = 128.
|
| 424 |
+
|
| 425 |
+
In both pre-training stages, we use the noise generation method and denoising objective provided by Coord (Zaidi et al., 2023), specifically energy function I as described in Section 2.2. The noise is added to atom positions as scaled mixture of isotropic Gaussian noise, with a scaling factor of 0.04. The denoising objective is defined in Eq. 2.
|
| 426 |
+
|
| 427 |
+
For baselines, we follow their recommended settings.
|
| 428 |
+
|
| 429 |
+
# D MORE EXPERIMENTAL RESULTS AND DISCUSSIONS
|
| 430 |
+
|
| 431 |
+
In addition to Coord, we evaluate the effect of incorporating SliDe (Ni et al., 2024) into our Mol-Spectra. SliDe is also a denoising-based pre-training method, utilizing the TorchMD-Net (Tholke ¨ & Fabritiis, 2022) as its encoder backbone, consistent with previous pre-training work (Zaidi et al., 2023; Feng et al., 2023). The results are presented in Table A1 and Table A2.
|
| 432 |
+
|
| 433 |
+
Integrating our method with SliDe effectively reduces the error in property prediction on the QM9 dataset and the MD17 dataset. Given that our method enhances both Coord and SliDe, this suggests that our approach is broadly effective across various denoising-based pretraining strategies. Furthermore, incorporating molecular spectra can guide the pre-trained model to acquire knowledge beyond what denoising objectives can offer, which proves beneficial for downstream property prediction.
|
| 434 |
+
|
| 435 |
+
Table A1: Performance (MAE↓) on QM9 dataset. The better result between the two variants of each pretraining method, w/ and w/o MolSpectra, is highlighted in bold.
|
| 436 |
+

|
| 437 |
+
|
| 438 |
+
Table A2: Performance (MAE↓) on MD17 dataset. The better result between the two variants of each pretraining method, w/ and w/o MolSpectra, is highlighted in bold.
|
| 439 |
+

|
| 440 |
+
|
| 441 |
+
# E VISUALIZATION OF ATTENTION PATTERNS AND LEARNED SPECTRA REPRESENTATIONS IN SPECFORMER
|
| 442 |
+
|
| 443 |
+

|
| 444 |
+
Figure A2: (a-c) Attention maps from three attention heads in SpecFormer. Different heads model distinct dependencies. (d) t-SNE visualization of the spectra representations produced by Spec-Former.
|
| 445 |
+
|
| 446 |
+
We visualize the attention patterns and learned spectra representations in SpecFormer. Based on the visualizations presented in Figure A2, we have made the following observations.
|
| 447 |
+
|
| 448 |
+
In Figure A2(a-c), we visualize attention maps from three attention heads in SpecFormer’s second layer. The attention weights within the three blocks along the main diagonal indicate intra-spectrum dependencies, while those outside reveal inter-spectrum dependencies, as explained in Section 3.1. It can be observed that different attention heads model distinct dependencies: Head 11 focuses on intra-spectrum dependencies, Head 13 focuses on inter-spectrum dependencies, and Head 12 models both types simultaneously. In inter-spectrum dependencies, the interaction between IR spectra and Raman spectra is relatively pronounced, which may be related to their mutual association with vibrational modes. Additionally, because the intensity peaks and dependencies in molecular spectra are sparse, the attention maps in SpecFormer are generally sparse as well.
|
| 449 |
+
|
| 450 |
+
In Figure A2(d), we use t-SNE to visualize the spectra representations produced by the final layer of SpecFormer. It can be observed that the distribution of representations in the latent space is relatively uniform and forms several potential clusters. This well-shaped distribution of representations reveals effective spectra representation learning and supports the structure-spectrum alignment.
|
| 451 |
+
|
| 452 |
+
# F LIMITATIONS AND POTENTIAL FUTURE DIRECTIONS
|
| 453 |
+
|
| 454 |
+
One limitation of our method is the availability, scale, and diversity of molecular spectral data. Our current dataset comprises geometric structures of 134,000 molecules, each with three types of spectra (UV-Vis, IR, Raman). To effectively explore the scaling laws of pre-training methods, larger and more diverse molecular spectral datasets are necessary. Encouragingly, molecular spectroscopy has been gaining increasing attention in the research community, with larger and more diverse datasets being released, such as the recent multimodal spectroscopic dataset (Alberts et al., 2024). This development supports advancements in molecular representation learning and other related tasks.
|
| 455 |
+
|
| 456 |
+
Another limitation is that our proposed SpecFormer can currently only handle one-dimensional molecular spectra. For higher-dimensional spectra, such as two-dimensional NMR and twodimensional correlation spectra, further development of sophisticated spectrum encoder is needed.
|
| 457 |
+
|
| 458 |
+
Looking ahead, we envision several future directions in this field. First, there is potential in investigating the scaling laws of pre-training on larger and more diverse molecular spectral datasets. Second, expanding the scope of molecular spectrum encoding to include a wider range, such as NMR, mass spectra, and two-dimensional spectra, could be highly beneficial. Third, while a pretrained spectral encoder has been developed in our method, we have so far only applied the pretrained 3D encoder to downstream tasks. Exploring the use of the pre-trained spectral encoder for molecular spectrum-related downstream tasks, such as automated molecular structure elucidation from spectra, represents an promising opportunity. Finally, current molecular 3D pre-training methods are designed based on TorchMD-Net (Tholke & Fabritiis ¨ , 2022). With the development of equivariant message passing neural networks, more expressive backbone architectures, such as Allegro (Musaelian et al., 2023) and MACE (Batatia et al., 2022) have been proposed, improving the prediction of molecular properties when trained from scratch. Extending pre-training strategies to these state-of-the-art architectures holds the promise of further advancing downstream tasks.
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_cleaned.md
ADDED
|
@@ -0,0 +1,205 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# MOLSPECTRA: PRE-TRAINING 3D MOLECULAR REP-RESENTATION WITH MULTI-MODAL ENERGY SPECTRA
|
| 2 |
+
|
| 3 |
+
Liang Wang1,2∗Shaozhen Liu1 Yu Rong3† Deli Zhao3 Qiang Liu1,2† Shu Wu1,2 Liang Wang1,2
|
| 4 |
+
|
| 5 |
+
1New Laboratory of Pattern Recognition (NLPR),
|
| 6 |
+
State Key Laboratory of Multimodal Artificial Intelligence Systems (MAIS),
|
| 7 |
+
Institute of Automation, Chinese Academy of Sciences (CASIA)
|
| 8 |
+
2School of Artificial Intelligence, University of Chinese Academy of Sciences
|
| 9 |
+
3DAMO Academy, Alibaba Group
|
| 10 |
+
|
| 11 |
+
# 1 INTRODUCTION
|
| 12 |
+
|
| 13 |
+
Learning 3D molecular representations from geometric conformations offers a promising approach for understanding molecular geometry and predicting quantum properties and interactions, which is significant in drug discovery and materials science (Musaelian et al., 2023; Batatia et al., 2022; Liao & Smidt, 2023; Wang et al., 2023b; Du et al., 2023b). Given the scarcity of molecular property labels, self-supervised representation pre-training has been proposed and utilized to provide generalizable representations (Hu et al., 2020; Rong et al., 2020; Ma et al., 2024).
|
| 14 |
+
|
| 15 |
+
In contrast to contrastive learning (Wang et al., 2022; Kim et al., 2022) and masked modeling (Hou et al., 2022; Liu et al., 2023c; Wang et al., 2024b) on 2D molecular graphs and molecular languages (e.g., SMILES), the design of pre-training strategies on 3D molecular geometries is more closely aligned with physical principles. Previous studies (Zaidi et al., 2023; Jiao et al., 2023) have guided representation learning through denoising processes on 3D molecular geometries, theoretically demonstrating that denoising 3D geometries is equivalent to learning molecular force fields, specifically the negative gradient of molecular potential energy with respect to position. Essentially, these studies reveal that establishing the relationship between 3D geometries and the energy states of molecular systems is an effective pathway to learn 3D molecular representations.
|
| 16 |
+
|
| 17 |
+
However, existing methods are limited to the continuous description (i.e., the potential energy function) of the molecular energy states within the classical mechanics, overlooking the quantized (discrete) energy level structures from the quantum mechanical perspective. From the quantum perspective, molecular systems exhibit quantized energy level structures, meaning that energy states can only assume specific discrete values. Specifically, different types of molecular motion, such as electronic, vibrational, and rotational motion, correspond to different energy level structures. Knowledge of these energy levels is crucial in molecular physics and quantum chemistry, as they determine the spectroscopic characteristics, chemical reactivity, and many other important molecular properties. Fortunately, experimental measurements of molecular energy spectra can reflect these structures. Meanwhile, there are many molecular spectra data obtained through experimental measurements or simulations (Zou et al., 2023; Alberts et al., 2024). Therefore, incorporating the knowledge of energy levels into molecular representation learning is expected to facilitate the development of more informative molecular representations.
|
| 18 |
+
|
| 19 |
+

|
| 20 |
+
Figure 1: The conceptual view of MolSpectra, which leverages both molecular conformation and spectra for pre-training. Prior works only model classical mechanics by denoising on conformations.
|
| 21 |
+
|
| 22 |
+
In this paper, we propose MolSpectra, a framework that incorporates molecular spectra into the pre-training of 3D molecular representations, thereby infusing the knowledge of quantized energy level structures into the representations, as shown in Figure 1. In MolSpectra, we introduce a multispectrum encoder, SpecFormer, to capture both intra-spectrum and inter-spectrum peak correlations by training with a masked patches reconstruction (MPR) objective. Additionally, we employ a contrastive objective to distills the spectral features and its inherent knowledge into the learning of 3D representations. After pre-training, the resulting 3D encoder can be fine-tuned for downstream tasks, providing expressive 3D molecular representations without the need for associated spectral data. Extensive experiments over different downstream molecular property prediction benchmarks shows the superiority of MolSpectra.
|
| 23 |
+
|
| 24 |
+
In summary, our contributions are as follows:
|
| 25 |
+
|
| 26 |
+
• We introduce quantized energy level structures and molecular spectra into 3D molecular representation pre-training for the first time, surpassing previous work that relied solely on physical knowledge within the scope of classical mechanics.
|
| 27 |
+
• We propose SpecFormer as an expressive multi-spectrum encoder, along with the masked patches reconstruction objective for spectral representation learning.
|
| 28 |
+
• We propose a contrastive objective to align molecular representations in the 3D modality and spectral modalities, enabling the pre-trained 3D encoder to infer molecular spectral features in downstream tasks without relying on spectral data.
|
| 29 |
+
• Experiments across different downstream benchmarks demonstrate that our method effectively enhances the expressiveness of the pre-trained 3D molecular representations.
|
| 30 |
+
|
| 31 |
+
# 2 PRELIMINARIES
|
| 32 |
+
|
| 33 |
+
# 2.1 NOTATIONS
|
| 34 |
+
|
| 35 |
+
Consider a molecule characterized by its 3D structure and spectra, represented as M = (a, x, S). Here, a ∈ {1, 2, . . . , 118}N specifies the atomic numbers, indicating the types of atoms within the molecule. The vector x ∈ R3N describes the conformation of the molecule, while S represents its spectra. The parameter N denotes the number of atoms in the molecule. Note that the atoms are arranged in the same order in both a and x, ensuring consistency between the atomic numbers and their corresponding spatial coordinates.
|
| 36 |
+
|
| 37 |
+
S = (s1, . . . , s|S|) represents the set of spectra for a molecule, where |S| denotes the number of spectrum types considered. In our study, we focus on three types, so |S| = 3. The first spectrum, s1 ∈ R601, is the UV-Vis spectrum, which spans from 1.5 to 13.5 eV with 601 data points at intervals of 0.02 eV. The second spectrum, s2 ∈ R3501, is the IR spectrum, covering a range from 500 to 4000 cm−1 with 3501 data points at intervals of 1 cm−1. The third spectrum, s3 ∈ R3501, is the Raman spectrum, with the same range and intervals as the IR spectrum. Together, these spectra provide a comprehensive description of the molecular characteristics across different spectral modalities.
|
| 38 |
+
|
| 39 |
+
# 2.2 PRE-TRAINING 3D MOLECULAR REPRESENTATION VIA DENOISING
|
| 40 |
+
|
| 41 |
+
Denoising has emerged as a prominent pre-training objective in 3D molecular representation learning, excelling in various downstream tasks. This method involves training models to predict and remove noise introduced deliberately into molecular structures. This approach is physically interpretable due to its proven equivalence to learning the molecular force field.
|
| 42 |
+
|
| 43 |
+
Equivalence between denoising and learning molecular force fields. The equivalence between coordinate denoising and force field learning is established by Zaidi et al. (2023). For a given molecule M, perturb its equilibrium structure x0 according to the distribution p(x|x0), where x is the noisy conformation. Assuming the molecular distribution adheres to the energy-based Boltzmann distribution with respect to the energy function E(·), then
|
| 44 |
+
|
| 45 |
+

|
| 46 |
+
|
| 47 |
+
where GNNθ(x) denotes a graph neural network parameterized by θ, which processes the conformation x to produce node-level predictions. The notation ≃ signifies the equivalence of different objectives. The proof of this equivalence is provided in the Appendix A. In prior research, the energy function E(·) has been defined in several forms. Below are three representative studies.
|
| 48 |
+
|
| 49 |
+
Energy function I: mixture of isotropic Gaussians. In Coord (Zaidi et al., 2023), the energy function is approximated using a mixture of isotropic Gaussians centered at the known equilibrium structures to replace the Boltzmann distribution, since these structures are local maxima of the Boltzman distribution. Leveraging the equivalence between the score-matching objective and denoising autoencoders (Vincent, 2011), the following denoising-based energy function ECoord(·) is derived:
|
| 50 |
+
|
| 51 |
+

|
| 52 |
+
|
| 53 |
+
Note that this objective is derived under the assumption of isotropic Gaussian noise, i.e., p(x|x0) ∼ N (x0, τ 2c I3N ), where I3N represents the identity matrix of size 3N , and the subscript c indicates the coordinate denoising approach.
|
| 54 |
+
|
| 55 |
+
Energy function II: mixture of anisotropic Gaussians. Considering rigid and flexible components in molecular structures, isotropic Gaussian can lead to significant approximation errors. To address the anisotropic distribution, Frad (Feng et al., 2023) introduces hybrid noise on dihedral angles of rotatable bonds and atomic coordinates, incorporating fractional denoising of the coordinate noise. The equilibrium structure x0 is initially perturbed by dihedral angle noise p(ψa|ψ0) ∼ N (ψ0, σ2f Im), followed by coordinate noise p(x|xa) ∼ N (xa, τ 2f I3N ). Here, ψa, ψ0 ∈ [0, 2π)m represent to the dihedral angles of rotatable bonds in structures xa and x0, respectively, with m denoting the number of rotatable bonds. The subscript f indicates the fractional denoising approach. Subsequently, the energy function is induced:
|
| 56 |
+
|
| 57 |
+

|
| 58 |
+
|
| 59 |
+
where Στf ,σf = τ 2f I3N + σ2f CC⊤, and C ∈ R3N×m is a matrix used to linearly transform the dihedral angle noise into coordinate change, expressed as ∆x ≈ C∆ψ.
|
| 60 |
+
|
| 61 |
+
Energy function III: classical potential energy theory. SliDe (Ni et al., 2024) derives energy function from classical molecular potential energy theory (Alavi, 2020; Zhou & Liu, 2022). In this
|
| 62 |
+
|
| 63 |
+

|
| 64 |
+
Figure 2: Overview of the MolSpectra pre-training framework. Our pre-training framework comprises three sub-objectives: the denoising objective and the MPR objective, which respectively guide the representation learning of the 3D and spectral modalities, and the contrastive objective, which aligns the representations of both modalities.
|
| 65 |
+
|
| 66 |
+
form, the total intramolecular potential energy is mainly attributed to three types of interactions: bond stretching, bond angle bending, and bond torsion. The following energy function is derived:
|
| 67 |
+
|
| 68 |
+

|
| 69 |
+
|
| 70 |
+
where r ∈ (R≥0)m1 , θ ∈ [0, 2π)m2 , ϕ ∈ [0, 2π)m3 represent vectors of the bond lengths, bond angles, and bond torsion angles of the molecule, respectively. r0, θ0, ϕ0 correspond to the respective equilibrium values. The parameter vectors kB, kA, kT determine the interaction strength.
|
| 71 |
+
|
| 72 |
+
# 3 THE PROPOSED MOLSPECTRA METHOD
|
| 73 |
+
|
| 74 |
+
Considering the complementarity of different spectra, we introduce multiple spectra into molecular representation learning. To effectively comprehend molecular spectra, we designed a Transformerbased multi-spectrum encoder, SpecFormer, along with a masked reconstruction objective to guide its training. Finally, a contrastive objective is employed to align the 3D encoding guided by the denoising objective with the spectra encoding guided by the reconstruction objective, endowing the 3D encoding with the capability to understand spectra and the knowledge they encompass.
|
| 75 |
+
|
| 76 |
+
# 3.1 SPECFORMER: A SINGLE-STREAM ENCODER FOR MULTI-MODAL ENERGY SPECTRA
|
| 77 |
+
|
| 78 |
+
For different types of spectra, each spectrum is independently patched and initially encoded. Then, all the resulting patch embeddings are concatenated and encoded using a Transformer-based encoder.
|
| 79 |
+
|
| 80 |
+
Patching. Compared to directly encoding individual frequency points, we divided each spectrum into multiple patches. This approach offers two distinct advantages: (i) By forming patches from adjacent frequency points, local semantic features, such as absorption peaks, can be captured more effectively. (ii) It reduces the computational overhead of subsequent Transformer layers. Technically, each spectrum si ∈ RLi where i = 1, · · · , |S| is first divided into patches according to the patch length Pi and the stride Di. When 0 < Di < Pi, the consecutive patches will be overlapped with overlapping region length Pi − Di. When Di = Pi, the consecutive patches will be nonoverlapped. Li denotes the length of si. The patching process on each spectrum will generate a sequence of patches pi ∈ RNi×Pi, where Ni = Li−PiDi + 1 is the number of patches.
|
| 81 |
+
|
| 82 |
+
Patch encoding and position encoding. Prior to be fed into the encoder, the patches of the i-th spectrum are mapped to the latent space of dimension d via a trainable linear projection Wi ∈ RPi×d. A learnable additive position encoding W posi ∈ RNi×d is applied to maintain the order of the patches: p′i = piWi + W posi , where p′i ∈ RNi×d denotes the latent representation of the spectrum si that will be fed into the subsequent SpecFormer encoder.
|
| 83 |
+
|
| 84 |
+
SpecFormer: multi-spectrum Transformer encoder. Although several encoders have been proposed to map molecular spectrum into implicit representations, such as the CNN-AM (Tao et al., 2024) based on one-dimensional convolution, these encoders are designed to encode only a single type of spectrum. In our approach, multiple molecular spectra (UV-Vis, IR, Raman) are jointly considered. When encoding multiple spectra of a molecule simultaneously, an observation caught our attention and led us to adopt a Transformer-based encoder with multiple spectra as input, similar to the single-stream Transformer in multi-modal learning (Shin et al., 2021).
|
| 85 |
+
|
| 86 |
+
The observation refers to the fact that the same functional group not only causes multiple peaks within a single spectrum, but also generates peaks across different spectra. As shown on the left of Figure 3, the different vibrational modes of the methyl group (-CH3) in methanol (CH3OH) result in three peaks in the IR spectrum, indicating intra-spectrum dependencies among these peaks. A similar phenomenon occurs with the hydroxyl group (-OH) in methanol. Additionally, the aromatic ring in phenol (C6H5OH), shown on the right of
|
| 87 |
+
|
| 88 |
+

|
| 89 |
+
Figure 3: Illustration of intra-spectrum (left) and interspectrum (right) dependencies.
|
| 90 |
+
|
| 91 |
+
Figure 3, not only produces multiple peaks in the IR spectrum due to different vibrational modes but also causes an absorption peak near 270 nm in the UV-Vis spectrum due to the π → π∗ transition in the aromatic ring, demonstrating the existence of inter-spectrum dependencies. Such dependencies have been theoretically studied, for example, in the context of vibronic coupling (Kong et al., 2021).
|
| 92 |
+
|
| 93 |
+
To capture intra-spectrum and inter-spectrum dependencies, we concatenate the embeddings obtained from patch encoding and position encoding of different spectra: pˆ = p′1∥ · · · ∥p′|S| ∈ R(P|S|i=1 Ni)×d, and then input them into the Transformer encoder as depicted in Figure 2. Then each head h = 1, . . . , H in multi-head attention will transform them into query matrices Qh = pWˆ Qh , key matrices Kh = pWˆ Kh and value matrices Vh = pWˆ Vh , where W Qh , W Kh ∈ Rd×dk and WVh ∈ Rd× dH . Afterward, a scaled product is utilized to obtain the attention output Oh ∈ R(P|S|i=1 Ni)× dH :
|
| 94 |
+
|
| 95 |
+

|
| 96 |
+
|
| 97 |
+
The multi-head attention block also includes BatchNorm layers and a feed forward network with residual connections as shown in Figure 2. After combining the outputs of all heads, it generates the representation denoted as z ∈ R(P|S|i=1 Ni)×d. Finally, a flatten layer with representation projection head is used to obtain the molecular spectra representation zs ∈ Rd.
|
| 98 |
+
|
| 99 |
+
# 3.2 MASKED PATCHES RECONSTRUCTION PRE-TRAINING FOR SPECTRA
|
| 100 |
+
|
| 101 |
+
Before distilling the spectra information into 3D molecular representation learning, we need first ensure that the spectrum encoder can effectively comprehend molecular spectra and generate spectral representations. Considering the success of masking modeling across various domains (Devlin et al., 2019; He et al., 2022; Hou et al., 2022; Xia et al., 2023; Wang et al., 2024b; Nie et al., 2023), we propose a masked patches reconstruction (MPR) objective to guide the training of SpecFormer.
|
| 102 |
+
|
| 103 |
+
After the patching step, we randomly select a portion of patches according to the mask ratio α and replace them with zero vectors to implement the masking. Subsequently, the masked patches undergo patch encoding and position encoding. In this way, the semantics of the masked patches (the absorption intensity at specific wavelengths) are obscured during patch encoding, while the positional information is retained to facilitate the reconstruction of the original semantics.
|
| 104 |
+
|
| 105 |
+
After encoding by SpecFormer, the encoded results corresponding to the masked patches are input into a spectrum-specific reconstruction head to reconstruct the original spectral values that were masked. The mean squared error (MSE) between the reconstruction results and the original masked spectra serves as the loss function for the MPR task, guiding the training of SpecFormer:
|
| 106 |
+
|
| 107 |
+

|
| 108 |
+
|
| 109 |
+
where Pi denotes the set of masked patches in the i-th type of molecular spectra, and pˆi,j denotes ethe reconstructed patch corresponding to the masked patch pi,j .
|
| 110 |
+
|
| 111 |
+
# 3.3 CONTRASTIVE LEARNING BETWEEN 3D STRUCTURES AND SPECTRA
|
| 112 |
+
|
| 113 |
+
Under the guidance of the denoising objective for 3D representation learning and the MPR objective for spectral representation learning, we further introduce a contrastive objective to align the representations across these two modalities. We treat the 3D representation zx ∈ Rd and spectral representation zs ∈ Rd of the same molecule as positive samples, and negative samples otherwise. Subsequently, the consistency between positive samples and the discrepancy between negative samples are maximized through the contrastive objective. Given the theoretical and empirical effectiveness, we employ InfoNCE (van den Oord et al., 2018) as the contrastive objective:
|
| 114 |
+
|
| 115 |
+

|
| 116 |
+
|
| 117 |
+
where zjx, zjs are randomly sampled 3D and spectra views regarding to the positive pair (zx, zs). fx(zx, zs) and fs(zs, zx) are scoring functions for the two corresponding views, with flexible formulations. Here we adopt fx(zx, zs) = fs(zs, zx) = ⟨zx, zs⟩.
|
| 118 |
+
|
| 119 |
+
Note that the denoising objective can utilize any form from existing 3D molecular representation pre-training studies, enabling seamless integration of our method into these frameworks.
|
| 120 |
+
|
| 121 |
+
# 3.4 TWO-STAGE PRE-TRAINING PIPELINE
|
| 122 |
+
|
| 123 |
+
Previous pre-training efforts for 3D molecular representation have been conducted on unlabeled datasets using denoising objective. These datasets typically provide only equilibrium 3D structures without offering spectra for all molecules. To enhance the pre-training effect by incorporating spectra while leveraging denoising pre-training, we employ a two-stage pre-training approach. The first stage involves training on a larger dataset (Nakata & Shimazaki, 2017) without spectra using only the denoising objective. Subsequently, the second stage involves training on a dataset that includes spectra using the complete objective as follows:
|
| 124 |
+
|
| 125 |
+

|
| 126 |
+
|
| 127 |
+
where βDenoising, βMPR, and βContrast denote the weights of each sub-objective.
|
| 128 |
+
|
| 129 |
+
# 4 EXPERIMENTS
|
| 130 |
+
|
| 131 |
+
To comprehensively evaluate the impact of molecular spectra on molecular tasks, we first verify the effectiveness of molecular spectra in the training-from-scratch method for the downstream task. Furthermore, we evaluate the effectiveness of our pre-training framework MolSpectra.
|
| 132 |
+
|
| 133 |
+
# 4.1 EFFECTIVENESS OF MOLECULAR SPECTRA IN TRAINING FROM SCRATCH
|
| 134 |
+
|
| 135 |
+
This pilot experiment aims to demonstrate the rationality for incorporating molecular spectra into pre-training. We introduce additional spectral features into a train-from-scratch molecular property prediction model to observe the impact of spectral information on prediction outcomes. We employ EGNN (Satorras et al., 2021), a representative 3D molecular encoder, equipped with an MLP-based prediction head as the baseline model. While EGNN encodes the 3D representations, the UV-Vis spectrum of each molecule provided by the QM9S (Zou et al., 2023) dataset is encoded into spectral representations by a spectrum encoder. Before making predictions with the final MLP, we concatenate the spectral and 3D representations for prediction. The results are presented in Table 1.
|
| 136 |
+
|
| 137 |
+
Table 1: Performance (MAE ↓) when training from scratch on QM9 dataset.
|
| 138 |
+

|
| 139 |
+
|
| 140 |
+
We observe that by directly concatenating spectral representations, the performance of molecular property prediction can be effectively enhanced. This indicates that the information from molecular spectra is beneficial for downstream molecular property prediction. Further incorporating molecular spectra into the pre-training phase of molecular representation has the potential to enhance the informativeness and generalization capability of the representations, thereby broadly improving the performance of downstream tasks.
|
| 141 |
+
|
| 142 |
+
# 4.2 EFFECTIVENESS OF MOLECULAR SPECTRA IN REPRESENTATION PRE-TRAINING
|
| 143 |
+
|
| 144 |
+
We conduct experiments to evaluate MolSpectra by first introducing spectral data into the pretraining of 3D representations, followed by evaluating the performance on downstream tasks. For a comprehensive comparison, two types of baselines are adopted: (1) training-from-scratch methods, including SchNet (Schutt et al. ¨ , 2017), EGNN, DimeNet (Klicpera et al., 2020b), DimeNet++ (Klicpera et al., 2020a), PaiNN (Schutt et al. ¨ , 2021), SphereNet (Liu et al., 2021), and TorchMD-Net (Tholke & Fabritiis ¨ , 2022); and (2) pre-training methods, including Transformer-M (Luo et al., 2023), SE(3)-DDM (Liu et al., 2023b), 3D-EMGP (Jiao et al., 2023), and Coord.
|
| 145 |
+
|
| 146 |
+
MolSpectra can be seamlessly plugged into any existing denoising method. To evaluate the enhancement provided by our method compared to denoising alone, we select the representative coordinate denoising (Coord) as our denoising sub-objective. This method also serves as our primary baseline.
|
| 147 |
+
|
| 148 |
+
# 4.2.1 PRE-TRAINING DATASET.
|
| 149 |
+
|
| 150 |
+
As described in Section 3.4, we first perform denoising pre-training on the PCQM4Mv2 (Nakata & Shimazaki, 2017) dataset, followed by a second stage of pre-training on the QM9Spectra (QM9S) (Zou et al., 2023) dataset, which includes multi-modal molecular energy spectra. In both stages, we adopt the denoising objective provided by Coord (Zaidi et al., 2023), as defined in Eq. 2.
|
| 151 |
+
|
| 152 |
+
The QM9S dataset comprises organic molecules from the QM9 (Ramakrishnan et al., 2014) dataset. The UV-Vis, IR, and Raman spectra of the molecules are calculated at the B3LYP/def-TZVP level of theory, through frequency analysis and time-dependent density functional theory (TD-DFT).
|
| 153 |
+
|
| 154 |
+
# 4.2.2 QM9
|
| 155 |
+
|
| 156 |
+
The QM9 dataset is a quantum chemistry dataset comprising over 134,000 small molecules, each consisting of up to 9 heavy atoms (C, N, O, F) and additional H atoms. This dataset provides an equilibrium geometric conformation for each molecule along with 12 property labels. The dataset is divided into a training set of 110k molecules, a validation set of 10k molecules, and a test set containing the remaining over 10k molecules. Prediction errors are measured using the mean absolute error (MAE). The experimental results are presented in Table 2.
|
| 157 |
+
|
| 158 |
+
The 3D molecular representations pre-trained using our method are fine-tuned and used for prediction across various properties, achieving state-of-the-art performance in 8 out of 12 properties and outperforms Coord in 10 out of 12 properties. In conjunction with the observations in Section 4.1, the performance improvement can be attributed to our incorporation of an understanding of molecular spectra and the knowledge they entail into the 3D molecular representations.
|
| 159 |
+
|
| 160 |
+
Table 2: Performance (MAE↓) on QM9 dataset. The compared methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are highlighted in bold.
|
| 161 |
+

|
| 162 |
+
|
| 163 |
+
Table 3: Performance (MAE↓) on MD17 force prediction (kcal/mol/ ˚A). The methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are in bold.
|
| 164 |
+

|
| 165 |
+
|
| 166 |
+
# 4.2.3 MD17
|
| 167 |
+
|
| 168 |
+
The MD17 dataset contains molecular dynamics trajectories for eight organic molecules, including aspirin, benzene, and ethanol. It offers 150k to nearly 1M conformations per molecule, with energy and force labels. Unlike QM9, MD17 emphasizes dynamic behavior in addition to static properties. We use a standard limited data split: models train on 1k samples, validate on 50, and test on the rest. Performance is evaluated using MAE, with results in Table 3.
|
| 169 |
+
|
| 170 |
+
Our approach also results in the expected performance improvement on MD17. MD17 is a dataset comprising a large number of non-equilibrium molecular structures and their corresponding force fields, which serves to evaluate a model’s understanding of molecular dynamics. However, previous pre-training methods based solely on denoising have only learned force field patterns at static equilibrium states, failing to adequately capture the dynamic evolution of molecular systems. In contrast, our MolSpectra learns the dynamic evolution of molecules by understanding energy level transition patterns, thereby outperforming denoising-based pre-training methods.
|
| 171 |
+
|
| 172 |
+
4.3 SENSITIVITY ANALYSIS OF PATCH LENGTH Pi, STRIDE Di, AND MASK RATIO α
|
| 173 |
+
|
| 174 |
+
We conduct experiments to evaluate the impact of patch length Pi, stride Di, and mask ratio α.
|
| 175 |
+
Results are summarized in Table 4 and Table 5.
|
| 176 |
+
|
| 177 |
+
From Table 4, we observe that when consecutive patches have overlap (Di < Pi), the performance of pre-training is superior compared to scenarios without overlap (Di = Pi). Specifically, the performance is optimal when the stride is half of the patch length. This is because appropriate overlap can better preserve and capture local features, particularly the information at the patch boundaries. Additionally, we find that choosing an appropriate patch length further enhances performance. In our experiments, the configuration of Pi = 20, Di = 10 yields the best results.
|
| 178 |
+
|
| 179 |
+
Table 4: Sensitivity of patch length and stride.
|
| 180 |
+

|
| 181 |
+
|
| 182 |
+
Table 5: Sensitivity of mask ratio.
|
| 183 |
+

|
| 184 |
+
|
| 185 |
+
Regarding the mask ratio, α = 0.10 is a preferable choice. A small mask ratio results in insufficient MPR optimization, hindering SpecFormer training. Conversely, a large mask ratio causes excessive spectral perturbation, degrading performance when aligning with the 3D representations with the contrastive objective. An appropriate mask ratio strikes a balance between these two aspects.
|
| 186 |
+
|
| 187 |
+
# 4.4 ABLATION STUDY
|
| 188 |
+
|
| 189 |
+
To rigorously demonstrate the contributions of masked patches reconstruction, the incorporation of molecular spectra, and each spectral modality, we conducted an ablation study on them.
|
| 190 |
+
|
| 191 |
+
Ablation study of masked patches reconstruction. We remove the MPR loss to analyze the impact of masked patches reconstruction, referred to as “w/o MPR” in Table 6. Removing the MPR objective leads to performance deterioration. This is consistent with the sensitivity analysis of the mask ratio α in Section 4.3, as removing MPR is an extreme case where α = 0. This decline is due to the lack of effective guidance in training SpecFormer. Using an undertrained SpecFormer for contrastive learning with 3D encoder outputs limits performance improvement.
|
| 192 |
+
|
| 193 |
+
Table 6: Ablation of optimization objectives.
|
| 194 |
+

|
| 195 |
+
|
| 196 |
+
Ablation study of molecular spectra. We retain only the denoising loss, removing both the MPR loss and contrastive loss, referred to as “w/o MPR, Contrast” in Table 6. The only difference between this variant and MolSpectra is the incorporation of molecular spectra into the pre-training. The ”w/o MPR, Contrast” results are inferior to those of MolSpectra, highlighting that incorporating molecular spectra effectively enhances the quality and generalizability of molecular 3D representations.
|
| 197 |
+
|
| 198 |
+
Ablation study of each spectral modality. To evaluate the contributions of each spectral modality to the performance, we conduct an ablation study for each modality. The results are presented in Table 7. It can be observed that each spectral modality contributes differently, with the UV-Vis spectrum having the smallest contribution and the IR spectrum the largest, likely due to the varying information content in each modality.
|
| 199 |
+
|
| 200 |
+
Table 7: Ablation of spectral modalities.
|
| 201 |
+

|
| 202 |
+
|
| 203 |
+
# 6 CONCLUSION
|
| 204 |
+
|
| 205 |
+
In this study, we explore pre-training molecular 3D representations beyond classical mechanics. By leveraging the correlation between molecular energy level structures and molecular spectra in quantum mechanics, we introduce molecular spectra for pre-training molecular 3D representations (MolSpectra). By aligning the 3D encoder trained with a denoising objective and the spectrum encoder trained with a masked patch reconstruction objective, we enhance the informativeness and transferability of the resulting 3D representations.
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_content_list.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_layout.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2bea3a8ab36d8527a4ee6555991a67905b9eab3ef2f66fa3c3403a1391ef0f8b
|
| 3 |
+
size 5904498
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_middle.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_model.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_origin.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8efe3d55230ed01ebeeae37cbf0d94a8cd5153571ffdc70bfce053790cdee45a
|
| 3 |
+
size 5683996
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/6027_MolSpectra_Pre_training_3_span.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c9950ec961b2b5f98f4c0d6bc615042e6fb0f8d82f6b4871d39821c7cf84c515
|
| 3 |
+
size 5896226
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/dag.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/032759948e20868c031881dac87a152765d78ea757955c32f34b83ca8e975d1d.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/07aae76295c011d4cdd34a6c00be2fe8427447017185173c5e25e9468ccf833d.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/084cd722defc01e058c1747c1103ac680c0cd8217a93077cab2a30ae06000a37.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/16005489fb642f13912d85a6d90523ef0a5e1d56b5c30d9d2ee35336fdb04f4f.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/1695d160ca54e8992fed88c81378f03bd31abccebafebc3c6a3d5dcadc6a747f.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/1ee1e5bd4e35d7acf3b80e7404b113daf11b6925490369aaa637e60964fb744d.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/27c884c07b2f80cded2a18e590a81c38cc2c409bbb73f478c4432366bc9ede6d.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/35c4ba087f1ae7299aa554638aef5647f9efa2de01450fec172beb01bf64538e.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/43ac77c0fc35ecea17e2a91075a4f1c272643487659bc83241629c4a80c6ae86.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/5237f9010e1b7bb79b84bdc91c83cb5152b351c1f8b05fe25b9f0b961f759e2d.jpg
ADDED
|
Git LFS Details
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/55d0ac7d29bb938c1a28dc8f66121322e7c07f72ae0f7cfa10cb5cd784b67c21.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/6b6a48b7abea8c9dc0c3c72345c9c94cbb0ac1d1ef8824d485ff0ace9b5a0a1e.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/6c50a187dd8276b7372af9e7e00b99521d3741fa210eee7a1af570b334053570.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/71b0bba244aef7a34cb53a9293a46808b73589fc9a53aa7b50884d5e277c2b36.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/7ad84d22cd3c911b24b3de89b1077c3b9f08c6b762840ecd38160beb882f3a16.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/7c18349955cc82bf082d353ddb8f1ff323edb09a34c2ad852dba320d3a0a3faa.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/88fc24d2ba2e08617faed6c58a55f1e92d5803ea4ae7a49eaaf118fe4ad56429.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/977b41496865477e7652249bec51630a1a037097fa74f17c10e8a72851cd7ce3.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/b1279274391a8ee385cd510483726424c1b7bb542dd3dfcaa664dd7784591ce0.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/b5f78fc1f86d0c7c6f996497a567b0462a6e0529447787d81ffe3fdbb71b3ccb.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/bdae202d54ee161317731819baf6967d189260b12ff333dc514a8cc065c47f97.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/d1786526fd710a309dcca5721e53bf82d19ca2622ac3314d3bfc46284f668f1d.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/dcbdd7f1637fddc90c9a9616e1e732fa41edaf190414db6aba1fd62b99a32716.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/images/e78ea26b8e8328ff5cf96a82a322d391d8879301b032bec5c51b78065f043e34.jpg
ADDED
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/success_dag.txt
ADDED
|
File without changes
|
mineru_outputs/6027_MolSpectra_Pre_training_3/auto/visual_dag.json
ADDED
|
@@ -0,0 +1,130 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"nodes": [
|
| 3 |
+
{
|
| 4 |
+
"name": "",
|
| 5 |
+
"caption": "Figure 1: The conceptual view of MolSpectra, which leverages both molecular conformation and spectra for pre-training. Prior works only model classical mechanics by denoising on conformations.",
|
| 6 |
+
"visual_node": 1,
|
| 7 |
+
"formula": 0,
|
| 8 |
+
"resolution": "1091x372"
|
| 9 |
+
},
|
| 10 |
+
{
|
| 11 |
+
"name": "",
|
| 12 |
+
"caption": "Equation representing the equivalence between denoising and learning molecular force fields.",
|
| 13 |
+
"visual_node": 1,
|
| 14 |
+
"formula": 1,
|
| 15 |
+
"resolution": "633x83"
|
| 16 |
+
},
|
| 17 |
+
{
|
| 18 |
+
"name": "",
|
| 19 |
+
"caption": "Equation defining the denoising-based energy function ECoord derived from isotropic Gaussian noise.",
|
| 20 |
+
"visual_node": 1,
|
| 21 |
+
"formula": 1,
|
| 22 |
+
"resolution": "436x69"
|
| 23 |
+
},
|
| 24 |
+
{
|
| 25 |
+
"name": "",
|
| 26 |
+
"caption": "Equation defining the energy function induced by anisotropic Gaussian noise on dihedral angles and coordinates.",
|
| 27 |
+
"visual_node": 1,
|
| 28 |
+
"formula": 1,
|
| 29 |
+
"resolution": "475x58"
|
| 30 |
+
},
|
| 31 |
+
{
|
| 32 |
+
"name": "",
|
| 33 |
+
"caption": "Figure 2: Overview of the MolSpectra pre-training framework. Our pre-training framework comprises three sub-objectives: the denoising objective and the MPR objective, which respectively guide the representation learning of the 3D and spectral modalities, and the contrastive objective, which aligns the representations of both modalities.",
|
| 34 |
+
"visual_node": 1,
|
| 35 |
+
"formula": 0,
|
| 36 |
+
"resolution": "1102x619"
|
| 37 |
+
},
|
| 38 |
+
{
|
| 39 |
+
"name": "",
|
| 40 |
+
"caption": "Equation defining the energy function derived from classical molecular potential energy theory involving bond stretching, bending, and torsion.",
|
| 41 |
+
"visual_node": 1,
|
| 42 |
+
"formula": 1,
|
| 43 |
+
"resolution": "877x131"
|
| 44 |
+
},
|
| 45 |
+
{
|
| 46 |
+
"name": "",
|
| 47 |
+
"caption": "Figure 3: Illustration of intra-spectrum (left) and interspectrum (right) dependencies.",
|
| 48 |
+
"visual_node": 1,
|
| 49 |
+
"formula": 0,
|
| 50 |
+
"resolution": "656x350"
|
| 51 |
+
},
|
| 52 |
+
{
|
| 53 |
+
"name": "",
|
| 54 |
+
"caption": "Equation for the scaled dot-product attention output in the multi-head attention block.",
|
| 55 |
+
"visual_node": 1,
|
| 56 |
+
"formula": 1,
|
| 57 |
+
"resolution": "659x75"
|
| 58 |
+
},
|
| 59 |
+
{
|
| 60 |
+
"name": "",
|
| 61 |
+
"caption": "Equation defining the masked patches reconstruction (MPR) loss function.",
|
| 62 |
+
"visual_node": 1,
|
| 63 |
+
"formula": 1,
|
| 64 |
+
"resolution": "420x95"
|
| 65 |
+
},
|
| 66 |
+
{
|
| 67 |
+
"name": "",
|
| 68 |
+
"caption": "Equation defining the InfoNCE contrastive objective function.",
|
| 69 |
+
"visual_node": 1,
|
| 70 |
+
"formula": 1,
|
| 71 |
+
"resolution": "1106x92"
|
| 72 |
+
},
|
| 73 |
+
{
|
| 74 |
+
"name": "",
|
| 75 |
+
"caption": "Equation defining the complete objective function combining denoising, MPR, and contrastive losses.",
|
| 76 |
+
"visual_node": 1,
|
| 77 |
+
"formula": 1,
|
| 78 |
+
"resolution": "628x36"
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"name": "",
|
| 82 |
+
"caption": "Table 1: Performance (MAE ↓) when training from scratch on QM9 dataset.",
|
| 83 |
+
"visual_node": 1,
|
| 84 |
+
"formula": 0,
|
| 85 |
+
"resolution": "1100x156"
|
| 86 |
+
},
|
| 87 |
+
{
|
| 88 |
+
"name": "",
|
| 89 |
+
"caption": "Table 2: Performance (MAE↓) on QM9 dataset. The compared methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are highlighted in bold.",
|
| 90 |
+
"visual_node": 1,
|
| 91 |
+
"formula": 0,
|
| 92 |
+
"resolution": "1095x394"
|
| 93 |
+
},
|
| 94 |
+
{
|
| 95 |
+
"name": "",
|
| 96 |
+
"caption": "Table 3: Performance (MAE↓) on MD17 force prediction (kcal/mol/ ˚A). The methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are in bold.",
|
| 97 |
+
"visual_node": 1,
|
| 98 |
+
"formula": 0,
|
| 99 |
+
"resolution": "986x327"
|
| 100 |
+
},
|
| 101 |
+
{
|
| 102 |
+
"name": "",
|
| 103 |
+
"caption": "Table 4: Sensitivity of patch length and stride.",
|
| 104 |
+
"visual_node": 1,
|
| 105 |
+
"formula": 0,
|
| 106 |
+
"resolution": "598x231"
|
| 107 |
+
},
|
| 108 |
+
{
|
| 109 |
+
"name": "",
|
| 110 |
+
"caption": "Table 5: Sensitivity of mask ratio.",
|
| 111 |
+
"visual_node": 1,
|
| 112 |
+
"formula": 0,
|
| 113 |
+
"resolution": "358x231"
|
| 114 |
+
},
|
| 115 |
+
{
|
| 116 |
+
"name": "",
|
| 117 |
+
"caption": "Table 6: Ablation of optimization objectives.",
|
| 118 |
+
"visual_node": 1,
|
| 119 |
+
"formula": 0,
|
| 120 |
+
"resolution": "486x150"
|
| 121 |
+
},
|
| 122 |
+
{
|
| 123 |
+
"name": "",
|
| 124 |
+
"caption": "Table 7: Ablation of spectral modalities.",
|
| 125 |
+
"visual_node": 1,
|
| 126 |
+
"formula": 0,
|
| 127 |
+
"resolution": "508x192"
|
| 128 |
+
}
|
| 129 |
+
]
|
| 130 |
+
}
|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/1 INTRODUCTION.md_dag.json
ADDED
|
@@ -0,0 +1,146 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"nodes": [
|
| 3 |
+
{
|
| 4 |
+
"name": "1 INTRODUCTION",
|
| 5 |
+
"content": "# 1 INTRODUCTION Learning 3D molecular representations from geometric conformations offers a promising approach for understanding molecular geometry and predicting quantum properties and interactions, which is significant in drug discovery and materials science (Musaelian et al., 2023; Batatia et al., 2022; Liao & Smidt, 2023; Wang et al., 2023b; Du et al., 2023b). Given the scarcity of molecular property labels, self-supervised representation pre-training has been proposed and utilized to provide generalizable representations (Hu et al., 2020; Rong et al., 2020; Ma et al., 2024). In contrast to contrastive learning (Wang et al., 2022; Kim et al., 2022) and masked modeling (Hou et al., 2022; Liu et al., 2023c; Wang et al., 2024b) on 2D molecular graphs and molecular languages (e.g., SMILES), the design of pre-training strategies on 3D molecular geometries is more closely aligned with physical principles. Previous studies (Zaidi et al., 2023; Jiao et al., 2023) have guided representation learning through denoising processes on 3D molecular geometries, theoretically demonstrating that denoising 3D geometries is equivalent to learning molecular force fields, specifically the negative gradient of molecular potential energy with respect to position. Essentially, these studies reveal that establishing the relationship between 3D geometries and the energy states of molecular systems is an effective pathway to learn 3D molecular representations. However, existing methods are limited to the continuous description (i.e., the potential energy function) of the molecular energy states within the classical mechanics, overlooking the quantized (discrete) energy level structures from the quantum mechanical perspective. From the quantum perspective, molecular systems exhibit quantized energy level structures, meaning that energy states can only assume specific discrete values. Specifically, different types of molecular motion, such as electronic, vibrational, and rotational motion, correspond to different energy level structures. Knowledge of these energy levels is crucial in molecular physics and quantum chemistry, as they determine the spectroscopic characteristics, chemical reactivity, and many other important molecular properties. Fortunately, experimental measurements of molecular energy spectra can reflect these structures. Meanwhile, there are many molecular spectra data obtained through experimental measurements or simulations (Zou et al., 2023; Alberts et al., 2024). Therefore, incorporating the knowledge of energy levels into molecular representation learning is expected to facilitate the development of more informative molecular representations.  Figure 1: The conceptual view of MolSpectra, which leverages both molecular conformation and spectra for pre-training. Prior works only model classical mechanics by denoising on conformations. In this paper, we propose MolSpectra, a framework that incorporates molecular spectra into the pre-training of 3D molecular representations, thereby infusing the knowledge of quantized energy level structures into the representations, as shown in Figure 1. In MolSpectra, we introduce a multispectrum encoder, SpecFormer, to capture both intra-spectrum and inter-spectrum peak correlations by training with a masked patches reconstruction (MPR) objective. Additionally, we employ a contrastive objective to distills the spectral features and its inherent knowledge into the learning of 3D representations. After pre-training, the resulting 3D encoder can be fine-tuned for downstream tasks, providing expressive 3D molecular representations without the need for associated spectral data. Extensive experiments over different downstream molecular property prediction benchmarks shows the superiority of MolSpectra. In summary, our contributions are as follows: • We introduce quantized energy level structures and molecular spectra into 3D molecular representation pre-training for the first time, surpassing previous work that relied solely on physical knowledge within the scope of classical mechanics. • We propose SpecFormer as an expressive multi-spectrum encoder, along with the masked patches reconstruction objective for spectral representation learning. • We propose a contrastive objective to align molecular representations in the 3D modality and spectral modalities, enabling the pre-trained 3D encoder to infer molecular spectral features in downstream tasks without relying on spectral data. • Experiments across different downstream benchmarks demonstrate that our method effectively enhances the expressiveness of the pre-trained 3D molecular representations.",
|
| 6 |
+
"edge": [
|
| 7 |
+
"Background and Current Methods",
|
| 8 |
+
"Limitations and Motivation",
|
| 9 |
+
"Proposed Method: MolSpectra",
|
| 10 |
+
"Contributions"
|
| 11 |
+
],
|
| 12 |
+
"level": 1,
|
| 13 |
+
"visual_node": []
|
| 14 |
+
},
|
| 15 |
+
{
|
| 16 |
+
"name": "Background and Current Methods",
|
| 17 |
+
"content": "Learning 3D molecular representations from geometric conformations offers a promising approach for understanding molecular geometry and predicting quantum properties and interactions, which is significant in drug discovery and materials science (Musaelian et al., 2023; Batatia et al., 2022; Liao & Smidt, 2023; Wang et al., 2023b; Du et al., 2023b). Given the scarcity of molecular property labels, self-supervised representation pre-training has been proposed and utilized to provide generalizable representations (Hu et al., 2020; Rong et al., 2020; Ma et al., 2024). In contrast to contrastive learning (Wang et al., 2022; Kim et al., 2022) and masked modeling (Hou et al., 2022; Liu et al., 2023c; Wang et al., 2024b) on 2D molecular graphs and molecular languages (e.g., SMILES), the design of pre-training strategies on 3D molecular geometries is more closely aligned with physical principles. Previous studies (Zaidi et al., 2023; Jiao et al., 2023) have guided representation learning through denoising processes on 3D molecular geometries, theoretically demonstrating that denoising 3D geometries is equivalent to learning molecular force fields, specifically the negative gradient of molecular potential energy with respect to position. Essentially, these studies reveal that establishing the relationship between 3D geometries and the energy states of molecular systems is an effective pathway to learn 3D molecular representations.",
|
| 18 |
+
"edge": [
|
| 19 |
+
"Importance of 3D Molecular Representations",
|
| 20 |
+
"Physical Principles in 3D Pre-training"
|
| 21 |
+
],
|
| 22 |
+
"level": 2,
|
| 23 |
+
"visual_node": []
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"name": "Limitations and Motivation",
|
| 27 |
+
"content": "However, existing methods are limited to the continuous description (i.e., the potential energy function) of the molecular energy states within the classical mechanics, overlooking the quantized (discrete) energy level structures from the quantum mechanical perspective. From the quantum perspective, molecular systems exhibit quantized energy level structures, meaning that energy states can only assume specific discrete values. Specifically, different types of molecular motion, such as electronic, vibrational, and rotational motion, correspond to different energy level structures. Knowledge of these energy levels is crucial in molecular physics and quantum chemistry, as they determine the spectroscopic characteristics, chemical reactivity, and many other important molecular properties. Fortunately, experimental measurements of molecular energy spectra can reflect these structures. Meanwhile, there are many molecular spectra data obtained through experimental measurements or simulations (Zou et al., 2023; Alberts et al., 2024). Therefore, incorporating the knowledge of energy levels into molecular representation learning is expected to facilitate the development of more informative molecular representations.",
|
| 28 |
+
"edge": [
|
| 29 |
+
"Limitations of Classical Mechanics",
|
| 30 |
+
"Quantum Perspective and Spectra Data"
|
| 31 |
+
],
|
| 32 |
+
"level": 2,
|
| 33 |
+
"visual_node": []
|
| 34 |
+
},
|
| 35 |
+
{
|
| 36 |
+
"name": "Proposed Method: MolSpectra",
|
| 37 |
+
"content": " Figure 1: The conceptual view of MolSpectra, which leverages both molecular conformation and spectra for pre-training. Prior works only model classical mechanics by denoising on conformations. In this paper, we propose MolSpectra, a framework that incorporates molecular spectra into the pre-training of 3D molecular representations, thereby infusing the knowledge of quantized energy level structures into the representations, as shown in Figure 1. In MolSpectra, we introduce a multispectrum encoder, SpecFormer, to capture both intra-spectrum and inter-spectrum peak correlations by training with a masked patches reconstruction (MPR) objective. Additionally, we employ a contrastive objective to distills the spectral features and its inherent knowledge into the learning of 3D representations. After pre-training, the resulting 3D encoder can be fine-tuned for downstream tasks, providing expressive 3D molecular representations without the need for associated spectral data. Extensive experiments over different downstream molecular property prediction benchmarks shows the superiority of MolSpectra.",
|
| 38 |
+
"edge": [
|
| 39 |
+
"Conceptual View of MolSpectra",
|
| 40 |
+
"MolSpectra Framework Details"
|
| 41 |
+
],
|
| 42 |
+
"level": 2,
|
| 43 |
+
"visual_node": []
|
| 44 |
+
},
|
| 45 |
+
{
|
| 46 |
+
"name": "Contributions",
|
| 47 |
+
"content": "In summary, our contributions are as follows: • We introduce quantized energy level structures and molecular spectra into 3D molecular representation pre-training for the first time, surpassing previous work that relied solely on physical knowledge within the scope of classical mechanics. • We propose SpecFormer as an expressive multi-spectrum encoder, along with the masked patches reconstruction objective for spectral representation learning. • We propose a contrastive objective to align molecular representations in the 3D modality and spectral modalities, enabling the pre-trained 3D encoder to infer molecular spectral features in downstream tasks without relying on spectral data. • Experiments across different downstream benchmarks demonstrate that our method effectively enhances the expressiveness of the pre-trained 3D molecular representations.",
|
| 48 |
+
"edge": [
|
| 49 |
+
"Summary Statement",
|
| 50 |
+
"List of Contributions"
|
| 51 |
+
],
|
| 52 |
+
"level": 2,
|
| 53 |
+
"visual_node": []
|
| 54 |
+
},
|
| 55 |
+
{
|
| 56 |
+
"name": "Importance of 3D Molecular Representations",
|
| 57 |
+
"content": "Learning 3D molecular representations from geometric conformations offers a promising approach for understanding molecular geometry and predicting quantum properties and interactions, which is significant in drug discovery and materials science (Musaelian et al., 2023; Batatia et al., 2022; Liao & Smidt, 2023; Wang et al., 2023b; Du et al., 2023b). Given the scarcity of molecular property labels, self-supervised representation pre-training has been proposed and utilized to provide generalizable representations (Hu et al., 2020; Rong et al., 2020; Ma et al., 2024).",
|
| 58 |
+
"edge": [],
|
| 59 |
+
"level": 3,
|
| 60 |
+
"visual_node": []
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"name": "Physical Principles in 3D Pre-training",
|
| 64 |
+
"content": "In contrast to contrastive learning (Wang et al., 2022; Kim et al., 2022) and masked modeling (Hou et al., 2022; Liu et al., 2023c; Wang et al., 2024b) on 2D molecular graphs and molecular languages (e.g., SMILES), the design of pre-training strategies on 3D molecular geometries is more closely aligned with physical principles. Previous studies (Zaidi et al., 2023; Jiao et al., 2023) have guided representation learning through denoising processes on 3D molecular geometries, theoretically demonstrating that denoising 3D geometries is equivalent to learning molecular force fields, specifically the negative gradient of molecular potential energy with respect to position. Essentially, these studies reveal that establishing the relationship between 3D geometries and the energy states of molecular systems is an effective pathway to learn 3D molecular representations.",
|
| 65 |
+
"edge": [],
|
| 66 |
+
"level": 3,
|
| 67 |
+
"visual_node": []
|
| 68 |
+
},
|
| 69 |
+
{
|
| 70 |
+
"name": "Limitations of Classical Mechanics",
|
| 71 |
+
"content": "However, existing methods are limited to the continuous description (i.e., the potential energy function) of the molecular energy states within the classical mechanics, overlooking the quantized (discrete) energy level structures from the quantum mechanical perspective. From the quantum perspective, molecular systems exhibit quantized energy level structures, meaning that energy states can only assume specific discrete values.",
|
| 72 |
+
"edge": [],
|
| 73 |
+
"level": 3,
|
| 74 |
+
"visual_node": []
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"name": "Quantum Perspective and Spectra Data",
|
| 78 |
+
"content": "Specifically, different types of molecular motion, such as electronic, vibrational, and rotational motion, correspond to different energy level structures. Knowledge of these energy levels is crucial in molecular physics and quantum chemistry, as they determine the spectroscopic characteristics, chemical reactivity, and many other important molecular properties. Fortunately, experimental measurements of molecular energy spectra can reflect these structures. Meanwhile, there are many molecular spectra data obtained through experimental measurements or simulations (Zou et al., 2023; Alberts et al., 2024). Therefore, incorporating the knowledge of energy levels into molecular representation learning is expected to facilitate the development of more informative molecular representations.",
|
| 79 |
+
"edge": [],
|
| 80 |
+
"level": 3,
|
| 81 |
+
"visual_node": []
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"name": "Conceptual View of MolSpectra",
|
| 85 |
+
"content": " Figure 1: The conceptual view of MolSpectra, which leverages both molecular conformation and spectra for pre-training. Prior works only model classical mechanics by denoising on conformations.",
|
| 86 |
+
"edge": [],
|
| 87 |
+
"level": 3,
|
| 88 |
+
"visual_node": []
|
| 89 |
+
},
|
| 90 |
+
{
|
| 91 |
+
"name": "MolSpectra Framework Details",
|
| 92 |
+
"content": "In this paper, we propose MolSpectra, a framework that incorporates molecular spectra into the pre-training of 3D molecular representations, thereby infusing the knowledge of quantized energy level structures into the representations, as shown in Figure 1. In MolSpectra, we introduce a multispectrum encoder, SpecFormer, to capture both intra-spectrum and inter-spectrum peak correlations by training with a masked patches reconstruction (MPR) objective. Additionally, we employ a contrastive objective to distills the spectral features and its inherent knowledge into the learning of 3D representations. After pre-training, the resulting 3D encoder can be fine-tuned for downstream tasks, providing expressive 3D molecular representations without the need for associated spectral data. Extensive experiments over different downstream molecular property prediction benchmarks shows the superiority of MolSpectra.",
|
| 93 |
+
"edge": [
|
| 94 |
+
"SpecFormer and MPR Objective",
|
| 95 |
+
"Contrastive Objective and Fine-tuning"
|
| 96 |
+
],
|
| 97 |
+
"level": 3,
|
| 98 |
+
"visual_node": []
|
| 99 |
+
},
|
| 100 |
+
{
|
| 101 |
+
"name": "Summary Statement",
|
| 102 |
+
"content": "In summary, our contributions are as follows:",
|
| 103 |
+
"edge": [],
|
| 104 |
+
"level": 3,
|
| 105 |
+
"visual_node": []
|
| 106 |
+
},
|
| 107 |
+
{
|
| 108 |
+
"name": "List of Contributions",
|
| 109 |
+
"content": "• We introduce quantized energy level structures and molecular spectra into 3D molecular representation pre-training for the first time, surpassing previous work that relied solely on physical knowledge within the scope of classical mechanics. • We propose SpecFormer as an expressive multi-spectrum encoder, along with the masked patches reconstruction objective for spectral representation learning. • We propose a contrastive objective to align molecular representations in the 3D modality and spectral modalities, enabling the pre-trained 3D encoder to infer molecular spectral features in downstream tasks without relying on spectral data. • Experiments across different downstream benchmarks demonstrate that our method effectively enhances the expressiveness of the pre-trained 3D molecular representations.",
|
| 110 |
+
"edge": [
|
| 111 |
+
"Contributions 1 and 2",
|
| 112 |
+
"Contributions 3 and 4"
|
| 113 |
+
],
|
| 114 |
+
"level": 3,
|
| 115 |
+
"visual_node": []
|
| 116 |
+
},
|
| 117 |
+
{
|
| 118 |
+
"name": "SpecFormer and MPR Objective",
|
| 119 |
+
"content": "In this paper, we propose MolSpectra, a framework that incorporates molecular spectra into the pre-training of 3D molecular representations, thereby infusing the knowledge of quantized energy level structures into the representations, as shown in Figure 1. In MolSpectra, we introduce a multispectrum encoder, SpecFormer, to capture both intra-spectrum and inter-spectrum peak correlations by training with a masked patches reconstruction (MPR) objective.",
|
| 120 |
+
"edge": [],
|
| 121 |
+
"level": 4,
|
| 122 |
+
"visual_node": []
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"name": "Contrastive Objective and Fine-tuning",
|
| 126 |
+
"content": "Additionally, we employ a contrastive objective to distills the spectral features and its inherent knowledge into the learning of 3D representations. After pre-training, the resulting 3D encoder can be fine-tuned for downstream tasks, providing expressive 3D molecular representations without the need for associated spectral data. Extensive experiments over different downstream molecular property prediction benchmarks shows the superiority of MolSpectra.",
|
| 127 |
+
"edge": [],
|
| 128 |
+
"level": 4,
|
| 129 |
+
"visual_node": []
|
| 130 |
+
},
|
| 131 |
+
{
|
| 132 |
+
"name": "Contributions 1 and 2",
|
| 133 |
+
"content": "• We introduce quantized energy level structures and molecular spectra into 3D molecular representation pre-training for the first time, surpassing previous work that relied solely on physical knowledge within the scope of classical mechanics. • We propose SpecFormer as an expressive multi-spectrum encoder, along with the masked patches reconstruction objective for spectral representation learning.",
|
| 134 |
+
"edge": [],
|
| 135 |
+
"level": 4,
|
| 136 |
+
"visual_node": []
|
| 137 |
+
},
|
| 138 |
+
{
|
| 139 |
+
"name": "Contributions 3 and 4",
|
| 140 |
+
"content": "• We propose a contrastive objective to align molecular representations in the 3D modality and spectral modalities, enabling the pre-trained 3D encoder to infer molecular spectral features in downstream tasks without relying on spectral data. • Experiments across different downstream benchmarks demonstrate that our method effectively enhances the expressiveness of the pre-trained 3D molecular representations.",
|
| 141 |
+
"edge": [],
|
| 142 |
+
"level": 4,
|
| 143 |
+
"visual_node": []
|
| 144 |
+
}
|
| 145 |
+
]
|
| 146 |
+
}
|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/2 PRELIMINARIES.md_dag.json
ADDED
|
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"nodes": [
|
| 3 |
+
{
|
| 4 |
+
"name": "2 PRELIMINARIES.md",
|
| 5 |
+
"content": "# 2 PRELIMINARIES # 2.1 NOTATIONS Consider a molecule characterized by its 3D structure and spectra, represented as M = (a, x, S). Here, a ∈ {1, 2, . . . , 118}N specifies the atomic numbers, indicating the types of atoms within the molecule. The vector x ∈ R3N describes the conformation of the molecule, while S represents its spectra. The parameter N denotes the number of atoms in the molecule. Note that the atoms are arranged in the same order in both a and x, ensuring consistency between the atomic numbers and their corresponding spatial coordinates. S = (s1, . . . , s|S|) represents the set of spectra for a molecule, where |S| denotes the number of spectrum types considered. In our study, we focus on three types, so |S| = 3. The first spectrum, s1 ∈ R601, is the UV-Vis spectrum, which spans from 1.5 to 13.5 eV with 601 data points at intervals of 0.02 eV. The second spectrum, s2 ∈ R3501, is the IR spectrum, covering a range from 500 to 4000 cm−1 with 3501 data points at intervals of 1 cm−1. The third spectrum, s3 ∈ R3501, is the Raman spectrum, with the same range and intervals as the IR spectrum. Together, these spectra provide a comprehensive description of the molecular characteristics across different spectral modalities. # 2.2 PRE-TRAINING 3D MOLECULAR REPRESENTATION VIA DENOISING Denoising has emerged as a prominent pre-training objective in 3D molecular representation learning, excelling in various downstream tasks. This method involves training models to predict and remove noise introduced deliberately into molecular structures. This approach is physically interpretable due to its proven equivalence to learning the molecular force field. Equivalence between denoising and learning molecular force fields. The equivalence between coordinate denoising and force field learning is established by Zaidi et al. (2023). For a given molecule M, perturb its equilibrium structure x0 according to the distribution p(x|x0), where x is the noisy conformation. Assuming the molecular distribution adheres to the energy-based Boltzmann distribution with respect to the energy function E(·), then  where GNNθ(x) denotes a graph neural network parameterized by θ, which processes the conformation x to produce node-level predictions. The notation ≃ signifies the equivalence of different objectives. The proof of this equivalence is provided in the Appendix A. In prior research, the energy function E(·) has been defined in several forms. Below are three representative studies. Energy function I: mixture of isotropic Gaussians. In Coord (Zaidi et al., 2023), the energy function is approximated using a mixture of isotropic Gaussians centered at the known equilibrium structures to replace the Boltzmann distribution, since these structures are local maxima of the Boltzman distribution. Leveraging the equivalence between the score-matching objective and denoising autoencoders (Vincent, 2011), the following denoising-based energy function ECoord(·) is derived:  Note that this objective is derived under the assumption of isotropic Gaussian noise, i.e., p(x|x0) ∼ N (x0, τ 2c I3N ), where I3N represents the identity matrix of size 3N , and the subscript c indicates the coordinate denoising approach. Energy function II: mixture of anisotropic Gaussians. Considering rigid and flexible components in molecular structures, isotropic Gaussian can lead to significant approximation errors. To address the anisotropic distribution, Frad (Feng et al., 2023) introduces hybrid noise on dihedral angles of rotatable bonds and atomic coordinates, incorporating fractional denoising of the coordinate noise. The equilibrium structure x0 is initially perturbed by dihedral angle noise p(ψa|ψ0) ∼ N (ψ0, σ2f Im), followed by coordinate noise p(x|xa) ∼ N (xa, τ 2f I3N ). Here, ψa, ψ0 ∈ [0, 2π)m represent to the dihedral angles of rotatable bonds in structures xa and x0, respectively, with m denoting the number of rotatable bonds. The subscript f indicates the fractional denoising approach. Subsequently, the energy function is induced:  where Στf ,σf = τ 2f I3N + σ2f CC⊤, and C ∈ R3N×m is a matrix used to linearly transform the dihedral angle noise into coordinate change, expressed as ∆x ≈ C∆ψ. Energy function III: classical potential energy theory. SliDe (Ni et al., 2024) derives energy function from classical molecular potential energy theory (Alavi, 2020; Zhou & Liu, 2022). In this  Figure 2: Overview of the MolSpectra pre-training framework. Our pre-training framework comprises three sub-objectives: the denoising objective and the MPR objective, which respectively guide the representation learning of the 3D and spectral modalities, and the contrastive objective, which aligns the representations of both modalities. form, the total intramolecular potential energy is mainly attributed to three types of interactions: bond stretching, bond angle bending, and bond torsion. The following energy function is derived:  where r ∈ (R≥0)m1 , θ ∈ [0, 2π)m2 , ϕ ∈ [0, 2π)m3 represent vectors of the bond lengths, bond angles, and bond torsion angles of the molecule, respectively. r0, θ0, ϕ0 correspond to the respective equilibrium values. The parameter vectors kB, kA, kT determine the interaction strength.",
|
| 6 |
+
"edge": [
|
| 7 |
+
"2.1 Notations",
|
| 8 |
+
"2.2 Pre-training 3D Molecular Representation via Denoising"
|
| 9 |
+
],
|
| 10 |
+
"level": 1,
|
| 11 |
+
"visual_node": []
|
| 12 |
+
},
|
| 13 |
+
{
|
| 14 |
+
"name": "2.1 Notations",
|
| 15 |
+
"content": "# 2.1 NOTATIONS Consider a molecule characterized by its 3D structure and spectra, represented as M = (a, x, S). Here, a ∈ {1, 2, . . . , 118}N specifies the atomic numbers, indicating the types of atoms within the molecule. The vector x ∈ R3N describes the conformation of the molecule, while S represents its spectra. The parameter N denotes the number of atoms in the molecule. Note that the atoms are arranged in the same order in both a and x, ensuring consistency between the atomic numbers and their corresponding spatial coordinates. S = (s1, . . . , s|S|) represents the set of spectra for a molecule, where |S| denotes the number of spectrum types considered. In our study, we focus on three types, so |S| = 3. The first spectrum, s1 ∈ R601, is the UV-Vis spectrum, which spans from 1.5 to 13.5 eV with 601 data points at intervals of 0.02 eV. The second spectrum, s2 ∈ R3501, is the IR spectrum, covering a range from 500 to 4000 cm−1 with 3501 data points at intervals of 1 cm−1. The third spectrum, s3 ∈ R3501, is the Raman spectrum, with the same range and intervals as the IR spectrum. Together, these spectra provide a comprehensive description of the molecular characteristics across different spectral modalities.",
|
| 16 |
+
"edge": [
|
| 17 |
+
"Molecule Structure Definition",
|
| 18 |
+
"Spectra Definition"
|
| 19 |
+
],
|
| 20 |
+
"level": 2,
|
| 21 |
+
"visual_node": []
|
| 22 |
+
},
|
| 23 |
+
{
|
| 24 |
+
"name": "2.2 Pre-training 3D Molecular Representation via Denoising",
|
| 25 |
+
"content": "# 2.2 PRE-TRAINING 3D MOLECULAR REPRESENTATION VIA DENOISING Denoising has emerged as a prominent pre-training objective in 3D molecular representation learning, excelling in various downstream tasks. This method involves training models to predict and remove noise introduced deliberately into molecular structures. This approach is physically interpretable due to its proven equivalence to learning the molecular force field. Equivalence between denoising and learning molecular force fields. The equivalence between coordinate denoising and force field learning is established by Zaidi et al. (2023). For a given molecule M, perturb its equilibrium structure x0 according to the distribution p(x|x0), where x is the noisy conformation. Assuming the molecular distribution adheres to the energy-based Boltzmann distribution with respect to the energy function E(·), then  where GNNθ(x) denotes a graph neural network parameterized by θ, which processes the conformation x to produce node-level predictions. The notation ≃ signifies the equivalence of different objectives. The proof of this equivalence is provided in the Appendix A. In prior research, the energy function E(·) has been defined in several forms. Below are three representative studies. Energy function I: mixture of isotropic Gaussians. In Coord (Zaidi et al., 2023), the energy function is approximated using a mixture of isotropic Gaussians centered at the known equilibrium structures to replace the Boltzmann distribution, since these structures are local maxima of the Boltzman distribution. Leveraging the equivalence between the score-matching objective and denoising autoencoders (Vincent, 2011), the following denoising-based energy function ECoord(·) is derived:  Note that this objective is derived under the assumption of isotropic Gaussian noise, i.e., p(x|x0) ∼ N (x0, τ 2c I3N ), where I3N represents the identity matrix of size 3N , and the subscript c indicates the coordinate denoising approach. Energy function II: mixture of anisotropic Gaussians. Considering rigid and flexible components in molecular structures, isotropic Gaussian can lead to significant approximation errors. To address the anisotropic distribution, Frad (Feng et al., 2023) introduces hybrid noise on dihedral angles of rotatable bonds and atomic coordinates, incorporating fractional denoising of the coordinate noise. The equilibrium structure x0 is initially perturbed by dihedral angle noise p(ψa|ψ0) ∼ N (ψ0, σ2f Im), followed by coordinate noise p(x|xa) ∼ N (xa, τ 2f I3N ). Here, ψa, ψ0 ∈ [0, 2π)m represent to the dihedral angles of rotatable bonds in structures xa and x0, respectively, with m denoting the number of rotatable bonds. The subscript f indicates the fractional denoising approach. Subsequently, the energy function is induced:  where Στf ,σf = τ 2f I3N + σ2f CC⊤, and C ∈ R3N×m is a matrix used to linearly transform the dihedral angle noise into coordinate change, expressed as ∆x ≈ C∆ψ. Energy function III: classical potential energy theory. SliDe (Ni et al., 2024) derives energy function from classical molecular potential energy theory (Alavi, 2020; Zhou & Liu, 2022). In this  Figure 2: Overview of the MolSpectra pre-training framework. Our pre-training framework comprises three sub-objectives: the denoising objective and the MPR objective, which respectively guide the representation learning of the 3D and spectral modalities, and the contrastive objective, which aligns the representations of both modalities. form, the total intramolecular potential energy is mainly attributed to three types of interactions: bond stretching, bond angle bending, and bond torsion. The following energy function is derived:  where r ∈ (R≥0)m1 , θ ∈ [0, 2π)m2 , ϕ ∈ [0, 2π)m3 represent vectors of the bond lengths, bond angles, and bond torsion angles of the molecule, respectively. r0, θ0, ϕ0 correspond to the respective equilibrium values. The parameter vectors kB, kA, kT determine the interaction strength.",
|
| 26 |
+
"edge": [
|
| 27 |
+
"Denoising Fundamentals and Equivalence",
|
| 28 |
+
"Representative Energy Functions"
|
| 29 |
+
],
|
| 30 |
+
"level": 2,
|
| 31 |
+
"visual_node": []
|
| 32 |
+
},
|
| 33 |
+
{
|
| 34 |
+
"name": "Molecule Structure Definition",
|
| 35 |
+
"content": "Consider a molecule characterized by its 3D structure and spectra, represented as M = (a, x, S). Here, a ∈ {1, 2, . . . , 118}N specifies the atomic numbers, indicating the types of atoms within the molecule. The vector x ∈ R3N describes the conformation of the molecule, while S represents its spectra. The parameter N denotes the number of atoms in the molecule. Note that the atoms are arranged in the same order in both a and x, ensuring consistency between the atomic numbers and their corresponding spatial coordinates.",
|
| 36 |
+
"edge": [],
|
| 37 |
+
"level": 3,
|
| 38 |
+
"visual_node": []
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"name": "Spectra Definition",
|
| 42 |
+
"content": "S = (s1, . . . , s|S|) represents the set of spectra for a molecule, where |S| denotes the number of spectrum types considered. In our study, we focus on three types, so |S| = 3. The first spectrum, s1 ∈ R601, is the UV-Vis spectrum, which spans from 1.5 to 13.5 eV with 601 data points at intervals of 0.02 eV. The second spectrum, s2 ∈ R3501, is the IR spectrum, covering a range from 500 to 4000 cm−1 with 3501 data points at intervals of 1 cm−1. The third spectrum, s3 ∈ R3501, is the Raman spectrum, with the same range and intervals as the IR spectrum. Together, these spectra provide a comprehensive description of the molecular characteristics across different spectral modalities.",
|
| 43 |
+
"edge": [],
|
| 44 |
+
"level": 3,
|
| 45 |
+
"visual_node": []
|
| 46 |
+
},
|
| 47 |
+
{
|
| 48 |
+
"name": "Denoising Fundamentals and Equivalence",
|
| 49 |
+
"content": "Denoising has emerged as a prominent pre-training objective in 3D molecular representation learning, excelling in various downstream tasks. This method involves training models to predict and remove noise introduced deliberately into molecular structures. This approach is physically interpretable due to its proven equivalence to learning the molecular force field. Equivalence between denoising and learning molecular force fields. The equivalence between coordinate denoising and force field learning is established by Zaidi et al. (2023). For a given molecule M, perturb its equilibrium structure x0 according to the distribution p(x|x0), where x is the noisy conformation. Assuming the molecular distribution adheres to the energy-based Boltzmann distribution with respect to the energy function E(·), then  where GNNθ(x) denotes a graph neural network parameterized by θ, which processes the conformation x to produce node-level predictions. The notation ≃ signifies the equivalence of different objectives. The proof of this equivalence is provided in the Appendix A.",
|
| 50 |
+
"edge": [
|
| 51 |
+
"Introduction to Denoising",
|
| 52 |
+
"Mathematical Equivalence"
|
| 53 |
+
],
|
| 54 |
+
"level": 3,
|
| 55 |
+
"visual_node": []
|
| 56 |
+
},
|
| 57 |
+
{
|
| 58 |
+
"name": "Representative Energy Functions",
|
| 59 |
+
"content": "In prior research, the energy function E(·) has been defined in several forms. Below are three representative studies. Energy function I: mixture of isotropic Gaussians. In Coord (Zaidi et al., 2023), the energy function is approximated using a mixture of isotropic Gaussians centered at the known equilibrium structures to replace the Boltzmann distribution, since these structures are local maxima of the Boltzman distribution. Leveraging the equivalence between the score-matching objective and denoising autoencoders (Vincent, 2011), the following denoising-based energy function ECoord(·) is derived:  Note that this objective is derived under the assumption of isotropic Gaussian noise, i.e., p(x|x0) ∼ N (x0, τ 2c I3N ), where I3N represents the identity matrix of size 3N , and the subscript c indicates the coordinate denoising approach. Energy function II: mixture of anisotropic Gaussians. Considering rigid and flexible components in molecular structures, isotropic Gaussian can lead to significant approximation errors. To address the anisotropic distribution, Frad (Feng et al., 2023) introduces hybrid noise on dihedral angles of rotatable bonds and atomic coordinates, incorporating fractional denoising of the coordinate noise. The equilibrium structure x0 is initially perturbed by dihedral angle noise p(ψa|ψ0) ∼ N (ψ0, σ2f Im), followed by coordinate noise p(x|xa) ∼ N (xa, τ 2f I3N ). Here, ψa, ψ0 ∈ [0, 2π)m represent to the dihedral angles of rotatable bonds in structures xa and x0, respectively, with m denoting the number of rotatable bonds. The subscript f indicates the fractional denoising approach. Subsequently, the energy function is induced:  where Στf ,σf = τ 2f I3N + σ2f CC⊤, and C ∈ R3N×m is a matrix used to linearly transform the dihedral angle noise into coordinate change, expressed as ∆x ≈ C∆ψ. Energy function III: classical potential energy theory. SliDe (Ni et al., 2024) derives energy function from classical molecular potential energy theory (Alavi, 2020; Zhou & Liu, 2022). In this  Figure 2: Overview of the MolSpectra pre-training framework. Our pre-training framework comprises three sub-objectives: the denoising objective and the MPR objective, which respectively guide the representation learning of the 3D and spectral modalities, and the contrastive objective, which aligns the representations of both modalities. form, the total intramolecular potential energy is mainly attributed to three types of interactions: bond stretching, bond angle bending, and bond torsion. The following energy function is derived:  where r ∈ (R≥0)m1 , θ ∈ [0, 2π)m2 , ϕ ∈ [0, 2π)m3 represent vectors of the bond lengths, bond angles, and bond torsion angles of the molecule, respectively. r0, θ0, ϕ0 correspond to the respective equilibrium values. The parameter vectors kB, kA, kT determine the interaction strength.",
|
| 60 |
+
"edge": [
|
| 61 |
+
"Energy Function I: Isotropic Gaussians",
|
| 62 |
+
"Energy Function II: Anisotropic Gaussians",
|
| 63 |
+
"Energy Function III: Classical Potential"
|
| 64 |
+
],
|
| 65 |
+
"level": 3,
|
| 66 |
+
"visual_node": []
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"name": "Introduction to Denoising",
|
| 70 |
+
"content": "Denoising has emerged as a prominent pre-training objective in 3D molecular representation learning, excelling in various downstream tasks. This method involves training models to predict and remove noise introduced deliberately into molecular structures. This approach is physically interpretable due to its proven equivalence to learning the molecular force field.",
|
| 71 |
+
"edge": [],
|
| 72 |
+
"level": 4,
|
| 73 |
+
"visual_node": []
|
| 74 |
+
},
|
| 75 |
+
{
|
| 76 |
+
"name": "Mathematical Equivalence",
|
| 77 |
+
"content": "Equivalence between denoising and learning molecular force fields. The equivalence between coordinate denoising and force field learning is established by Zaidi et al. (2023). For a given molecule M, perturb its equilibrium structure x0 according to the distribution p(x|x0), where x is the noisy conformation. Assuming the molecular distribution adheres to the energy-based Boltzmann distribution with respect to the energy function E(·), then  where GNNθ(x) denotes a graph neural network parameterized by θ, which processes the conformation x to produce node-level predictions. The notation ≃ signifies the equivalence of different objectives. The proof of this equivalence is provided in the Appendix A.",
|
| 78 |
+
"edge": [],
|
| 79 |
+
"level": 4,
|
| 80 |
+
"visual_node": []
|
| 81 |
+
},
|
| 82 |
+
{
|
| 83 |
+
"name": "Energy Function I: Isotropic Gaussians",
|
| 84 |
+
"content": "In prior research, the energy function E(·) has been defined in several forms. Below are three representative studies. Energy function I: mixture of isotropic Gaussians. In Coord (Zaidi et al., 2023), the energy function is approximated using a mixture of isotropic Gaussians centered at the known equilibrium structures to replace the Boltzmann distribution, since these structures are local maxima of the Boltzman distribution. Leveraging the equivalence between the score-matching objective and denoising autoencoders (Vincent, 2011), the following denoising-based energy function ECoord(·) is derived:  Note that this objective is derived under the assumption of isotropic Gaussian noise, i.e., p(x|x0) ∼ N (x0, τ 2c I3N ), where I3N represents the identity matrix of size 3N , and the subscript c indicates the coordinate denoising approach.",
|
| 85 |
+
"edge": [],
|
| 86 |
+
"level": 4,
|
| 87 |
+
"visual_node": []
|
| 88 |
+
},
|
| 89 |
+
{
|
| 90 |
+
"name": "Energy Function II: Anisotropic Gaussians",
|
| 91 |
+
"content": "Energy function II: mixture of anisotropic Gaussians. Considering rigid and flexible components in molecular structures, isotropic Gaussian can lead to significant approximation errors. To address the anisotropic distribution, Frad (Feng et al., 2023) introduces hybrid noise on dihedral angles of rotatable bonds and atomic coordinates, incorporating fractional denoising of the coordinate noise. The equilibrium structure x0 is initially perturbed by dihedral angle noise p(ψa|ψ0) ∼ N (ψ0, σ2f Im), followed by coordinate noise p(x|xa) ∼ N (xa, τ 2f I3N ). Here, ψa, ψ0 ∈ [0, 2π)m represent to the dihedral angles of rotatable bonds in structures xa and x0, respectively, with m denoting the number of rotatable bonds. The subscript f indicates the fractional denoising approach. Subsequently, the energy function is induced:  where Στf ,σf = τ 2f I3N + σ2f CC⊤, and C ∈ R3N×m is a matrix used to linearly transform the dihedral angle noise into coordinate change, expressed as ∆x ≈ C∆ψ.",
|
| 92 |
+
"edge": [],
|
| 93 |
+
"level": 4,
|
| 94 |
+
"visual_node": []
|
| 95 |
+
},
|
| 96 |
+
{
|
| 97 |
+
"name": "Energy Function III: Classical Potential",
|
| 98 |
+
"content": "Energy function III: classical potential energy theory. SliDe (Ni et al., 2024) derives energy function from classical molecular potential energy theory (Alavi, 2020; Zhou & Liu, 2022). In this  Figure 2: Overview of the MolSpectra pre-training framework. Our pre-training framework comprises three sub-objectives: the denoising objective and the MPR objective, which respectively guide the representation learning of the 3D and spectral modalities, and the contrastive objective, which aligns the representations of both modalities. form, the total intramolecular potential energy is mainly attributed to three types of interactions: bond stretching, bond angle bending, and bond torsion. The following energy function is derived:  where r ∈ (R≥0)m1 , θ ∈ [0, 2π)m2 , ϕ ∈ [0, 2π)m3 represent vectors of the bond lengths, bond angles, and bond torsion angles of the molecule, respectively. r0, θ0, ϕ0 correspond to the respective equilibrium values. The parameter vectors kB, kA, kT determine the interaction strength.",
|
| 99 |
+
"edge": [],
|
| 100 |
+
"level": 4,
|
| 101 |
+
"visual_node": []
|
| 102 |
+
}
|
| 103 |
+
]
|
| 104 |
+
}
|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/3 THE PROPOSED MOLSPECTRA METHOD.md_dag.json
ADDED
|
@@ -0,0 +1,187 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"nodes": [
|
| 3 |
+
{
|
| 4 |
+
"name": "3 THE PROPOSED MOLSPECTRA METHOD",
|
| 5 |
+
"content": "# 3 THE PROPOSED MOLSPECTRA METHOD Considering the complementarity of different spectra, we introduce multiple spectra into molecular representation learning. To effectively comprehend molecular spectra, we designed a Transformerbased multi-spectrum encoder, SpecFormer, along with a masked reconstruction objective to guide its training. Finally, a contrastive objective is employed to align the 3D encoding guided by the denoising objective with the spectra encoding guided by the reconstruction objective, endowing the 3D encoding with the capability to understand spectra and the knowledge they encompass. # 3.1 SPECFORMER: A SINGLE-STREAM ENCODER FOR MULTI-MODAL ENERGY SPECTRA For different types of spectra, each spectrum is independently patched and initially encoded. Then, all the resulting patch embeddings are concatenated and encoded using a Transformer-based encoder. Patching. Compared to directly encoding individual frequency points, we divided each spectrum into multiple patches. This approach offers two distinct advantages: (i) By forming patches from adjacent frequency points, local semantic features, such as absorption peaks, can be captured more effectively. (ii) It reduces the computational overhead of subsequent Transformer layers. Technically, each spectrum si ∈ RLi where i = 1, · · · , |S| is first divided into patches according to the patch length Pi and the stride Di. When 0 < Di < Pi, the consecutive patches will be overlapped with overlapping region length Pi − Di. When Di = Pi, the consecutive patches will be nonoverlapped. Li denotes the length of si. The patching process on each spectrum will generate a sequence of patches pi ∈ RNi×Pi, where Ni = Li−PiDi + 1 is the number of patches. Patch encoding and position encoding. Prior to be fed into the encoder, the patches of the i-th spectrum are mapped to the latent space of dimension d via a trainable linear projection Wi ∈ RPi×d. A learnable additive position encoding W posi ∈ RNi×d is applied to maintain the order of the patches: p′i = piWi + W posi , where p′i ∈ RNi×d denotes the latent representation of the spectrum si that will be fed into the subsequent SpecFormer encoder. SpecFormer: multi-spectrum Transformer encoder. Although several encoders have been proposed to map molecular spectrum into implicit representations, such as the CNN-AM (Tao et al., 2024) based on one-dimensional convolution, these encoders are designed to encode only a single type of spectrum. In our approach, multiple molecular spectra (UV-Vis, IR, Raman) are jointly considered. When encoding multiple spectra of a molecule simultaneously, an observation caught our attention and led us to adopt a Transformer-based encoder with multiple spectra as input, similar to the single-stream Transformer in multi-modal learning (Shin et al., 2021). The observation refers to the fact that the same functional group not only causes multiple peaks within a single spectrum, but also generates peaks across different spectra. As shown on the left of Figure 3, the different vibrational modes of the methyl group (-CH3) in methanol (CH3OH) result in three peaks in the IR spectrum, indicating intra-spectrum dependencies among these peaks. A similar phenomenon occurs with the hydroxyl group (-OH) in methanol. Additionally, the aromatic ring in phenol (C6H5OH), shown on the right of  Figure 3: Illustration of intra-spectrum (left) and interspectrum (right) dependencies. Figure 3, not only produces multiple peaks in the IR spectrum due to different vibrational modes but also causes an absorption peak near 270 nm in the UV-Vis spectrum due to the π → π∗ transition in the aromatic ring, demonstrating the existence of inter-spectrum dependencies. Such dependencies have been theoretically studied, for example, in the context of vibronic coupling (Kong et al., 2021). To capture intra-spectrum and inter-spectrum dependencies, we concatenate the embeddings obtained from patch encoding and position encoding of different spectra: pˆ = p′1∥ · · · ∥p′|S| ∈ R(P|S|i=1 Ni)×d, and then input them into the Transformer encoder as depicted in Figure 2. Then each head h = 1, . . . , H in multi-head attention will transform them into query matrices Qh = pWˆ Qh , key matrices Kh = pWˆ Kh and value matrices Vh = pWˆ Vh , where W Qh , W Kh ∈ Rd×dk and WVh ∈ Rd× dH . Afterward, a scaled product is utilized to obtain the attention output Oh ∈ R(P|S|i=1 Ni)× dH :  The multi-head attention block also includes BatchNorm layers and a feed forward network with residual connections as shown in Figure 2. After combining the outputs of all heads, it generates the representation denoted as z ∈ R(P|S|i=1 Ni)×d. Finally, a flatten layer with representation projection head is used to obtain the molecular spectra representation zs ∈ Rd. # 3.2 MASKED PATCHES RECONSTRUCTION PRE-TRAINING FOR SPECTRA Before distilling the spectra information into 3D molecular representation learning, we need first ensure that the spectrum encoder can effectively comprehend molecular spectra and generate spectral representations. Considering the success of masking modeling across various domains (Devlin et al., 2019; He et al., 2022; Hou et al., 2022; Xia et al., 2023; Wang et al., 2024b; Nie et al., 2023), we propose a masked patches reconstruction (MPR) objective to guide the training of SpecFormer. After the patching step, we randomly select a portion of patches according to the mask ratio α and replace them with zero vectors to implement the masking. Subsequently, the masked patches undergo patch encoding and position encoding. In this way, the semantics of the masked patches (the absorption intensity at specific wavelengths) are obscured during patch encoding, while the positional information is retained to facilitate the reconstruction of the original semantics. After encoding by SpecFormer, the encoded results corresponding to the masked patches are input into a spectrum-specific reconstruction head to reconstruct the original spectral values that were masked. The mean squared error (MSE) between the reconstruction results and the original masked spectra serves as the loss function for the MPR task, guiding the training of SpecFormer:  where Pi denotes the set of masked patches in the i-th type of molecular spectra, and pˆi,j denotes ethe reconstructed patch corresponding to the masked patch pi,j . # 3.3 CONTRASTIVE LEARNING BETWEEN 3D STRUCTURES AND SPECTRA Under the guidance of the denoising objective for 3D representation learning and the MPR objective for spectral representation learning, we further introduce a contrastive objective to align the representations across these two modalities. We treat the 3D representation zx ∈ Rd and spectral representation zs ∈ Rd of the same molecule as positive samples, and negative samples otherwise. Subsequently, the consistency between positive samples and the discrepancy between negative samples are maximized through the contrastive objective. Given the theoretical and empirical effectiveness, we employ InfoNCE (van den Oord et al., 2018) as the contrastive objective:  where zjx, zjs are randomly sampled 3D and spectra views regarding to the positive pair (zx, zs). fx(zx, zs) and fs(zs, zx) are scoring functions for the two corresponding views, with flexible formulations. Here we adopt fx(zx, zs) = fs(zs, zx) = ⟨zx, zs⟩. Note that the denoising objective can utilize any form from existing 3D molecular representation pre-training studies, enabling seamless integration of our method into these frameworks. # 3.4 TWO-STAGE PRE-TRAINING PIPELINE Previous pre-training efforts for 3D molecular representation have been conducted on unlabeled datasets using denoising objective. These datasets typically provide only equilibrium 3D structures without offering spectra for all molecules. To enhance the pre-training effect by incorporating spectra while leveraging denoising pre-training, we employ a two-stage pre-training approach. The first stage involves training on a larger dataset (Nakata & Shimazaki, 2017) without spectra using only the denoising objective. Subsequently, the second stage involves training on a dataset that includes spectra using the complete objective as follows:  where βDenoising, βMPR, and βContrast denote the weights of each sub-objective.",
|
| 6 |
+
"edge": [
|
| 7 |
+
"Method Overview",
|
| 8 |
+
"SpecFormer Encoder (3.1)",
|
| 9 |
+
"Masked Patches Reconstruction (3.2)",
|
| 10 |
+
"Contrastive Learning (3.3)",
|
| 11 |
+
"Two-Stage Pipeline (3.4)"
|
| 12 |
+
],
|
| 13 |
+
"level": 1,
|
| 14 |
+
"visual_node": []
|
| 15 |
+
},
|
| 16 |
+
{
|
| 17 |
+
"name": "Method Overview",
|
| 18 |
+
"content": "Considering the complementarity of different spectra, we introduce multiple spectra into molecular representation learning. To effectively comprehend molecular spectra, we designed a Transformerbased multi-spectrum encoder, SpecFormer, along with a masked reconstruction objective to guide its training. Finally, a contrastive objective is employed to align the 3D encoding guided by the denoising objective with the spectra encoding guided by the reconstruction objective, endowing the 3D encoding with the capability to understand spectra and the knowledge they encompass.",
|
| 19 |
+
"edge": [],
|
| 20 |
+
"level": 2,
|
| 21 |
+
"visual_node": []
|
| 22 |
+
},
|
| 23 |
+
{
|
| 24 |
+
"name": "SpecFormer Encoder (3.1)",
|
| 25 |
+
"content": "# 3.1 SPECFORMER: A SINGLE-STREAM ENCODER FOR MULTI-MODAL ENERGY SPECTRA For different types of spectra, each spectrum is independently patched and initially encoded. Then, all the resulting patch embeddings are concatenated and encoded using a Transformer-based encoder. Patching. Compared to directly encoding individual frequency points, we divided each spectrum into multiple patches. This approach offers two distinct advantages: (i) By forming patches from adjacent frequency points, local semantic features, such as absorption peaks, can be captured more effectively. (ii) It reduces the computational overhead of subsequent Transformer layers. Technically, each spectrum si ∈ RLi where i = 1, · · · , |S| is first divided into patches according to the patch length Pi and the stride Di. When 0 < Di < Pi, the consecutive patches will be overlapped with overlapping region length Pi − Di. When Di = Pi, the consecutive patches will be nonoverlapped. Li denotes the length of si. The patching process on each spectrum will generate a sequence of patches pi ∈ RNi×Pi, where Ni = Li−PiDi + 1 is the number of patches. Patch encoding and position encoding. Prior to be fed into the encoder, the patches of the i-th spectrum are mapped to the latent space of dimension d via a trainable linear projection Wi ∈ RPi×d. A learnable additive position encoding W posi ∈ RNi×d is applied to maintain the order of the patches: p′i = piWi + W posi , where p′i ∈ RNi×d denotes the latent representation of the spectrum si that will be fed into the subsequent SpecFormer encoder. SpecFormer: multi-spectrum Transformer encoder. Although several encoders have been proposed to map molecular spectrum into implicit representations, such as the CNN-AM (Tao et al., 2024) based on one-dimensional convolution, these encoders are designed to encode only a single type of spectrum. In our approach, multiple molecular spectra (UV-Vis, IR, Raman) are jointly considered. When encoding multiple spectra of a molecule simultaneously, an observation caught our attention and led us to adopt a Transformer-based encoder with multiple spectra as input, similar to the single-stream Transformer in multi-modal learning (Shin et al., 2021). The observation refers to the fact that the same functional group not only causes multiple peaks within a single spectrum, but also generates peaks across different spectra. As shown on the left of Figure 3, the different vibrational modes of the methyl group (-CH3) in methanol (CH3OH) result in three peaks in the IR spectrum, indicating intra-spectrum dependencies among these peaks. A similar phenomenon occurs with the hydroxyl group (-OH) in methanol. Additionally, the aromatic ring in phenol (C6H5OH), shown on the right of  Figure 3: Illustration of intra-spectrum (left) and interspectrum (right) dependencies. Figure 3, not only produces multiple peaks in the IR spectrum due to different vibrational modes but also causes an absorption peak near 270 nm in the UV-Vis spectrum due to the π → π∗ transition in the aromatic ring, demonstrating the existence of inter-spectrum dependencies. Such dependencies have been theoretically studied, for example, in the context of vibronic coupling (Kong et al., 2021). To capture intra-spectrum and inter-spectrum dependencies, we concatenate the embeddings obtained from patch encoding and position encoding of different spectra: pˆ = p′1∥ · · · ∥p′|S| ∈ R(P|S|i=1 Ni)×d, and then input them into the Transformer encoder as depicted in Figure 2. Then each head h = 1, . . . , H in multi-head attention will transform them into query matrices Qh = pWˆ Qh , key matrices Kh = pWˆ Kh and value matrices Vh = pWˆ Vh , where W Qh , W Kh ∈ Rd×dk and WVh ∈ Rd× dH . Afterward, a scaled product is utilized to obtain the attention output Oh ∈ R(P|S|i=1 Ni)× dH :  The multi-head attention block also includes BatchNorm layers and a feed forward network with residual connections as shown in Figure 2. After combining the outputs of all heads, it generates the representation denoted as z ∈ R(P|S|i=1 Ni)×d. Finally, a flatten layer with representation projection head is used to obtain the molecular spectra representation zs ∈ Rd.",
|
| 26 |
+
"edge": [
|
| 27 |
+
"Patching and Initial Encoding",
|
| 28 |
+
"Transformer Architecture and Dependencies"
|
| 29 |
+
],
|
| 30 |
+
"level": 2,
|
| 31 |
+
"visual_node": []
|
| 32 |
+
},
|
| 33 |
+
{
|
| 34 |
+
"name": "Patching and Initial Encoding",
|
| 35 |
+
"content": "For different types of spectra, each spectrum is independently patched and initially encoded. Then, all the resulting patch embeddings are concatenated and encoded using a Transformer-based encoder. Patching. Compared to directly encoding individual frequency points, we divided each spectrum into multiple patches. This approach offers two distinct advantages: (i) By forming patches from adjacent frequency points, local semantic features, such as absorption peaks, can be captured more effectively. (ii) It reduces the computational overhead of subsequent Transformer layers. Technically, each spectrum si ∈ RLi where i = 1, · · · , |S| is first divided into patches according to the patch length Pi and the stride Di. When 0 < Di < Pi, the consecutive patches will be overlapped with overlapping region length Pi − Di. When Di = Pi, the consecutive patches will be nonoverlapped. Li denotes the length of si. The patching process on each spectrum will generate a sequence of patches pi ∈ RNi×Pi, where Ni = Li−PiDi + 1 is the number of patches. Patch encoding and position encoding. Prior to be fed into the encoder, the patches of the i-th spectrum are mapped to the latent space of dimension d via a trainable linear projection Wi ∈ RPi×d. A learnable additive position encoding W posi ∈ RNi×d is applied to maintain the order of the patches: p′i = piWi + W posi , where p′i ∈ RNi×d denotes the latent representation of the spectrum si that will be fed into the subsequent SpecFormer encoder.",
|
| 36 |
+
"edge": [
|
| 37 |
+
"Process Overview",
|
| 38 |
+
"Patching Strategy",
|
| 39 |
+
"Patch and Position Encoding"
|
| 40 |
+
],
|
| 41 |
+
"level": 3,
|
| 42 |
+
"visual_node": []
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"name": "Process Overview",
|
| 46 |
+
"content": "For different types of spectra, each spectrum is independently patched and initially encoded. Then, all the resulting patch embeddings are concatenated and encoded using a Transformer-based encoder.",
|
| 47 |
+
"edge": [],
|
| 48 |
+
"level": 4,
|
| 49 |
+
"visual_node": []
|
| 50 |
+
},
|
| 51 |
+
{
|
| 52 |
+
"name": "Patching Strategy",
|
| 53 |
+
"content": "Patching. Compared to directly encoding individual frequency points, we divided each spectrum into multiple patches. This approach offers two distinct advantages: (i) By forming patches from adjacent frequency points, local semantic features, such as absorption peaks, can be captured more effectively. (ii) It reduces the computational overhead of subsequent Transformer layers. Technically, each spectrum si ∈ RLi where i = 1, · · · , |S| is first divided into patches according to the patch length Pi and the stride Di. When 0 < Di < Pi, the consecutive patches will be overlapped with overlapping region length Pi − Di. When Di = Pi, the consecutive patches will be nonoverlapped. Li denotes the length of si. The patching process on each spectrum will generate a sequence of patches pi ∈ RNi×Pi, where Ni = Li−PiDi + 1 is the number of patches.",
|
| 54 |
+
"edge": [],
|
| 55 |
+
"level": 4,
|
| 56 |
+
"visual_node": []
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"name": "Patch and Position Encoding",
|
| 60 |
+
"content": "Patch encoding and position encoding. Prior to be fed into the encoder, the patches of the i-th spectrum are mapped to the latent space of dimension d via a trainable linear projection Wi ∈ RPi×d. A learnable additive position encoding W posi ∈ RNi×d is applied to maintain the order of the patches: p′i = piWi + W posi , where p′i ∈ RNi×d denotes the latent representation of the spectrum si that will be fed into the subsequent SpecFormer encoder.",
|
| 61 |
+
"edge": [],
|
| 62 |
+
"level": 4,
|
| 63 |
+
"visual_node": []
|
| 64 |
+
},
|
| 65 |
+
{
|
| 66 |
+
"name": "Transformer Architecture and Dependencies",
|
| 67 |
+
"content": "SpecFormer: multi-spectrum Transformer encoder. Although several encoders have been proposed to map molecular spectrum into implicit representations, such as the CNN-AM (Tao et al., 2024) based on one-dimensional convolution, these encoders are designed to encode only a single type of spectrum. In our approach, multiple molecular spectra (UV-Vis, IR, Raman) are jointly considered. When encoding multiple spectra of a molecule simultaneously, an observation caught our attention and led us to adopt a Transformer-based encoder with multiple spectra as input, similar to the single-stream Transformer in multi-modal learning (Shin et al., 2021). The observation refers to the fact that the same functional group not only causes multiple peaks within a single spectrum, but also generates peaks across different spectra. As shown on the left of Figure 3, the different vibrational modes of the methyl group (-CH3) in methanol (CH3OH) result in three peaks in the IR spectrum, indicating intra-spectrum dependencies among these peaks. A similar phenomenon occurs with the hydroxyl group (-OH) in methanol. Additionally, the aromatic ring in phenol (C6H5OH), shown on the right of  Figure 3: Illustration of intra-spectrum (left) and interspectrum (right) dependencies. Figure 3, not only produces multiple peaks in the IR spectrum due to different vibrational modes but also causes an absorption peak near 270 nm in the UV-Vis spectrum due to the π → π∗ transition in the aromatic ring, demonstrating the existence of inter-spectrum dependencies. Such dependencies have been theoretically studied, for example, in the context of vibronic coupling (Kong et al., 2021). To capture intra-spectrum and inter-spectrum dependencies, we concatenate the embeddings obtained from patch encoding and position encoding of different spectra: pˆ = p′1∥ · · · ∥p′|S| ∈ R(P|S|i=1 Ni)×d, and then input them into the Transformer encoder as depicted in Figure 2. Then each head h = 1, . . . , H in multi-head attention will transform them into query matrices Qh = pWˆ Qh , key matrices Kh = pWˆ Kh and value matrices Vh = pWˆ Vh , where W Qh , W Kh ∈ Rd×dk and WVh ∈ Rd× dH . Afterward, a scaled product is utilized to obtain the attention output Oh ∈ R(P|S|i=1 Ni)× dH :  The multi-head attention block also includes BatchNorm layers and a feed forward network with residual connections as shown in Figure 2. After combining the outputs of all heads, it generates the representation denoted as z ∈ R(P|S|i=1 Ni)×d. Finally, a flatten layer with representation projection head is used to obtain the molecular spectra representation zs ∈ Rd.",
|
| 68 |
+
"edge": [
|
| 69 |
+
"Encoder Motivation",
|
| 70 |
+
"Spectral Dependencies Observation",
|
| 71 |
+
"Attention Mechanism and Output"
|
| 72 |
+
],
|
| 73 |
+
"level": 3,
|
| 74 |
+
"visual_node": []
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"name": "Encoder Motivation",
|
| 78 |
+
"content": "SpecFormer: multi-spectrum Transformer encoder. Although several encoders have been proposed to map molecular spectrum into implicit representations, such as the CNN-AM (Tao et al., 2024) based on one-dimensional convolution, these encoders are designed to encode only a single type of spectrum. In our approach, multiple molecular spectra (UV-Vis, IR, Raman) are jointly considered. When encoding multiple spectra of a molecule simultaneously, an observation caught our attention and led us to adopt a Transformer-based encoder with multiple spectra as input, similar to the single-stream Transformer in multi-modal learning (Shin et al., 2021).",
|
| 79 |
+
"edge": [],
|
| 80 |
+
"level": 4,
|
| 81 |
+
"visual_node": []
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"name": "Spectral Dependencies Observation",
|
| 85 |
+
"content": "The observation refers to the fact that the same functional group not only causes multiple peaks within a single spectrum, but also generates peaks across different spectra. As shown on the left of Figure 3, the different vibrational modes of the methyl group (-CH3) in methanol (CH3OH) result in three peaks in the IR spectrum, indicating intra-spectrum dependencies among these peaks. A similar phenomenon occurs with the hydroxyl group (-OH) in methanol. Additionally, the aromatic ring in phenol (C6H5OH), shown on the right of  Figure 3: Illustration of intra-spectrum (left) and interspectrum (right) dependencies. Figure 3, not only produces multiple peaks in the IR spectrum due to different vibrational modes but also causes an absorption peak near 270 nm in the UV-Vis spectrum due to the π → π∗ transition in the aromatic ring, demonstrating the existence of inter-spectrum dependencies. Such dependencies have been theoretically studied, for example, in the context of vibronic coupling (Kong et al., 2021).",
|
| 86 |
+
"edge": [],
|
| 87 |
+
"level": 4,
|
| 88 |
+
"visual_node": []
|
| 89 |
+
},
|
| 90 |
+
{
|
| 91 |
+
"name": "Attention Mechanism and Output",
|
| 92 |
+
"content": "To capture intra-spectrum and inter-spectrum dependencies, we concatenate the embeddings obtained from patch encoding and position encoding of different spectra: pˆ = p′1∥ · · · ∥p′|S| ∈ R(P|S|i=1 Ni)×d, and then input them into the Transformer encoder as depicted in Figure 2. Then each head h = 1, . . . , H in multi-head attention will transform them into query matrices Qh = pWˆ Qh , key matrices Kh = pWˆ Kh and value matrices Vh = pWˆ Vh , where W Qh , W Kh ∈ Rd×dk and WVh ∈ Rd× dH . Afterward, a scaled product is utilized to obtain the attention output Oh ∈ R(P|S|i=1 Ni)× dH :  The multi-head attention block also includes BatchNorm layers and a feed forward network with residual connections as shown in Figure 2. After combining the outputs of all heads, it generates the representation denoted as z ∈ R(P|S|i=1 Ni)×d. Finally, a flatten layer with representation projection head is used to obtain the molecular spectra representation zs ∈ Rd.",
|
| 93 |
+
"edge": [],
|
| 94 |
+
"level": 4,
|
| 95 |
+
"visual_node": []
|
| 96 |
+
},
|
| 97 |
+
{
|
| 98 |
+
"name": "Masked Patches Reconstruction (3.2)",
|
| 99 |
+
"content": "# 3.2 MASKED PATCHES RECONSTRUCTION PRE-TRAINING FOR SPECTRA Before distilling the spectra information into 3D molecular representation learning, we need first ensure that the spectrum encoder can effectively comprehend molecular spectra and generate spectral representations. Considering the success of masking modeling across various domains (Devlin et al., 2019; He et al., 2022; Hou et al., 2022; Xia et al., 2023; Wang et al., 2024b; Nie et al., 2023), we propose a masked patches reconstruction (MPR) objective to guide the training of SpecFormer. After the patching step, we randomly select a portion of patches according to the mask ratio α and replace them with zero vectors to implement the masking. Subsequently, the masked patches undergo patch encoding and position encoding. In this way, the semantics of the masked patches (the absorption intensity at specific wavelengths) are obscured during patch encoding, while the positional information is retained to facilitate the reconstruction of the original semantics. After encoding by SpecFormer, the encoded results corresponding to the masked patches are input into a spectrum-specific reconstruction head to reconstruct the original spectral values that were masked. The mean squared error (MSE) between the reconstruction results and the original masked spectra serves as the loss function for the MPR task, guiding the training of SpecFormer:  where Pi denotes the set of masked patches in the i-th type of molecular spectra, and pˆi,j denotes ethe reconstructed patch corresponding to the masked patch pi,j .",
|
| 100 |
+
"edge": [
|
| 101 |
+
"MPR Motivation and Masking",
|
| 102 |
+
"Reconstruction Head and Loss"
|
| 103 |
+
],
|
| 104 |
+
"level": 2,
|
| 105 |
+
"visual_node": []
|
| 106 |
+
},
|
| 107 |
+
{
|
| 108 |
+
"name": "MPR Motivation and Masking",
|
| 109 |
+
"content": "Before distilling the spectra information into 3D molecular representation learning, we need first ensure that the spectrum encoder can effectively comprehend molecular spectra and generate spectral representations. Considering the success of masking modeling across various domains (Devlin et al., 2019; He et al., 2022; Hou et al., 2022; Xia et al., 2023; Wang et al., 2024b; Nie et al., 2023), we propose a masked patches reconstruction (MPR) objective to guide the training of SpecFormer. After the patching step, we randomly select a portion of patches according to the mask ratio α and replace them with zero vectors to implement the masking. Subsequently, the masked patches undergo patch encoding and position encoding. In this way, the semantics of the masked patches (the absorption intensity at specific wavelengths) are obscured during patch encoding, while the positional information is retained to facilitate the reconstruction of the original semantics.",
|
| 110 |
+
"edge": [
|
| 111 |
+
"MPR Motivation",
|
| 112 |
+
"Masking Strategy"
|
| 113 |
+
],
|
| 114 |
+
"level": 3,
|
| 115 |
+
"visual_node": []
|
| 116 |
+
},
|
| 117 |
+
{
|
| 118 |
+
"name": "MPR Motivation",
|
| 119 |
+
"content": "Before distilling the spectra information into 3D molecular representation learning, we need first ensure that the spectrum encoder can effectively comprehend molecular spectra and generate spectral representations. Considering the success of masking modeling across various domains (Devlin et al., 2019; He et al., 2022; Hou et al., 2022; Xia et al., 2023; Wang et al., 2024b; Nie et al., 2023), we propose a masked patches reconstruction (MPR) objective to guide the training of SpecFormer.",
|
| 120 |
+
"edge": [],
|
| 121 |
+
"level": 4,
|
| 122 |
+
"visual_node": []
|
| 123 |
+
},
|
| 124 |
+
{
|
| 125 |
+
"name": "Masking Strategy",
|
| 126 |
+
"content": "After the patching step, we randomly select a portion of patches according to the mask ratio α and replace them with zero vectors to implement the masking. Subsequently, the masked patches undergo patch encoding and position encoding. In this way, the semantics of the masked patches (the absorption intensity at specific wavelengths) are obscured during patch encoding, while the positional information is retained to facilitate the reconstruction of the original semantics.",
|
| 127 |
+
"edge": [],
|
| 128 |
+
"level": 4,
|
| 129 |
+
"visual_node": []
|
| 130 |
+
},
|
| 131 |
+
{
|
| 132 |
+
"name": "Reconstruction Head and Loss",
|
| 133 |
+
"content": "After encoding by SpecFormer, the encoded results corresponding to the masked patches are input into a spectrum-specific reconstruction head to reconstruct the original spectral values that were masked. The mean squared error (MSE) between the reconstruction results and the original masked spectra serves as the loss function for the MPR task, guiding the training of SpecFormer:  where Pi denotes the set of masked patches in the i-th type of molecular spectra, and pˆi,j denotes ethe reconstructed patch corresponding to the masked patch pi,j .",
|
| 134 |
+
"edge": [],
|
| 135 |
+
"level": 3,
|
| 136 |
+
"visual_node": []
|
| 137 |
+
},
|
| 138 |
+
{
|
| 139 |
+
"name": "Contrastive Learning (3.3)",
|
| 140 |
+
"content": "# 3.3 CONTRASTIVE LEARNING BETWEEN 3D STRUCTURES AND SPECTRA Under the guidance of the denoising objective for 3D representation learning and the MPR objective for spectral representation learning, we further introduce a contrastive objective to align the representations across these two modalities. We treat the 3D representation zx ∈ Rd and spectral representation zs ∈ Rd of the same molecule as positive samples, and negative samples otherwise. Subsequently, the consistency between positive samples and the discrepancy between negative samples are maximized through the contrastive objective. Given the theoretical and empirical effectiveness, we employ InfoNCE (van den Oord et al., 2018) as the contrastive objective:  where zjx, zjs are randomly sampled 3D and spectra views regarding to the positive pair (zx, zs). fx(zx, zs) and fs(zs, zx) are scoring functions for the two corresponding views, with flexible formulations. Here we adopt fx(zx, zs) = fs(zs, zx) = ⟨zx, zs⟩. Note that the denoising objective can utilize any form from existing 3D molecular representation pre-training studies, enabling seamless integration of our method into these frameworks.",
|
| 141 |
+
"edge": [
|
| 142 |
+
"Contrastive Objective Formulation",
|
| 143 |
+
"Integration with Denoising"
|
| 144 |
+
],
|
| 145 |
+
"level": 2,
|
| 146 |
+
"visual_node": []
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"name": "Contrastive Objective Formulation",
|
| 150 |
+
"content": "Under the guidance of the denoising objective for 3D representation learning and the MPR objective for spectral representation learning, we further introduce a contrastive objective to align the representations across these two modalities. We treat the 3D representation zx ∈ Rd and spectral representation zs ∈ Rd of the same molecule as positive samples, and negative samples otherwise. Subsequently, the consistency between positive samples and the discrepancy between negative samples are maximized through the contrastive objective. Given the theoretical and empirical effectiveness, we employ InfoNCE (van den Oord et al., 2018) as the contrastive objective:  where zjx, zjs are randomly sampled 3D and spectra views regarding to the positive pair (zx, zs). fx(zx, zs) and fs(zs, zx) are scoring functions for the two corresponding views, with flexible formulations. Here we adopt fx(zx, zs) = fs(zs, zx) = ⟨zx, zs⟩.",
|
| 151 |
+
"edge": [],
|
| 152 |
+
"level": 3,
|
| 153 |
+
"visual_node": []
|
| 154 |
+
},
|
| 155 |
+
{
|
| 156 |
+
"name": "Integration with Denoising",
|
| 157 |
+
"content": "Note that the denoising objective can utilize any form from existing 3D molecular representation pre-training studies, enabling seamless integration of our method into these frameworks.",
|
| 158 |
+
"edge": [],
|
| 159 |
+
"level": 3,
|
| 160 |
+
"visual_node": []
|
| 161 |
+
},
|
| 162 |
+
{
|
| 163 |
+
"name": "Two-Stage Pipeline (3.4)",
|
| 164 |
+
"content": "# 3.4 TWO-STAGE PRE-TRAINING PIPELINE Previous pre-training efforts for 3D molecular representation have been conducted on unlabeled datasets using denoising objective. These datasets typically provide only equilibrium 3D structures without offering spectra for all molecules. To enhance the pre-training effect by incorporating spectra while leveraging denoising pre-training, we employ a two-stage pre-training approach. The first stage involves training on a larger dataset (Nakata & Shimazaki, 2017) without spectra using only the denoising objective. Subsequently, the second stage involves training on a dataset that includes spectra using the complete objective as follows:  where βDenoising, βMPR, and βContrast denote the weights of each sub-objective.",
|
| 165 |
+
"edge": [
|
| 166 |
+
"Pre-training Context",
|
| 167 |
+
"Two-Stage Method"
|
| 168 |
+
],
|
| 169 |
+
"level": 2,
|
| 170 |
+
"visual_node": []
|
| 171 |
+
},
|
| 172 |
+
{
|
| 173 |
+
"name": "Pre-training Context",
|
| 174 |
+
"content": "Previous pre-training efforts for 3D molecular representation have been conducted on unlabeled datasets using denoising objective. These datasets typically provide only equilibrium 3D structures without offering spectra for all molecules.",
|
| 175 |
+
"edge": [],
|
| 176 |
+
"level": 3,
|
| 177 |
+
"visual_node": []
|
| 178 |
+
},
|
| 179 |
+
{
|
| 180 |
+
"name": "Two-Stage Method",
|
| 181 |
+
"content": "To enhance the pre-training effect by incorporating spectra while leveraging denoising pre-training, we employ a two-stage pre-training approach. The first stage involves training on a larger dataset (Nakata & Shimazaki, 2017) without spectra using only the denoising objective. Subsequently, the second stage involves training on a dataset that includes spectra using the complete objective as follows:  where βDenoising, βMPR, and βContrast denote the weights of each sub-objective.",
|
| 182 |
+
"edge": [],
|
| 183 |
+
"level": 3,
|
| 184 |
+
"visual_node": []
|
| 185 |
+
}
|
| 186 |
+
]
|
| 187 |
+
}
|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/4 EXPERIMENTS.md_dag.json
ADDED
|
@@ -0,0 +1,187 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"nodes": [
|
| 3 |
+
{
|
| 4 |
+
"name": "4 EXPERIMENTS",
|
| 5 |
+
"content": "# 4 EXPERIMENTS To comprehensively evaluate the impact of molecular spectra on molecular tasks, we first verify the effectiveness of molecular spectra in the training-from-scratch method for the downstream task. Furthermore, we evaluate the effectiveness of our pre-training framework MolSpectra. # 4.1 EFFECTIVENESS OF MOLECULAR SPECTRA IN TRAINING FROM SCRATCH This pilot experiment aims to demonstrate the rationality for incorporating molecular spectra into pre-training. We introduce additional spectral features into a train-from-scratch molecular property prediction model to observe the impact of spectral information on prediction outcomes. We employ EGNN (Satorras et al., 2021), a representative 3D molecular encoder, equipped with an MLP-based prediction head as the baseline model. While EGNN encodes the 3D representations, the UV-Vis spectrum of each molecule provided by the QM9S (Zou et al., 2023) dataset is encoded into spectral representations by a spectrum encoder. Before making predictions with the final MLP, we concatenate the spectral and 3D representations for prediction. The results are presented in Table 1. Table 1: Performance (MAE ↓) when training from scratch on QM9 dataset.  We observe that by directly concatenating spectral representations, the performance of molecular property prediction can be effectively enhanced. This indicates that the information from molecular spectra is beneficial for downstream molecular property prediction. Further incorporating molecular spectra into the pre-training phase of molecular representation has the potential to enhance the informativeness and generalization capability of the representations, thereby broadly improving the performance of downstream tasks. # 4.2 EFFECTIVENESS OF MOLECULAR SPECTRA IN REPRESENTATION PRE-TRAINING We conduct experiments to evaluate MolSpectra by first introducing spectral data into the pretraining of 3D representations, followed by evaluating the performance on downstream tasks. For a comprehensive comparison, two types of baselines are adopted: (1) training-from-scratch methods, including SchNet (Schutt et al. ¨ , 2017), EGNN, DimeNet (Klicpera et al., 2020b), DimeNet++ (Klicpera et al., 2020a), PaiNN (Schutt et al. ¨ , 2021), SphereNet (Liu et al., 2021), and TorchMD-Net (Tholke & Fabritiis ¨ , 2022); and (2) pre-training methods, including Transformer-M (Luo et al., 2023), SE(3)-DDM (Liu et al., 2023b), 3D-EMGP (Jiao et al., 2023), and Coord. MolSpectra can be seamlessly plugged into any existing denoising method. To evaluate the enhancement provided by our method compared to denoising alone, we select the representative coordinate denoising (Coord) as our denoising sub-objective. This method also serves as our primary baseline. # 4.2.1 PRE-TRAINING DATASET. As described in Section 3.4, we first perform denoising pre-training on the PCQM4Mv2 (Nakata & Shimazaki, 2017) dataset, followed by a second stage of pre-training on the QM9Spectra (QM9S) (Zou et al., 2023) dataset, which includes multi-modal molecular energy spectra. In both stages, we adopt the denoising objective provided by Coord (Zaidi et al., 2023), as defined in Eq. 2. The QM9S dataset comprises organic molecules from the QM9 (Ramakrishnan et al., 2014) dataset. The UV-Vis, IR, and Raman spectra of the molecules are calculated at the B3LYP/def-TZVP level of theory, through frequency analysis and time-dependent density functional theory (TD-DFT). # 4.2.2 QM9 The QM9 dataset is a quantum chemistry dataset comprising over 134,000 small molecules, each consisting of up to 9 heavy atoms (C, N, O, F) and additional H atoms. This dataset provides an equilibrium geometric conformation for each molecule along with 12 property labels. The dataset is divided into a training set of 110k molecules, a validation set of 10k molecules, and a test set containing the remaining over 10k molecules. Prediction errors are measured using the mean absolute error (MAE). The experimental results are presented in Table 2. The 3D molecular representations pre-trained using our method are fine-tuned and used for prediction across various properties, achieving state-of-the-art performance in 8 out of 12 properties and outperforms Coord in 10 out of 12 properties. In conjunction with the observations in Section 4.1, the performance improvement can be attributed to our incorporation of an understanding of molecular spectra and the knowledge they entail into the 3D molecular representations. Table 2: Performance (MAE↓) on QM9 dataset. The compared methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are highlighted in bold.  Table 3: Performance (MAE↓) on MD17 force prediction (kcal/mol/ ˚A). The methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are in bold.  # 4.2.3 MD17 The MD17 dataset contains molecular dynamics trajectories for eight organic molecules, including aspirin, benzene, and ethanol. It offers 150k to nearly 1M conformations per molecule, with energy and force labels. Unlike QM9, MD17 emphasizes dynamic behavior in addition to static properties. We use a standard limited data split: models train on 1k samples, validate on 50, and test on the rest. Performance is evaluated using MAE, with results in Table 3. Our approach also results in the expected performance improvement on MD17. MD17 is a dataset comprising a large number of non-equilibrium molecular structures and their corresponding force fields, which serves to evaluate a model’s understanding of molecular dynamics. However, previous pre-training methods based solely on denoising have only learned force field patterns at static equilibrium states, failing to adequately capture the dynamic evolution of molecular systems. In contrast, our MolSpectra learns the dynamic evolution of molecules by understanding energy level transition patterns, thereby outperforming denoising-based pre-training methods. 4.3 SENSITIVITY ANALYSIS OF PATCH LENGTH Pi, STRIDE Di, AND MASK RATIO α We conduct experiments to evaluate the impact of patch length Pi, stride Di, and mask ratio α. Results are summarized in Table 4 and Table 5. From Table 4, we observe that when consecutive patches have overlap (Di < Pi), the performance of pre-training is superior compared to scenarios without overlap (Di = Pi). Specifically, the performance is optimal when the stride is half of the patch length. This is because appropriate overlap can better preserve and capture local features, particularly the information at the patch boundaries. Additionally, we find that choosing an appropriate patch length further enhances performance. In our experiments, the configuration of Pi = 20, Di = 10 yields the best results. Table 4: Sensitivity of patch length and stride.  Table 5: Sensitivity of mask ratio.  Regarding the mask ratio, α = 0.10 is a preferable choice. A small mask ratio results in insufficient MPR optimization, hindering SpecFormer training. Conversely, a large mask ratio causes excessive spectral perturbation, degrading performance when aligning with the 3D representations with the contrastive objective. An appropriate mask ratio strikes a balance between these two aspects. # 4.4 ABLATION STUDY To rigorously demonstrate the contributions of masked patches reconstruction, the incorporation of molecular spectra, and each spectral modality, we conducted an ablation study on them. Ablation study of masked patches reconstruction. We remove the MPR loss to analyze the impact of masked patches reconstruction, referred to as “w/o MPR” in Table 6. Removing the MPR objective leads to performance deterioration. This is consistent with the sensitivity analysis of the mask ratio α in Section 4.3, as removing MPR is an extreme case where α = 0. This decline is due to the lack of effective guidance in training SpecFormer. Using an undertrained SpecFormer for contrastive learning with 3D encoder outputs limits performance improvement. Table 6: Ablation of optimization objectives.  Ablation study of molecular spectra. We retain only the denoising loss, removing both the MPR loss and contrastive loss, referred to as “w/o MPR, Contrast” in Table 6. The only difference between this variant and MolSpectra is the incorporation of molecular spectra into the pre-training. The ”w/o MPR, Contrast” results are inferior to those of MolSpectra, highlighting that incorporating molecular spectra effectively enhances the quality and generalizability of molecular 3D representations. Ablation study of each spectral modality. To evaluate the contributions of each spectral modality to the performance, we conduct an ablation study for each modality. The results are presented in Table 7. It can be observed that each spectral modality contributes differently, with the UV-Vis spectrum having the smallest contribution and the IR spectrum the largest, likely due to the varying information content in each modality. Table 7: Ablation of spectral modalities. ",
|
| 6 |
+
"edge": [
|
| 7 |
+
"Experiment Overview",
|
| 8 |
+
"4.1 Effectiveness in Training from Scratch",
|
| 9 |
+
"4.2 Effectiveness in Representation Pre-training",
|
| 10 |
+
"4.3 Sensitivity Analysis",
|
| 11 |
+
"4.4 Ablation Study"
|
| 12 |
+
],
|
| 13 |
+
"level": 1,
|
| 14 |
+
"visual_node": []
|
| 15 |
+
},
|
| 16 |
+
{
|
| 17 |
+
"name": "Experiment Overview",
|
| 18 |
+
"content": "To comprehensively evaluate the impact of molecular spectra on molecular tasks, we first verify the effectiveness of molecular spectra in the training-from-scratch method for the downstream task. Furthermore, we evaluate the effectiveness of our pre-training framework MolSpectra.",
|
| 19 |
+
"edge": [],
|
| 20 |
+
"level": 2,
|
| 21 |
+
"visual_node": []
|
| 22 |
+
},
|
| 23 |
+
{
|
| 24 |
+
"name": "4.1 Effectiveness in Training from Scratch",
|
| 25 |
+
"content": "# 4.1 EFFECTIVENESS OF MOLECULAR SPECTRA IN TRAINING FROM SCRATCH This pilot experiment aims to demonstrate the rationality for incorporating molecular spectra into pre-training. We introduce additional spectral features into a train-from-scratch molecular property prediction model to observe the impact of spectral information on prediction outcomes. We employ EGNN (Satorras et al., 2021), a representative 3D molecular encoder, equipped with an MLP-based prediction head as the baseline model. While EGNN encodes the 3D representations, the UV-Vis spectrum of each molecule provided by the QM9S (Zou et al., 2023) dataset is encoded into spectral representations by a spectrum encoder. Before making predictions with the final MLP, we concatenate the spectral and 3D representations for prediction. The results are presented in Table 1. Table 1: Performance (MAE ↓) when training from scratch on QM9 dataset.  We observe that by directly concatenating spectral representations, the performance of molecular property prediction can be effectively enhanced. This indicates that the information from molecular spectra is beneficial for downstream molecular property prediction. Further incorporating molecular spectra into the pre-training phase of molecular representation has the potential to enhance the informativeness and generalization capability of the representations, thereby broadly improving the performance of downstream tasks.",
|
| 26 |
+
"edge": [
|
| 27 |
+
"4.1 Methodology",
|
| 28 |
+
"4.1 Results Analysis"
|
| 29 |
+
],
|
| 30 |
+
"level": 2,
|
| 31 |
+
"visual_node": []
|
| 32 |
+
},
|
| 33 |
+
{
|
| 34 |
+
"name": "4.1 Methodology",
|
| 35 |
+
"content": "This pilot experiment aims to demonstrate the rationality for incorporating molecular spectra into pre-training. We introduce additional spectral features into a train-from-scratch molecular property prediction model to observe the impact of spectral information on prediction outcomes. We employ EGNN (Satorras et al., 2021), a representative 3D molecular encoder, equipped with an MLP-based prediction head as the baseline model. While EGNN encodes the 3D representations, the UV-Vis spectrum of each molecule provided by the QM9S (Zou et al., 2023) dataset is encoded into spectral representations by a spectrum encoder. Before making predictions with the final MLP, we concatenate the spectral and 3D representations for prediction. The results are presented in Table 1. Table 1: Performance (MAE ↓) when training from scratch on QM9 dataset. ",
|
| 36 |
+
"edge": [],
|
| 37 |
+
"level": 3,
|
| 38 |
+
"visual_node": []
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"name": "4.1 Results Analysis",
|
| 42 |
+
"content": "We observe that by directly concatenating spectral representations, the performance of molecular property prediction can be effectively enhanced. This indicates that the information from molecular spectra is beneficial for downstream molecular property prediction. Further incorporating molecular spectra into the pre-training phase of molecular representation has the potential to enhance the informativeness and generalization capability of the representations, thereby broadly improving the performance of downstream tasks.",
|
| 43 |
+
"edge": [],
|
| 44 |
+
"level": 3,
|
| 45 |
+
"visual_node": []
|
| 46 |
+
},
|
| 47 |
+
{
|
| 48 |
+
"name": "4.2 Effectiveness in Representation Pre-training",
|
| 49 |
+
"content": "# 4.2 EFFECTIVENESS OF MOLECULAR SPECTRA IN REPRESENTATION PRE-TRAINING We conduct experiments to evaluate MolSpectra by first introducing spectral data into the pretraining of 3D representations, followed by evaluating the performance on downstream tasks. For a comprehensive comparison, two types of baselines are adopted: (1) training-from-scratch methods, including SchNet (Schutt et al. ¨ , 2017), EGNN, DimeNet (Klicpera et al., 2020b), DimeNet++ (Klicpera et al., 2020a), PaiNN (Schutt et al. ¨ , 2021), SphereNet (Liu et al., 2021), and TorchMD-Net (Tholke & Fabritiis ¨ , 2022); and (2) pre-training methods, including Transformer-M (Luo et al., 2023), SE(3)-DDM (Liu et al., 2023b), 3D-EMGP (Jiao et al., 2023), and Coord. MolSpectra can be seamlessly plugged into any existing denoising method. To evaluate the enhancement provided by our method compared to denoising alone, we select the representative coordinate denoising (Coord) as our denoising sub-objective. This method also serves as our primary baseline. # 4.2.1 PRE-TRAINING DATASET. As described in Section 3.4, we first perform denoising pre-training on the PCQM4Mv2 (Nakata & Shimazaki, 2017) dataset, followed by a second stage of pre-training on the QM9Spectra (QM9S) (Zou et al., 2023) dataset, which includes multi-modal molecular energy spectra. In both stages, we adopt the denoising objective provided by Coord (Zaidi et al., 2023), as defined in Eq. 2. The QM9S dataset comprises organic molecules from the QM9 (Ramakrishnan et al., 2014) dataset. The UV-Vis, IR, and Raman spectra of the molecules are calculated at the B3LYP/def-TZVP level of theory, through frequency analysis and time-dependent density functional theory (TD-DFT). # 4.2.2 QM9 The QM9 dataset is a quantum chemistry dataset comprising over 134,000 small molecules, each consisting of up to 9 heavy atoms (C, N, O, F) and additional H atoms. This dataset provides an equilibrium geometric conformation for each molecule along with 12 property labels. The dataset is divided into a training set of 110k molecules, a validation set of 10k molecules, and a test set containing the remaining over 10k molecules. Prediction errors are measured using the mean absolute error (MAE). The experimental results are presented in Table 2. The 3D molecular representations pre-trained using our method are fine-tuned and used for prediction across various properties, achieving state-of-the-art performance in 8 out of 12 properties and outperforms Coord in 10 out of 12 properties. In conjunction with the observations in Section 4.1, the performance improvement can be attributed to our incorporation of an understanding of molecular spectra and the knowledge they entail into the 3D molecular representations. Table 2: Performance (MAE↓) on QM9 dataset. The compared methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are highlighted in bold.  Table 3: Performance (MAE↓) on MD17 force prediction (kcal/mol/ ˚A). The methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are in bold.  # 4.2.3 MD17 The MD17 dataset contains molecular dynamics trajectories for eight organic molecules, including aspirin, benzene, and ethanol. It offers 150k to nearly 1M conformations per molecule, with energy and force labels. Unlike QM9, MD17 emphasizes dynamic behavior in addition to static properties. We use a standard limited data split: models train on 1k samples, validate on 50, and test on the rest. Performance is evaluated using MAE, with results in Table 3. Our approach also results in the expected performance improvement on MD17. MD17 is a dataset comprising a large number of non-equilibrium molecular structures and their corresponding force fields, which serves to evaluate a model’s understanding of molecular dynamics. However, previous pre-training methods based solely on denoising have only learned force field patterns at static equilibrium states, failing to adequately capture the dynamic evolution of molecular systems. In contrast, our MolSpectra learns the dynamic evolution of molecules by understanding energy level transition patterns, thereby outperforming denoising-based pre-training methods.",
|
| 50 |
+
"edge": [
|
| 51 |
+
"4.2 Baselines and Setup",
|
| 52 |
+
"4.2.1 Pre-training Dataset",
|
| 53 |
+
"4.2.2 QM9 Evaluation",
|
| 54 |
+
"4.2.3 MD17 Evaluation"
|
| 55 |
+
],
|
| 56 |
+
"level": 2,
|
| 57 |
+
"visual_node": []
|
| 58 |
+
},
|
| 59 |
+
{
|
| 60 |
+
"name": "4.2 Baselines and Setup",
|
| 61 |
+
"content": "We conduct experiments to evaluate MolSpectra by first introducing spectral data into the pretraining of 3D representations, followed by evaluating the performance on downstream tasks. For a comprehensive comparison, two types of baselines are adopted: (1) training-from-scratch methods, including SchNet (Schutt et al. ¨ , 2017), EGNN, DimeNet (Klicpera et al., 2020b), DimeNet++ (Klicpera et al., 2020a), PaiNN (Schutt et al. ¨ , 2021), SphereNet (Liu et al., 2021), and TorchMD-Net (Tholke & Fabritiis ¨ , 2022); and (2) pre-training methods, including Transformer-M (Luo et al., 2023), SE(3)-DDM (Liu et al., 2023b), 3D-EMGP (Jiao et al., 2023), and Coord. MolSpectra can be seamlessly plugged into any existing denoising method. To evaluate the enhancement provided by our method compared to denoising alone, we select the representative coordinate denoising (Coord) as our denoising sub-objective. This method also serves as our primary baseline.",
|
| 62 |
+
"edge": [],
|
| 63 |
+
"level": 3,
|
| 64 |
+
"visual_node": []
|
| 65 |
+
},
|
| 66 |
+
{
|
| 67 |
+
"name": "4.2.1 Pre-training Dataset",
|
| 68 |
+
"content": "# 4.2.1 PRE-TRAINING DATASET. As described in Section 3.4, we first perform denoising pre-training on the PCQM4Mv2 (Nakata & Shimazaki, 2017) dataset, followed by a second stage of pre-training on the QM9Spectra (QM9S) (Zou et al., 2023) dataset, which includes multi-modal molecular energy spectra. In both stages, we adopt the denoising objective provided by Coord (Zaidi et al., 2023), as defined in Eq. 2. The QM9S dataset comprises organic molecules from the QM9 (Ramakrishnan et al., 2014) dataset. The UV-Vis, IR, and Raman spectra of the molecules are calculated at the B3LYP/def-TZVP level of theory, through frequency analysis and time-dependent density functional theory (TD-DFT).",
|
| 69 |
+
"edge": [],
|
| 70 |
+
"level": 3,
|
| 71 |
+
"visual_node": []
|
| 72 |
+
},
|
| 73 |
+
{
|
| 74 |
+
"name": "4.2.2 QM9 Evaluation",
|
| 75 |
+
"content": "# 4.2.2 QM9 The QM9 dataset is a quantum chemistry dataset comprising over 134,000 small molecules, each consisting of up to 9 heavy atoms (C, N, O, F) and additional H atoms. This dataset provides an equilibrium geometric conformation for each molecule along with 12 property labels. The dataset is divided into a training set of 110k molecules, a validation set of 10k molecules, and a test set containing the remaining over 10k molecules. Prediction errors are measured using the mean absolute error (MAE). The experimental results are presented in Table 2. The 3D molecular representations pre-trained using our method are fine-tuned and used for prediction across various properties, achieving state-of-the-art performance in 8 out of 12 properties and outperforms Coord in 10 out of 12 properties. In conjunction with the observations in Section 4.1, the performance improvement can be attributed to our incorporation of an understanding of molecular spectra and the knowledge they entail into the 3D molecular representations. Table 2: Performance (MAE↓) on QM9 dataset. The compared methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are highlighted in bold. ",
|
| 76 |
+
"edge": [
|
| 77 |
+
"QM9 Dataset Description",
|
| 78 |
+
"QM9 Performance Analysis"
|
| 79 |
+
],
|
| 80 |
+
"level": 3,
|
| 81 |
+
"visual_node": []
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"name": "QM9 Dataset Description",
|
| 85 |
+
"content": "The QM9 dataset is a quantum chemistry dataset comprising over 134,000 small molecules, each consisting of up to 9 heavy atoms (C, N, O, F) and additional H atoms. This dataset provides an equilibrium geometric conformation for each molecule along with 12 property labels. The dataset is divided into a training set of 110k molecules, a validation set of 10k molecules, and a test set containing the remaining over 10k molecules. Prediction errors are measured using the mean absolute error (MAE). The experimental results are presented in Table 2. Table 2: Performance (MAE↓) on QM9 dataset. The compared methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are highlighted in bold. ",
|
| 86 |
+
"edge": [],
|
| 87 |
+
"level": 4,
|
| 88 |
+
"visual_node": []
|
| 89 |
+
},
|
| 90 |
+
{
|
| 91 |
+
"name": "QM9 Performance Analysis",
|
| 92 |
+
"content": "The 3D molecular representations pre-trained using our method are fine-tuned and used for prediction across various properties, achieving state-of-the-art performance in 8 out of 12 properties and outperforms Coord in 10 out of 12 properties. In conjunction with the observations in Section 4.1, the performance improvement can be attributed to our incorporation of an understanding of molecular spectra and the knowledge they entail into the 3D molecular representations.",
|
| 93 |
+
"edge": [],
|
| 94 |
+
"level": 4,
|
| 95 |
+
"visual_node": []
|
| 96 |
+
},
|
| 97 |
+
{
|
| 98 |
+
"name": "4.2.3 MD17 Evaluation",
|
| 99 |
+
"content": "Table 3: Performance (MAE↓) on MD17 force prediction (kcal/mol/ ˚A). The methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are in bold.  # 4.2.3 MD17 The MD17 dataset contains molecular dynamics trajectories for eight organic molecules, including aspirin, benzene, and ethanol. It offers 150k to nearly 1M conformations per molecule, with energy and force labels. Unlike QM9, MD17 emphasizes dynamic behavior in addition to static properties. We use a standard limited data split: models train on 1k samples, validate on 50, and test on the rest. Performance is evaluated using MAE, with results in Table 3. Our approach also results in the expected performance improvement on MD17. MD17 is a dataset comprising a large number of non-equilibrium molecular structures and their corresponding force fields, which serves to evaluate a model’s understanding of molecular dynamics. However, previous pre-training methods based solely on denoising have only learned force field patterns at static equilibrium states, failing to adequately capture the dynamic evolution of molecular systems. In contrast, our MolSpectra learns the dynamic evolution of molecules by understanding energy level transition patterns, thereby outperforming denoising-based pre-training methods.",
|
| 100 |
+
"edge": [
|
| 101 |
+
"MD17 Dataset Description",
|
| 102 |
+
"MD17 Performance Analysis"
|
| 103 |
+
],
|
| 104 |
+
"level": 3,
|
| 105 |
+
"visual_node": []
|
| 106 |
+
},
|
| 107 |
+
{
|
| 108 |
+
"name": "MD17 Dataset Description",
|
| 109 |
+
"content": "Table 3: Performance (MAE↓) on MD17 force prediction (kcal/mol/ ˚A). The methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are in bold.  # 4.2.3 MD17 The MD17 dataset contains molecular dynamics trajectories for eight organic molecules, including aspirin, benzene, and ethanol. It offers 150k to nearly 1M conformations per molecule, with energy and force labels. Unlike QM9, MD17 emphasizes dynamic behavior in addition to static properties. We use a standard limited data split: models train on 1k samples, validate on 50, and test on the rest. Performance is evaluated using MAE, with results in Table 3.",
|
| 110 |
+
"edge": [],
|
| 111 |
+
"level": 4,
|
| 112 |
+
"visual_node": []
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"name": "MD17 Performance Analysis",
|
| 116 |
+
"content": "Our approach also results in the expected performance improvement on MD17. MD17 is a dataset comprising a large number of non-equilibrium molecular structures and their corresponding force fields, which serves to evaluate a model’s understanding of molecular dynamics. However, previous pre-training methods based solely on denoising have only learned force field patterns at static equilibrium states, failing to adequately capture the dynamic evolution of molecular systems. In contrast, our MolSpectra learns the dynamic evolution of molecules by understanding energy level transition patterns, thereby outperforming denoising-based pre-training methods.",
|
| 117 |
+
"edge": [],
|
| 118 |
+
"level": 4,
|
| 119 |
+
"visual_node": []
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"name": "4.3 Sensitivity Analysis",
|
| 123 |
+
"content": "4.3 SENSITIVITY ANALYSIS OF PATCH LENGTH Pi, STRIDE Di, AND MASK RATIO α We conduct experiments to evaluate the impact of patch length Pi, stride Di, and mask ratio α. Results are summarized in Table 4 and Table 5. From Table 4, we observe that when consecutive patches have overlap (Di < Pi), the performance of pre-training is superior compared to scenarios without overlap (Di = Pi). Specifically, the performance is optimal when the stride is half of the patch length. This is because appropriate overlap can better preserve and capture local features, particularly the information at the patch boundaries. Additionally, we find that choosing an appropriate patch length further enhances performance. In our experiments, the configuration of Pi = 20, Di = 10 yields the best results. Table 4: Sensitivity of patch length and stride.  Table 5: Sensitivity of mask ratio.  Regarding the mask ratio, α = 0.10 is a preferable choice. A small mask ratio results in insufficient MPR optimization, hindering SpecFormer training. Conversely, a large mask ratio causes excessive spectral perturbation, degrading performance when aligning with the 3D representations with the contrastive objective. An appropriate mask ratio strikes a balance between these two aspects.",
|
| 124 |
+
"edge": [
|
| 125 |
+
"Patch Length and Stride Analysis",
|
| 126 |
+
"Mask Ratio Analysis"
|
| 127 |
+
],
|
| 128 |
+
"level": 2,
|
| 129 |
+
"visual_node": []
|
| 130 |
+
},
|
| 131 |
+
{
|
| 132 |
+
"name": "Patch Length and Stride Analysis",
|
| 133 |
+
"content": "We conduct experiments to evaluate the impact of patch length Pi, stride Di, and mask ratio α. Results are summarized in Table 4 and Table 5. From Table 4, we observe that when consecutive patches have overlap (Di < Pi), the performance of pre-training is superior compared to scenarios without overlap (Di = Pi). Specifically, the performance is optimal when the stride is half of the patch length. This is because appropriate overlap can better preserve and capture local features, particularly the information at the patch boundaries. Additionally, we find that choosing an appropriate patch length further enhances performance. In our experiments, the configuration of Pi = 20, Di = 10 yields the best results. Table 4: Sensitivity of patch length and stride. ",
|
| 134 |
+
"edge": [],
|
| 135 |
+
"level": 3,
|
| 136 |
+
"visual_node": []
|
| 137 |
+
},
|
| 138 |
+
{
|
| 139 |
+
"name": "Mask Ratio Analysis",
|
| 140 |
+
"content": "Table 5: Sensitivity of mask ratio.  Regarding the mask ratio, α = 0.10 is a preferable choice. A small mask ratio results in insufficient MPR optimization, hindering SpecFormer training. Conversely, a large mask ratio causes excessive spectral perturbation, degrading performance when aligning with the 3D representations with the contrastive objective. An appropriate mask ratio strikes a balance between these two aspects.",
|
| 141 |
+
"edge": [],
|
| 142 |
+
"level": 3,
|
| 143 |
+
"visual_node": []
|
| 144 |
+
},
|
| 145 |
+
{
|
| 146 |
+
"name": "4.4 Ablation Study",
|
| 147 |
+
"content": "# 4.4 ABLATION STUDY To rigorously demonstrate the contributions of masked patches reconstruction, the incorporation of molecular spectra, and each spectral modality, we conducted an ablation study on them. Ablation study of masked patches reconstruction. We remove the MPR loss to analyze the impact of masked patches reconstruction, referred to as “w/o MPR” in Table 6. Removing the MPR objective leads to performance deterioration. This is consistent with the sensitivity analysis of the mask ratio α in Section 4.3, as removing MPR is an extreme case where α = 0. This decline is due to the lack of effective guidance in training SpecFormer. Using an undertrained SpecFormer for contrastive learning with 3D encoder outputs limits performance improvement. Table 6: Ablation of optimization objectives.  Ablation study of molecular spectra. We retain only the denoising loss, removing both the MPR loss and contrastive loss, referred to as “w/o MPR, Contrast” in Table 6. The only difference between this variant and MolSpectra is the incorporation of molecular spectra into the pre-training. The ”w/o MPR, Contrast” results are inferior to those of MolSpectra, highlighting that incorporating molecular spectra effectively enhances the quality and generalizability of molecular 3D representations. Ablation study of each spectral modality. To evaluate the contributions of each spectral modality to the performance, we conduct an ablation study for each modality. The results are presented in Table 7. It can be observed that each spectral modality contributes differently, with the UV-Vis spectrum having the smallest contribution and the IR spectrum the largest, likely due to the varying information content in each modality. Table 7: Ablation of spectral modalities. ",
|
| 148 |
+
"edge": [
|
| 149 |
+
"MPR Ablation",
|
| 150 |
+
"Spectra and Modality Ablation"
|
| 151 |
+
],
|
| 152 |
+
"level": 2,
|
| 153 |
+
"visual_node": []
|
| 154 |
+
},
|
| 155 |
+
{
|
| 156 |
+
"name": "MPR Ablation",
|
| 157 |
+
"content": "To rigorously demonstrate the contributions of masked patches reconstruction, the incorporation of molecular spectra, and each spectral modality, we conducted an ablation study on them. Ablation study of masked patches reconstruction. We remove the MPR loss to analyze the impact of masked patches reconstruction, referred to as “w/o MPR” in Table 6. Removing the MPR objective leads to performance deterioration. This is consistent with the sensitivity analysis of the mask ratio α in Section 4.3, as removing MPR is an extreme case where α = 0. This decline is due to the lack of effective guidance in training SpecFormer. Using an undertrained SpecFormer for contrastive learning with 3D encoder outputs limits performance improvement. Table 6: Ablation of optimization objectives. ",
|
| 158 |
+
"edge": [],
|
| 159 |
+
"level": 3,
|
| 160 |
+
"visual_node": []
|
| 161 |
+
},
|
| 162 |
+
{
|
| 163 |
+
"name": "Spectra and Modality Ablation",
|
| 164 |
+
"content": "Ablation study of molecular spectra. We retain only the denoising loss, removing both the MPR loss and contrastive loss, referred to as “w/o MPR, Contrast” in Table 6. The only difference between this variant and MolSpectra is the incorporation of molecular spectra into the pre-training. The ”w/o MPR, Contrast” results are inferior to those of MolSpectra, highlighting that incorporating molecular spectra effectively enhances the quality and generalizability of molecular 3D representations. Ablation study of each spectral modality. To evaluate the contributions of each spectral modality to the performance, we conduct an ablation study for each modality. The results are presented in Table 7. It can be observed that each spectral modality contributes differently, with the UV-Vis spectrum having the smallest contribution and the IR spectrum the largest, likely due to the varying information content in each modality. Table 7: Ablation of spectral modalities. ",
|
| 165 |
+
"edge": [
|
| 166 |
+
"Spectra Ablation",
|
| 167 |
+
"Modality Ablation"
|
| 168 |
+
],
|
| 169 |
+
"level": 3,
|
| 170 |
+
"visual_node": []
|
| 171 |
+
},
|
| 172 |
+
{
|
| 173 |
+
"name": "Spectra Ablation",
|
| 174 |
+
"content": "Ablation study of molecular spectra. We retain only the denoising loss, removing both the MPR loss and contrastive loss, referred to as “w/o MPR, Contrast” in Table 6. The only difference between this variant and MolSpectra is the incorporation of molecular spectra into the pre-training. The ”w/o MPR, Contrast” results are inferior to those of MolSpectra, highlighting that incorporating molecular spectra effectively enhances the quality and generalizability of molecular 3D representations.",
|
| 175 |
+
"edge": [],
|
| 176 |
+
"level": 4,
|
| 177 |
+
"visual_node": []
|
| 178 |
+
},
|
| 179 |
+
{
|
| 180 |
+
"name": "Modality Ablation",
|
| 181 |
+
"content": "Ablation study of each spectral modality. To evaluate the contributions of each spectral modality to the performance, we conduct an ablation study for each modality. The results are presented in Table 7. It can be observed that each spectral modality contributes differently, with the UV-Vis spectrum having the smallest contribution and the IR spectrum the largest, likely due to the varying information content in each modality. Table 7: Ablation of spectral modalities. ",
|
| 182 |
+
"edge": [],
|
| 183 |
+
"level": 4,
|
| 184 |
+
"visual_node": []
|
| 185 |
+
}
|
| 186 |
+
]
|
| 187 |
+
}
|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_dag/6 CONCLUSION.md_dag.json
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"nodes": [
|
| 3 |
+
{
|
| 4 |
+
"name": "6 CONCLUSION.md",
|
| 5 |
+
"content": "# 6 CONCLUSION In this study, we explore pre-training molecular 3D representations beyond classical mechanics. By leveraging the correlation between molecular energy level structures and molecular spectra in quantum mechanics, we introduce molecular spectra for pre-training molecular 3D representations (MolSpectra). By aligning the 3D encoder trained with a denoising objective and the spectrum encoder trained with a masked patch reconstruction objective, we enhance the informativeness and transferability of the resulting 3D representations.",
|
| 6 |
+
"edge": [
|
| 7 |
+
"Study Overview",
|
| 8 |
+
"Methodology and Outcomes"
|
| 9 |
+
],
|
| 10 |
+
"level": 1,
|
| 11 |
+
"visual_node": []
|
| 12 |
+
},
|
| 13 |
+
{
|
| 14 |
+
"name": "Study Overview",
|
| 15 |
+
"content": "In this study, we explore pre-training molecular 3D representations beyond classical mechanics.",
|
| 16 |
+
"edge": [],
|
| 17 |
+
"level": 2,
|
| 18 |
+
"visual_node": []
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"name": "Methodology and Outcomes",
|
| 22 |
+
"content": "By leveraging the correlation between molecular energy level structures and molecular spectra in quantum mechanics, we introduce molecular spectra for pre-training molecular 3D representations (MolSpectra). By aligning the 3D encoder trained with a denoising objective and the spectrum encoder trained with a masked patch reconstruction objective, we enhance the informativeness and transferability of the resulting 3D representations.",
|
| 23 |
+
"edge": [
|
| 24 |
+
"Proposed Method",
|
| 25 |
+
"Technical Alignment and Results"
|
| 26 |
+
],
|
| 27 |
+
"level": 2,
|
| 28 |
+
"visual_node": []
|
| 29 |
+
},
|
| 30 |
+
{
|
| 31 |
+
"name": "Proposed Method",
|
| 32 |
+
"content": "By leveraging the correlation between molecular energy level structures and molecular spectra in quantum mechanics, we introduce molecular spectra for pre-training molecular 3D representations (MolSpectra).",
|
| 33 |
+
"edge": [],
|
| 34 |
+
"level": 3,
|
| 35 |
+
"visual_node": []
|
| 36 |
+
},
|
| 37 |
+
{
|
| 38 |
+
"name": "Technical Alignment and Results",
|
| 39 |
+
"content": "By aligning the 3D encoder trained with a denoising objective and the spectrum encoder trained with a masked patch reconstruction objective, we enhance the informativeness and transferability of the resulting 3D representations.",
|
| 40 |
+
"edge": [],
|
| 41 |
+
"level": 3,
|
| 42 |
+
"visual_node": []
|
| 43 |
+
}
|
| 44 |
+
]
|
| 45 |
+
}
|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/1 INTRODUCTION.md
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# 1 INTRODUCTION
|
| 2 |
+
|
| 3 |
+
Learning 3D molecular representations from geometric conformations offers a promising approach for understanding molecular geometry and predicting quantum properties and interactions, which is significant in drug discovery and materials science (Musaelian et al., 2023; Batatia et al., 2022; Liao & Smidt, 2023; Wang et al., 2023b; Du et al., 2023b). Given the scarcity of molecular property labels, self-supervised representation pre-training has been proposed and utilized to provide generalizable representations (Hu et al., 2020; Rong et al., 2020; Ma et al., 2024).
|
| 4 |
+
|
| 5 |
+
In contrast to contrastive learning (Wang et al., 2022; Kim et al., 2022) and masked modeling (Hou et al., 2022; Liu et al., 2023c; Wang et al., 2024b) on 2D molecular graphs and molecular languages (e.g., SMILES), the design of pre-training strategies on 3D molecular geometries is more closely aligned with physical principles. Previous studies (Zaidi et al., 2023; Jiao et al., 2023) have guided representation learning through denoising processes on 3D molecular geometries, theoretically demonstrating that denoising 3D geometries is equivalent to learning molecular force fields, specifically the negative gradient of molecular potential energy with respect to position. Essentially, these studies reveal that establishing the relationship between 3D geometries and the energy states of molecular systems is an effective pathway to learn 3D molecular representations.
|
| 6 |
+
|
| 7 |
+
However, existing methods are limited to the continuous description (i.e., the potential energy function) of the molecular energy states within the classical mechanics, overlooking the quantized (discrete) energy level structures from the quantum mechanical perspective. From the quantum perspective, molecular systems exhibit quantized energy level structures, meaning that energy states can only assume specific discrete values. Specifically, different types of molecular motion, such as electronic, vibrational, and rotational motion, correspond to different energy level structures. Knowledge of these energy levels is crucial in molecular physics and quantum chemistry, as they determine the spectroscopic characteristics, chemical reactivity, and many other important molecular properties. Fortunately, experimental measurements of molecular energy spectra can reflect these structures. Meanwhile, there are many molecular spectra data obtained through experimental measurements or simulations (Zou et al., 2023; Alberts et al., 2024). Therefore, incorporating the knowledge of energy levels into molecular representation learning is expected to facilitate the development of more informative molecular representations.
|
| 8 |
+
|
| 9 |
+

|
| 10 |
+
Figure 1: The conceptual view of MolSpectra, which leverages both molecular conformation and spectra for pre-training. Prior works only model classical mechanics by denoising on conformations.
|
| 11 |
+
|
| 12 |
+
In this paper, we propose MolSpectra, a framework that incorporates molecular spectra into the pre-training of 3D molecular representations, thereby infusing the knowledge of quantized energy level structures into the representations, as shown in Figure 1. In MolSpectra, we introduce a multispectrum encoder, SpecFormer, to capture both intra-spectrum and inter-spectrum peak correlations by training with a masked patches reconstruction (MPR) objective. Additionally, we employ a contrastive objective to distills the spectral features and its inherent knowledge into the learning of 3D representations. After pre-training, the resulting 3D encoder can be fine-tuned for downstream tasks, providing expressive 3D molecular representations without the need for associated spectral data. Extensive experiments over different downstream molecular property prediction benchmarks shows the superiority of MolSpectra.
|
| 13 |
+
|
| 14 |
+
In summary, our contributions are as follows:
|
| 15 |
+
|
| 16 |
+
• We introduce quantized energy level structures and molecular spectra into 3D molecular representation pre-training for the first time, surpassing previous work that relied solely on physical knowledge within the scope of classical mechanics.
|
| 17 |
+
• We propose SpecFormer as an expressive multi-spectrum encoder, along with the masked patches reconstruction objective for spectral representation learning.
|
| 18 |
+
• We propose a contrastive objective to align molecular representations in the 3D modality and spectral modalities, enabling the pre-trained 3D encoder to infer molecular spectral features in downstream tasks without relying on spectral data.
|
| 19 |
+
• Experiments across different downstream benchmarks demonstrate that our method effectively enhances the expressiveness of the pre-trained 3D molecular representations.
|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/2 PRELIMINARIES.md
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# 2 PRELIMINARIES
|
| 2 |
+
|
| 3 |
+
# 2.1 NOTATIONS
|
| 4 |
+
|
| 5 |
+
Consider a molecule characterized by its 3D structure and spectra, represented as M = (a, x, S). Here, a ∈ {1, 2, . . . , 118}N specifies the atomic numbers, indicating the types of atoms within the molecule. The vector x ∈ R3N describes the conformation of the molecule, while S represents its spectra. The parameter N denotes the number of atoms in the molecule. Note that the atoms are arranged in the same order in both a and x, ensuring consistency between the atomic numbers and their corresponding spatial coordinates.
|
| 6 |
+
|
| 7 |
+
S = (s1, . . . , s|S|) represents the set of spectra for a molecule, where |S| denotes the number of spectrum types considered. In our study, we focus on three types, so |S| = 3. The first spectrum, s1 ∈ R601, is the UV-Vis spectrum, which spans from 1.5 to 13.5 eV with 601 data points at intervals of 0.02 eV. The second spectrum, s2 ∈ R3501, is the IR spectrum, covering a range from 500 to 4000 cm−1 with 3501 data points at intervals of 1 cm−1. The third spectrum, s3 ∈ R3501, is the Raman spectrum, with the same range and intervals as the IR spectrum. Together, these spectra provide a comprehensive description of the molecular characteristics across different spectral modalities.
|
| 8 |
+
|
| 9 |
+
# 2.2 PRE-TRAINING 3D MOLECULAR REPRESENTATION VIA DENOISING
|
| 10 |
+
|
| 11 |
+
Denoising has emerged as a prominent pre-training objective in 3D molecular representation learning, excelling in various downstream tasks. This method involves training models to predict and remove noise introduced deliberately into molecular structures. This approach is physically interpretable due to its proven equivalence to learning the molecular force field.
|
| 12 |
+
|
| 13 |
+
Equivalence between denoising and learning molecular force fields. The equivalence between coordinate denoising and force field learning is established by Zaidi et al. (2023). For a given molecule M, perturb its equilibrium structure x0 according to the distribution p(x|x0), where x is the noisy conformation. Assuming the molecular distribution adheres to the energy-based Boltzmann distribution with respect to the energy function E(·), then
|
| 14 |
+
|
| 15 |
+

|
| 16 |
+
|
| 17 |
+
where GNNθ(x) denotes a graph neural network parameterized by θ, which processes the conformation x to produce node-level predictions. The notation ≃ signifies the equivalence of different objectives. The proof of this equivalence is provided in the Appendix A. In prior research, the energy function E(·) has been defined in several forms. Below are three representative studies.
|
| 18 |
+
|
| 19 |
+
Energy function I: mixture of isotropic Gaussians. In Coord (Zaidi et al., 2023), the energy function is approximated using a mixture of isotropic Gaussians centered at the known equilibrium structures to replace the Boltzmann distribution, since these structures are local maxima of the Boltzman distribution. Leveraging the equivalence between the score-matching objective and denoising autoencoders (Vincent, 2011), the following denoising-based energy function ECoord(·) is derived:
|
| 20 |
+
|
| 21 |
+

|
| 22 |
+
|
| 23 |
+
Note that this objective is derived under the assumption of isotropic Gaussian noise, i.e., p(x|x0) ∼ N (x0, τ 2c I3N ), where I3N represents the identity matrix of size 3N , and the subscript c indicates the coordinate denoising approach.
|
| 24 |
+
|
| 25 |
+
Energy function II: mixture of anisotropic Gaussians. Considering rigid and flexible components in molecular structures, isotropic Gaussian can lead to significant approximation errors. To address the anisotropic distribution, Frad (Feng et al., 2023) introduces hybrid noise on dihedral angles of rotatable bonds and atomic coordinates, incorporating fractional denoising of the coordinate noise. The equilibrium structure x0 is initially perturbed by dihedral angle noise p(ψa|ψ0) ∼ N (ψ0, σ2f Im), followed by coordinate noise p(x|xa) ∼ N (xa, τ 2f I3N ). Here, ψa, ψ0 ∈ [0, 2π)m represent to the dihedral angles of rotatable bonds in structures xa and x0, respectively, with m denoting the number of rotatable bonds. The subscript f indicates the fractional denoising approach. Subsequently, the energy function is induced:
|
| 26 |
+
|
| 27 |
+

|
| 28 |
+
|
| 29 |
+
where Στf ,σf = τ 2f I3N + σ2f CC⊤, and C ∈ R3N×m is a matrix used to linearly transform the dihedral angle noise into coordinate change, expressed as ∆x ≈ C∆ψ.
|
| 30 |
+
|
| 31 |
+
Energy function III: classical potential energy theory. SliDe (Ni et al., 2024) derives energy function from classical molecular potential energy theory (Alavi, 2020; Zhou & Liu, 2022). In this
|
| 32 |
+
|
| 33 |
+

|
| 34 |
+
Figure 2: Overview of the MolSpectra pre-training framework. Our pre-training framework comprises three sub-objectives: the denoising objective and the MPR objective, which respectively guide the representation learning of the 3D and spectral modalities, and the contrastive objective, which aligns the representations of both modalities.
|
| 35 |
+
|
| 36 |
+
form, the total intramolecular potential energy is mainly attributed to three types of interactions: bond stretching, bond angle bending, and bond torsion. The following energy function is derived:
|
| 37 |
+
|
| 38 |
+

|
| 39 |
+
|
| 40 |
+
where r ∈ (R≥0)m1 , θ ∈ [0, 2π)m2 , ϕ ∈ [0, 2π)m3 represent vectors of the bond lengths, bond angles, and bond torsion angles of the molecule, respectively. r0, θ0, ϕ0 correspond to the respective equilibrium values. The parameter vectors kB, kA, kT determine the interaction strength.
|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/3 THE PROPOSED MOLSPECTRA METHOD.md
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# 3 THE PROPOSED MOLSPECTRA METHOD
|
| 2 |
+
|
| 3 |
+
Considering the complementarity of different spectra, we introduce multiple spectra into molecular representation learning. To effectively comprehend molecular spectra, we designed a Transformerbased multi-spectrum encoder, SpecFormer, along with a masked reconstruction objective to guide its training. Finally, a contrastive objective is employed to align the 3D encoding guided by the denoising objective with the spectra encoding guided by the reconstruction objective, endowing the 3D encoding with the capability to understand spectra and the knowledge they encompass.
|
| 4 |
+
|
| 5 |
+
# 3.1 SPECFORMER: A SINGLE-STREAM ENCODER FOR MULTI-MODAL ENERGY SPECTRA
|
| 6 |
+
|
| 7 |
+
For different types of spectra, each spectrum is independently patched and initially encoded. Then, all the resulting patch embeddings are concatenated and encoded using a Transformer-based encoder.
|
| 8 |
+
|
| 9 |
+
Patching. Compared to directly encoding individual frequency points, we divided each spectrum into multiple patches. This approach offers two distinct advantages: (i) By forming patches from adjacent frequency points, local semantic features, such as absorption peaks, can be captured more effectively. (ii) It reduces the computational overhead of subsequent Transformer layers. Technically, each spectrum si ∈ RLi where i = 1, · · · , |S| is first divided into patches according to the patch length Pi and the stride Di. When 0 < Di < Pi, the consecutive patches will be overlapped with overlapping region length Pi − Di. When Di = Pi, the consecutive patches will be nonoverlapped. Li denotes the length of si. The patching process on each spectrum will generate a sequence of patches pi ∈ RNi×Pi, where Ni = Li−PiDi + 1 is the number of patches.
|
| 10 |
+
|
| 11 |
+
Patch encoding and position encoding. Prior to be fed into the encoder, the patches of the i-th spectrum are mapped to the latent space of dimension d via a trainable linear projection Wi ∈ RPi×d. A learnable additive position encoding W posi ∈ RNi×d is applied to maintain the order of the patches: p′i = piWi + W posi , where p′i ∈ RNi×d denotes the latent representation of the spectrum si that will be fed into the subsequent SpecFormer encoder.
|
| 12 |
+
|
| 13 |
+
SpecFormer: multi-spectrum Transformer encoder. Although several encoders have been proposed to map molecular spectrum into implicit representations, such as the CNN-AM (Tao et al., 2024) based on one-dimensional convolution, these encoders are designed to encode only a single type of spectrum. In our approach, multiple molecular spectra (UV-Vis, IR, Raman) are jointly considered. When encoding multiple spectra of a molecule simultaneously, an observation caught our attention and led us to adopt a Transformer-based encoder with multiple spectra as input, similar to the single-stream Transformer in multi-modal learning (Shin et al., 2021).
|
| 14 |
+
|
| 15 |
+
The observation refers to the fact that the same functional group not only causes multiple peaks within a single spectrum, but also generates peaks across different spectra. As shown on the left of Figure 3, the different vibrational modes of the methyl group (-CH3) in methanol (CH3OH) result in three peaks in the IR spectrum, indicating intra-spectrum dependencies among these peaks. A similar phenomenon occurs with the hydroxyl group (-OH) in methanol. Additionally, the aromatic ring in phenol (C6H5OH), shown on the right of
|
| 16 |
+
|
| 17 |
+

|
| 18 |
+
Figure 3: Illustration of intra-spectrum (left) and interspectrum (right) dependencies.
|
| 19 |
+
|
| 20 |
+
Figure 3, not only produces multiple peaks in the IR spectrum due to different vibrational modes but also causes an absorption peak near 270 nm in the UV-Vis spectrum due to the π → π∗ transition in the aromatic ring, demonstrating the existence of inter-spectrum dependencies. Such dependencies have been theoretically studied, for example, in the context of vibronic coupling (Kong et al., 2021).
|
| 21 |
+
|
| 22 |
+
To capture intra-spectrum and inter-spectrum dependencies, we concatenate the embeddings obtained from patch encoding and position encoding of different spectra: pˆ = p′1∥ · · · ∥p′|S| ∈ R(P|S|i=1 Ni)×d, and then input them into the Transformer encoder as depicted in Figure 2. Then each head h = 1, . . . , H in multi-head attention will transform them into query matrices Qh = pWˆ Qh , key matrices Kh = pWˆ Kh and value matrices Vh = pWˆ Vh , where W Qh , W Kh ∈ Rd×dk and WVh ∈ Rd× dH . Afterward, a scaled product is utilized to obtain the attention output Oh ∈ R(P|S|i=1 Ni)× dH :
|
| 23 |
+
|
| 24 |
+

|
| 25 |
+
|
| 26 |
+
The multi-head attention block also includes BatchNorm layers and a feed forward network with residual connections as shown in Figure 2. After combining the outputs of all heads, it generates the representation denoted as z ∈ R(P|S|i=1 Ni)×d. Finally, a flatten layer with representation projection head is used to obtain the molecular spectra representation zs ∈ Rd.
|
| 27 |
+
|
| 28 |
+
# 3.2 MASKED PATCHES RECONSTRUCTION PRE-TRAINING FOR SPECTRA
|
| 29 |
+
|
| 30 |
+
Before distilling the spectra information into 3D molecular representation learning, we need first ensure that the spectrum encoder can effectively comprehend molecular spectra and generate spectral representations. Considering the success of masking modeling across various domains (Devlin et al., 2019; He et al., 2022; Hou et al., 2022; Xia et al., 2023; Wang et al., 2024b; Nie et al., 2023), we propose a masked patches reconstruction (MPR) objective to guide the training of SpecFormer.
|
| 31 |
+
|
| 32 |
+
After the patching step, we randomly select a portion of patches according to the mask ratio α and replace them with zero vectors to implement the masking. Subsequently, the masked patches undergo patch encoding and position encoding. In this way, the semantics of the masked patches (the absorption intensity at specific wavelengths) are obscured during patch encoding, while the positional information is retained to facilitate the reconstruction of the original semantics.
|
| 33 |
+
|
| 34 |
+
After encoding by SpecFormer, the encoded results corresponding to the masked patches are input into a spectrum-specific reconstruction head to reconstruct the original spectral values that were masked. The mean squared error (MSE) between the reconstruction results and the original masked spectra serves as the loss function for the MPR task, guiding the training of SpecFormer:
|
| 35 |
+
|
| 36 |
+

|
| 37 |
+
|
| 38 |
+
where Pi denotes the set of masked patches in the i-th type of molecular spectra, and pˆi,j denotes ethe reconstructed patch corresponding to the masked patch pi,j .
|
| 39 |
+
|
| 40 |
+
# 3.3 CONTRASTIVE LEARNING BETWEEN 3D STRUCTURES AND SPECTRA
|
| 41 |
+
|
| 42 |
+
Under the guidance of the denoising objective for 3D representation learning and the MPR objective for spectral representation learning, we further introduce a contrastive objective to align the representations across these two modalities. We treat the 3D representation zx ∈ Rd and spectral representation zs ∈ Rd of the same molecule as positive samples, and negative samples otherwise. Subsequently, the consistency between positive samples and the discrepancy between negative samples are maximized through the contrastive objective. Given the theoretical and empirical effectiveness, we employ InfoNCE (van den Oord et al., 2018) as the contrastive objective:
|
| 43 |
+
|
| 44 |
+

|
| 45 |
+
|
| 46 |
+
where zjx, zjs are randomly sampled 3D and spectra views regarding to the positive pair (zx, zs). fx(zx, zs) and fs(zs, zx) are scoring functions for the two corresponding views, with flexible formulations. Here we adopt fx(zx, zs) = fs(zs, zx) = ⟨zx, zs⟩.
|
| 47 |
+
|
| 48 |
+
Note that the denoising objective can utilize any form from existing 3D molecular representation pre-training studies, enabling seamless integration of our method into these frameworks.
|
| 49 |
+
|
| 50 |
+
# 3.4 TWO-STAGE PRE-TRAINING PIPELINE
|
| 51 |
+
|
| 52 |
+
Previous pre-training efforts for 3D molecular representation have been conducted on unlabeled datasets using denoising objective. These datasets typically provide only equilibrium 3D structures without offering spectra for all molecules. To enhance the pre-training effect by incorporating spectra while leveraging denoising pre-training, we employ a two-stage pre-training approach. The first stage involves training on a larger dataset (Nakata & Shimazaki, 2017) without spectra using only the denoising objective. Subsequently, the second stage involves training on a dataset that includes spectra using the complete objective as follows:
|
| 53 |
+
|
| 54 |
+

|
| 55 |
+
|
| 56 |
+
where βDenoising, βMPR, and βContrast denote the weights of each sub-objective.
|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/4 EXPERIMENTS.md
ADDED
|
@@ -0,0 +1,73 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# 4 EXPERIMENTS
|
| 2 |
+
|
| 3 |
+
To comprehensively evaluate the impact of molecular spectra on molecular tasks, we first verify the effectiveness of molecular spectra in the training-from-scratch method for the downstream task. Furthermore, we evaluate the effectiveness of our pre-training framework MolSpectra.
|
| 4 |
+
|
| 5 |
+
# 4.1 EFFECTIVENESS OF MOLECULAR SPECTRA IN TRAINING FROM SCRATCH
|
| 6 |
+
|
| 7 |
+
This pilot experiment aims to demonstrate the rationality for incorporating molecular spectra into pre-training. We introduce additional spectral features into a train-from-scratch molecular property prediction model to observe the impact of spectral information on prediction outcomes. We employ EGNN (Satorras et al., 2021), a representative 3D molecular encoder, equipped with an MLP-based prediction head as the baseline model. While EGNN encodes the 3D representations, the UV-Vis spectrum of each molecule provided by the QM9S (Zou et al., 2023) dataset is encoded into spectral representations by a spectrum encoder. Before making predictions with the final MLP, we concatenate the spectral and 3D representations for prediction. The results are presented in Table 1.
|
| 8 |
+
|
| 9 |
+
Table 1: Performance (MAE ↓) when training from scratch on QM9 dataset.
|
| 10 |
+

|
| 11 |
+
|
| 12 |
+
We observe that by directly concatenating spectral representations, the performance of molecular property prediction can be effectively enhanced. This indicates that the information from molecular spectra is beneficial for downstream molecular property prediction. Further incorporating molecular spectra into the pre-training phase of molecular representation has the potential to enhance the informativeness and generalization capability of the representations, thereby broadly improving the performance of downstream tasks.
|
| 13 |
+
|
| 14 |
+
# 4.2 EFFECTIVENESS OF MOLECULAR SPECTRA IN REPRESENTATION PRE-TRAINING
|
| 15 |
+
|
| 16 |
+
We conduct experiments to evaluate MolSpectra by first introducing spectral data into the pretraining of 3D representations, followed by evaluating the performance on downstream tasks. For a comprehensive comparison, two types of baselines are adopted: (1) training-from-scratch methods, including SchNet (Schutt et al. ¨ , 2017), EGNN, DimeNet (Klicpera et al., 2020b), DimeNet++ (Klicpera et al., 2020a), PaiNN (Schutt et al. ¨ , 2021), SphereNet (Liu et al., 2021), and TorchMD-Net (Tholke & Fabritiis ¨ , 2022); and (2) pre-training methods, including Transformer-M (Luo et al., 2023), SE(3)-DDM (Liu et al., 2023b), 3D-EMGP (Jiao et al., 2023), and Coord.
|
| 17 |
+
|
| 18 |
+
MolSpectra can be seamlessly plugged into any existing denoising method. To evaluate the enhancement provided by our method compared to denoising alone, we select the representative coordinate denoising (Coord) as our denoising sub-objective. This method also serves as our primary baseline.
|
| 19 |
+
|
| 20 |
+
# 4.2.1 PRE-TRAINING DATASET.
|
| 21 |
+
|
| 22 |
+
As described in Section 3.4, we first perform denoising pre-training on the PCQM4Mv2 (Nakata & Shimazaki, 2017) dataset, followed by a second stage of pre-training on the QM9Spectra (QM9S) (Zou et al., 2023) dataset, which includes multi-modal molecular energy spectra. In both stages, we adopt the denoising objective provided by Coord (Zaidi et al., 2023), as defined in Eq. 2.
|
| 23 |
+
|
| 24 |
+
The QM9S dataset comprises organic molecules from the QM9 (Ramakrishnan et al., 2014) dataset. The UV-Vis, IR, and Raman spectra of the molecules are calculated at the B3LYP/def-TZVP level of theory, through frequency analysis and time-dependent density functional theory (TD-DFT).
|
| 25 |
+
|
| 26 |
+
# 4.2.2 QM9
|
| 27 |
+
|
| 28 |
+
The QM9 dataset is a quantum chemistry dataset comprising over 134,000 small molecules, each consisting of up to 9 heavy atoms (C, N, O, F) and additional H atoms. This dataset provides an equilibrium geometric conformation for each molecule along with 12 property labels. The dataset is divided into a training set of 110k molecules, a validation set of 10k molecules, and a test set containing the remaining over 10k molecules. Prediction errors are measured using the mean absolute error (MAE). The experimental results are presented in Table 2.
|
| 29 |
+
|
| 30 |
+
The 3D molecular representations pre-trained using our method are fine-tuned and used for prediction across various properties, achieving state-of-the-art performance in 8 out of 12 properties and outperforms Coord in 10 out of 12 properties. In conjunction with the observations in Section 4.1, the performance improvement can be attributed to our incorporation of an understanding of molecular spectra and the knowledge they entail into the 3D molecular representations.
|
| 31 |
+
|
| 32 |
+
Table 2: Performance (MAE↓) on QM9 dataset. The compared methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are highlighted in bold.
|
| 33 |
+

|
| 34 |
+
|
| 35 |
+
Table 3: Performance (MAE↓) on MD17 force prediction (kcal/mol/ ˚A). The methods are divided into two groups: training from scratch and pre-training then fine-tuning. The best results are in bold.
|
| 36 |
+

|
| 37 |
+
|
| 38 |
+
# 4.2.3 MD17
|
| 39 |
+
|
| 40 |
+
The MD17 dataset contains molecular dynamics trajectories for eight organic molecules, including aspirin, benzene, and ethanol. It offers 150k to nearly 1M conformations per molecule, with energy and force labels. Unlike QM9, MD17 emphasizes dynamic behavior in addition to static properties. We use a standard limited data split: models train on 1k samples, validate on 50, and test on the rest. Performance is evaluated using MAE, with results in Table 3.
|
| 41 |
+
|
| 42 |
+
Our approach also results in the expected performance improvement on MD17. MD17 is a dataset comprising a large number of non-equilibrium molecular structures and their corresponding force fields, which serves to evaluate a model’s understanding of molecular dynamics. However, previous pre-training methods based solely on denoising have only learned force field patterns at static equilibrium states, failing to adequately capture the dynamic evolution of molecular systems. In contrast, our MolSpectra learns the dynamic evolution of molecules by understanding energy level transition patterns, thereby outperforming denoising-based pre-training methods.
|
| 43 |
+
|
| 44 |
+
4.3 SENSITIVITY ANALYSIS OF PATCH LENGTH Pi, STRIDE Di, AND MASK RATIO α
|
| 45 |
+
|
| 46 |
+
We conduct experiments to evaluate the impact of patch length Pi, stride Di, and mask ratio α.
|
| 47 |
+
Results are summarized in Table 4 and Table 5.
|
| 48 |
+
|
| 49 |
+
From Table 4, we observe that when consecutive patches have overlap (Di < Pi), the performance of pre-training is superior compared to scenarios without overlap (Di = Pi). Specifically, the performance is optimal when the stride is half of the patch length. This is because appropriate overlap can better preserve and capture local features, particularly the information at the patch boundaries. Additionally, we find that choosing an appropriate patch length further enhances performance. In our experiments, the configuration of Pi = 20, Di = 10 yields the best results.
|
| 50 |
+
|
| 51 |
+
Table 4: Sensitivity of patch length and stride.
|
| 52 |
+

|
| 53 |
+
|
| 54 |
+
Table 5: Sensitivity of mask ratio.
|
| 55 |
+

|
| 56 |
+
|
| 57 |
+
Regarding the mask ratio, α = 0.10 is a preferable choice. A small mask ratio results in insufficient MPR optimization, hindering SpecFormer training. Conversely, a large mask ratio causes excessive spectral perturbation, degrading performance when aligning with the 3D representations with the contrastive objective. An appropriate mask ratio strikes a balance between these two aspects.
|
| 58 |
+
|
| 59 |
+
# 4.4 ABLATION STUDY
|
| 60 |
+
|
| 61 |
+
To rigorously demonstrate the contributions of masked patches reconstruction, the incorporation of molecular spectra, and each spectral modality, we conducted an ablation study on them.
|
| 62 |
+
|
| 63 |
+
Ablation study of masked patches reconstruction. We remove the MPR loss to analyze the impact of masked patches reconstruction, referred to as “w/o MPR” in Table 6. Removing the MPR objective leads to performance deterioration. This is consistent with the sensitivity analysis of the mask ratio α in Section 4.3, as removing MPR is an extreme case where α = 0. This decline is due to the lack of effective guidance in training SpecFormer. Using an undertrained SpecFormer for contrastive learning with 3D encoder outputs limits performance improvement.
|
| 64 |
+
|
| 65 |
+
Table 6: Ablation of optimization objectives.
|
| 66 |
+

|
| 67 |
+
|
| 68 |
+
Ablation study of molecular spectra. We retain only the denoising loss, removing both the MPR loss and contrastive loss, referred to as “w/o MPR, Contrast” in Table 6. The only difference between this variant and MolSpectra is the incorporation of molecular spectra into the pre-training. The ”w/o MPR, Contrast” results are inferior to those of MolSpectra, highlighting that incorporating molecular spectra effectively enhances the quality and generalizability of molecular 3D representations.
|
| 69 |
+
|
| 70 |
+
Ablation study of each spectral modality. To evaluate the contributions of each spectral modality to the performance, we conduct an ablation study for each modality. The results are presented in Table 7. It can be observed that each spectral modality contributes differently, with the UV-Vis spectrum having the smallest contribution and the IR spectrum the largest, likely due to the varying information content in each modality.
|
| 71 |
+
|
| 72 |
+
Table 7: Ablation of spectral modalities.
|
| 73 |
+

|
mineru_outputs/6027_MolSpectra_Pre_training_3/section_split_output/6 CONCLUSION.md
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# 6 CONCLUSION
|
| 2 |
+
|
| 3 |
+
In this study, we explore pre-training molecular 3D representations beyond classical mechanics. By leveraging the correlation between molecular energy level structures and molecular spectra in quantum mechanics, we introduce molecular spectra for pre-training molecular 3D representations (MolSpectra). By aligning the 3D encoder trained with a denoising objective and the spectrum encoder trained with a masked patch reconstruction objective, we enhance the informativeness and transferability of the resulting 3D representations.
|