Dataset Viewer
Auto-converted to Parquet Duplicate
authors
stringlengths
11
2.41k
title
stringlengths
38
184
journal-ref
stringclasses
115 values
doi
stringlengths
17
34
report-no
stringclasses
3 values
categories
stringlengths
17
83
abstract
stringlengths
124
1.92k
versions
stringlengths
62
689
update_date
stringdate
2007-09-13 00:00:00
2025-05-15 00:00:00
Z. Suszynski, M. Kosikowski, R. Duer
The application of Artificial Neural Network for the assessment of thermal properties of multi-layer semiconductor structure
Dans Proceedings of 12th International Workshop on Thermal investigations of ICs - THERMINIC 2006, Nice : France (2006)
null
null
cond-mat.mtrl-sci
In this paper, the solution of the problem of identification of thermal properties of investigated multi-layer structure is presented. In order of that, artificial neural network was used to find the set of thermal properties for which the complex contrast characteric derived fits the best to the one evaluated basing upon experimenatal data.
[{'version': 'v1', 'created': 'Wed, 12 Sep 2007 13:30:55 GMT'}]
2007-09-13
Gabriele C. Sosso (1), Giacomo Miceli (1), Sebastiano Caravati (2), J\"org Behler (3), and Marco Bernasconi (1) ((1) Dipartimento di Scienza dei Materiali, Universit\`a di Milano-Bicocca, Milano, Italy, (2) Computational Science, Department of Chemistry and Applied Biosciences ETH Zurich, USI Campus, Lugano, Switzerland, (3) Lehrstuhl f\"ur Theoretische Chemie, Ruhr-Universit\"at Bochum, Bochum, Germany)
A neural network interatomic potential for the phase change material GeTe
Phys. Rev. B 85, 174103 (2012)
10.1103/PhysRevB.85.174103
null
cond-mat.mtrl-sci cond-mat.dis-nn
GeTe is a prototypical phase change material of high interest for applications in optical and electronic non-volatile memories. We present an interatomic potential for the bulk phases of GeTe, which is created using a neural network (NN) representation of the potential-energy surface obtained from reference calculations based on density functional theory. It is demonstrated that the NN potential provides a close to ab initio quality description of a number of properties of liquid, crystalline and amorphous GeTe. The availability of a reliable classical potential allows addressing a number of issues of interest for the technological applications of phase change materials, which are presently beyond the capability of first principles molecular dynamics simulations.
[{'version': 'v1', 'created': 'Tue, 10 Jan 2012 11:35:59 GMT'}]
2012-08-02
Francesco Bonanno, Giacomo Capizzi, Grazia Lo Sciuto, Christian Napoli, Giuseppe Pappalardo, Emiliano Tramontana
A Cascade Neural Network Architecture investigating Surface Plasmon Polaritons propagation for thin metals in OpenMP
International conference on Artificial Intelligence and Soft Computing (ICAISC 2014), Vol I, 22-33 (2014)
null
null
cs.NE cond-mat.mes-hall cond-mat.mtrl-sci cs.DC cs.LG
Surface plasmon polaritons (SPPs) confined along metal-dielectric interface have attracted a relevant interest in the area of ultracompact photonic circuits, photovoltaic devices and other applications due to their strong field confinement and enhancement. This paper investigates a novel cascade neural network (NN) architecture to find the dependance of metal thickness on the SPP propagation. Additionally, a novel training procedure for the proposed cascade NN has been developed using an OpenMP-based framework, thus greatly reducing training time. The performed experiments confirm the effectiveness of the proposed NN architecture for the problem at hand.
[{'version': 'v1', 'created': 'Thu, 12 Jun 2014 08:40:04 GMT'}]
2014-06-13
S. Alireza Ghasemi, Albert Hofstetter, Santanu Saha, Stefan Goedecker
Interatomic potentials for ionic systems with density functional accuracy based on charge densities obtained by a neural network
Phys. Rev. B 92, 045131 (2015)
10.1103/PhysRevB.92.045131
null
cond-mat.mtrl-sci physics.chem-ph
Based on an analysis of the short range chemical environment of each atom in a system, standard machine learning based approaches to the construction of interatomic potentials aim at determining directly the central quantity which is the total energy. This prevents for instance an accurate description of the energetics of systems where long range charge transfer is important as well as of ionized systems. We propose therefore not to target directly with machine learning methods the total energy but an intermediate physical quantity namely the charge density, which then in turn allows to determine the total energy. By allowing the electronic charge to distribute itself in an optimal way over the system, we can describe not only neutral but also ionized systems with unprecedented accuracy. We demonstrate the power of our approach for both neutral and ionized NaCl clusters where charge redistribution plays a decisive role for the energetics. We are able to obtain chemical accuracy, i.e. errors of less than a milli Hartree per atom compared to the reference density functional results. The introduction of physically motivated quantities which are determined by the short range atomic environment via a neural network leads also to an increased stability of the machine learning process and transferability of the potential.
[{'version': 'v1', 'created': 'Thu, 29 Jan 2015 05:19:01 GMT'}]
2015-08-05
Samad Hajinazar, Junping Shao, Aleksey N. Kolmogorov
Stratified construction of neural network based interatomic models for multicomponent materials
Phys. Rev. B 95, 014114 (2017)
10.1103/PhysRevB.95.014114
null
cond-mat.mtrl-sci
Recent application of neural networks (NNs) to modeling interatomic interactions has shown the learning machines' encouragingly accurate performance for select elemental and multicomponent systems. In this study, we explore the possibility of building a library of NN-based models by introducing a hierarchical NN training. In such a stratified procedure NNs for multicomponent systems are obtained by sequential training from the bottom up: first unaries, then binaries, and so on. Advantages of constructing NN sets with shared parameters include acceleration of the training process and intact description of the constituent systems. We use an automated generation of diverse structure sets for NN training on density functional theory-level reference energies. In the test case of Cu, Pd, Ag, Cu-Pd, Cu-Ag, Pd-Ag, and Cu-Pd-Ag systems, NNs trained in the traditional and stratified fashions are found to have essentially identical accuracy for defect energies, phonon dispersions, formation energies, etc. The models' robustness is further illustrated via unconstrained evolutionary structure searches in which the NN is used for the local optimization of crystal unit cells.
[{'version': 'v1', 'created': 'Tue, 27 Sep 2016 14:06:16 GMT'}, {'version': 'v2', 'created': 'Wed, 1 Feb 2017 01:15:44 GMT'}]
2017-02-08
P. Anees, M. C. Valsakumar and B. K. Panigrahi
Delineating the role of ripples on thermal expansion of honeycomb materials:graphene, 2D-h-BN and monolayer(ML)-MoS2
Phys. Chem. Chem. Phys., 19, 10518, (2017)
10.1039/C6CP08635G
null
cond-mat.mtrl-sci
We delineated the role of thermally excited ripples on thermal expansion properties of 2D honeycomb materials (free-standing graphene, 2D h-BN, and ML-MoS2), by explicitly carrying out three-dimensional (3D) and two-dimensional (2D) molecular dynamics simulations. In 3D simulations, the in-plane lattice parameter (a-lattice) of graphene and 2D h-BN shows thermal contraction over a wide range of temperatures and exhibits a strong system size dependence. The 2D simulations of the very same system show a reverse trend, where the a-lattice is expanding in the whole computed temperature range. Contrary to graphene and 2D h-BN, the a-lattice of ML-MoS2 shows thermal expansion in both 2D and 3D simulations and their system size dependence is marginal. By analyzing the phonon dispersion at 300 K, we found that the discrepancy between 2D and 3D simulations of graphene and 2D h-BN is due to the absence of out-of-plane bending mode (ZA) in 2D simulations, which is responsible for thermal contraction of a-lattice at low temperature. Meanwhile, all the phonon modes are present in 2D phonon dispersion of ML-MoS2, which indicates that the origin of ZA mode is not purely due to out-of-plane movement of atoms and also its effect on thermal expansion is not significant as found in graphene and 2D h-BN
[{'version': 'v1', 'created': 'Tue, 25 Oct 2016 08:43:49 GMT'}]
2017-07-25
Daniel Valencia, Evan Wilson, Zhengping Jiang, Gustavo A. Valencia-Zapata, Gerhard Klimeck and Michael Povolotskyi
Grain Boundary Resistance in Copper Interconnects from an Atomistic Model to a Neural Network
Phys. Rev. Applied 9, 044005 (2018)
10.1103/PhysRevApplied.9.044005
null
cond-mat.mtrl-sci
Orientation effects on the resistivity of copper grain boundaries are studied systematically with two different atomistic tight binding methods. A methodology is developed to model the resistivity of grain boundaries using the Embedded Atom Model, tight binding methods and non-equilibrum Green's functions (NEGF). The methodology is validated against first principles calculations for small, ultra-thin body grain boundaries (<5nm) with 6.4% deviation in the resistivity. A statistical ensemble of 600 large, random structures with grains is studied. For structures with three grains, it is found that the distribution of resistivities is close to normal. Finally, a compact model for grain boundary resistivity is constructed based on a neural network.
[{'version': 'v1', 'created': 'Tue, 17 Jan 2017 23:24:10 GMT'}, {'version': 'v2', 'created': 'Sat, 4 Feb 2017 04:16:11 GMT'}, {'version': 'v3', 'created': 'Sun, 8 Oct 2017 20:03:43 GMT'}]
2018-04-11
Kyle Mills, Michael Spanner, and Isaac Tamblyn
Deep learning and the Schr\"odinger equation
Phys. Rev. A 96, 042113 (2017)
10.1103/PhysRevA.96.042113
null
cond-mat.mtrl-sci cs.LG physics.chem-ph
We have trained a deep (convolutional) neural network to predict the ground-state energy of an electron in four classes of confining two-dimensional electrostatic potentials. On randomly generated potentials, for which there is no analytic form for either the potential or the ground-state energy, the neural network model was able to predict the ground-state energy to within chemical accuracy, with a median absolute error of 1.49 mHa. We also investigate the performance of the model in predicting other quantities such as the kinetic energy and the first excited-state energy of random potentials.
[{'version': 'v1', 'created': 'Sun, 5 Feb 2017 02:58:58 GMT'}, {'version': 'v2', 'created': 'Thu, 8 Jun 2017 20:39:27 GMT'}, {'version': 'v3', 'created': 'Fri, 3 Nov 2017 13:10:51 GMT'}]
2017-11-06
Seyed Majid Azimi, Dominik Britz, Michael Engstler, Mario Fritz, Frank M\"ucklich
Advanced Steel Microstructural Classification by Deep Learning Methods
null
10.1038/s41598-018-20037-5
null
cs.CV cond-mat.mtrl-sci
The inner structure of a material is called microstructure. It stores the genesis of a material and determines all its physical and chemical properties. While microstructural characterization is widely spread and well known, the microstructural classification is mostly done manually by human experts, which gives rise to uncertainties due to subjectivity. Since the microstructure could be a combination of different phases or constituents with complex substructures its automatic classification is very challenging and only a few prior studies exist. Prior works focused on designed and engineered features by experts and classified microstructures separately from the feature extraction step. Recently, Deep Learning methods have shown strong performance in vision applications by learning the features from data together with the classification step. In this work, we propose a Deep Learning method for microstructural classification in the examples of certain microstructural constituents of low carbon steel. This novel method employs pixel-wise segmentation via Fully Convolutional Neural Networks (FCNN) accompanied by a max-voting scheme. Our system achieves 93.94% classification accuracy, drastically outperforming the state-of-the-art method of 48.89% accuracy. Beyond the strong performance of our method, this line of research offers a more robust and first of all objective way for the difficult task of steel quality appreciation.
[{'version': 'v1', 'created': 'Tue, 20 Jun 2017 14:29:42 GMT'}, {'version': 'v2', 'created': 'Thu, 15 Feb 2018 14:30:16 GMT'}]
2018-02-16
Weizong Xu and James M. LeBeau
A Deep Convolutional Neural Network to Analyze Position Averaged Convergent Beam Electron Diffraction Patterns
null
10.1016/j.ultramic.2018.03.004
null
physics.data-an cond-mat.mtrl-sci
We establish a series of deep convolutional neural networks to automatically analyze position averaged convergent beam electron diffraction patterns. The networks first calibrate the zero-order disk size, center position, and rotation without the need for pretreating the data. With the aligned data, additional networks then measure the sample thickness and tilt. The performance of the network is explored as a function of a variety of variables including thickness, tilt, and dose. A methodology to explore the response of the neural network to various pattern features is also presented. Processing patterns at a rate of $\sim$0.1 s/pattern, the network is shown to be orders of magnitude faster than a brute force method while maintaining accuracy. The approach is thus suitable for automatically processing big, 4D STEM data. We also discuss the generality of the method to other materials/orientations as well as a hybrid approach that combines the features of the neural network with least squares fitting for even more robust analysis. The source code is available at https://github.com/subangstrom/DeepDiffraction.
[{'version': 'v1', 'created': 'Thu, 3 Aug 2017 14:38:30 GMT'}]
2018-06-05
A. Ziletti, D. Kumar, M. Scheffler, L. M. Ghiringhelli
Insightful classification of crystal structures using deep learning
Nature Communications 9, 2775 (2018)
10.1038/s41467-018-05169-6
null
cond-mat.mtrl-sci cond-mat.dis-nn
Computational methods that automatically extract knowledge from data are critical for enabling data-driven materials science. A reliable identification of lattice symmetry is a crucial first step for materials characterization and analytics. Current methods require a user-specified threshold, and are unable to detect average symmetries for defective structures. Here, we propose a machine-learning-based approach to automatically classify structures by crystal symmetry. First, we represent crystals by calculating a diffraction image, then construct a deep-learning neural-network model for classification. Our approach is able to correctly classify a dataset comprising more than 100 000 simulated crystal structures, including heavily defective ones. The internal operations of the neural network are unraveled through attentive response maps, demonstrating that it uses the same landmarks a materials scientist would use, although never explicitly instructed to do so. Our study paves the way for crystal-structure recognition of - possibly noisy and incomplete - three-dimensional structural data in big-data materials science.
[{'version': 'v1', 'created': 'Thu, 7 Sep 2017 15:09:27 GMT'}, {'version': 'v2', 'created': 'Wed, 30 May 2018 06:11:23 GMT'}]
2018-07-19
M. Carrillo, J. A. Gonz\'alez, S. Hern\'andez-Ortiz, C. E. L\'opez, A. Raya
Bloch oscillations in graphene from an artificial neural network study
null
null
null
cond-mat.mtrl-sci cond-mat.mes-hall
We develop an artificial neural network (ANN) approach to classify simulated signals corrsponding to the semi-classical description of Bloch oscillations in pristine graphene. After the ANN is properly trained, we consider the inverse problem of Bloch oscillations (BO),namely, a new signal is classified according to the external electric field strength oriented along either the zig-zag or arm-chair edges of the graphene membrane, with a correct classification that ranges from 82.6% to 99.3% depending on the accuracy of the predicted electric field. This approach can be improved depending on the time spent in training the network and the computational power available. Findings in this work can be straightforwardly extended to a variety of Dirac-Weyl materials.
[{'version': 'v1', 'created': 'Wed, 4 Oct 2017 17:46:07 GMT'}]
2017-10-05
Adrien Bouhon and Annica M. Black-Schaffer
Bulk topology of line-nodal structures protected by space group symmetries in class AI
null
null
null
cond-mat.mtrl-sci
We give an exhaustive characterization of the topology of band structures in class AI, using nonsymmorphic space group 33 ($Pna2_1$) as a representative example where a great variety of symmetry protected line-nodal structures can be formed. We start with the topological classification of all line-nodal structures given through the combinatorics of valence irreducible representations (IRREPs) at a few high-symmetry points (HSPs) at a fixed filling. We decompose the total topology of nodal valence band bundles through the local topology of elementary (i.e. inseparable) nodal structures and the global topology that constrains distinct elementary nodal elements over the Brillouin zone (BZ). Generalizing from the cases of simple point nodes and simple nodal lines (NLs), we argue that the local topology of every elementary nodal structure is characterized by a set of poloidal-toroidal charges, one monopole, and one thread charge (when threading the BZ torus), while the global topology only allows pairs of nontrivial monopole and thread charges. We show that all these charges are given in terms of symmetry protected topological invariants, defined through quantized Wilson loop phases over symmetry constrained momentum loops, which we derive entirely algebraically from the valence IRREPs at the HSPs. In particular, we find highly connected line-nodal structures, line-nodal monopole pairs, and line-nodal thread pairs, that are all protected by the unitary crystalline symmetries only. Furthermore, we show symmetry preserving topological Lifshitz transitions through which independent NLs can be connected, disconnected, or linked. Our work constitutes a heuristic approach to the systematic topological classification and characterization of all momentum space line-nodal structures protected by space group symmetries in class AI.
[{'version': 'v1', 'created': 'Fri, 13 Oct 2017 11:10:07 GMT'}]
2017-10-16
Maxwell L. Hutchinson, Erin Antono, Brenna M. Gibbons, Sean Paradiso, Julia Ling, Bryce Meredig
Overcoming data scarcity with transfer learning
null
null
null
cs.LG cond-mat.mtrl-sci stat.ML
Despite increasing focus on data publication and discovery in materials science and related fields, the global view of materials data is highly sparse. This sparsity encourages training models on the union of multiple datasets, but simple unions can prove problematic as (ostensibly) equivalent properties may be measured or computed differently depending on the data source. These hidden contextual differences introduce irreducible errors into analyses, fundamentally limiting their accuracy. Transfer learning, where information from one dataset is used to inform a model on another, can be an effective tool for bridging sparse data while preserving the contextual differences in the underlying measurements. Here, we describe and compare three techniques for transfer learning: multi-task, difference, and explicit latent variable architectures. We show that difference architectures are most accurate in the multi-fidelity case of mixed DFT and experimental band gaps, while multi-task most improves classification performance of color with band gaps. For activation energies of steps in NO reduction, the explicit latent variable method is not only the most accurate, but also enjoys cancellation of errors in functions that depend on multiple tasks. These results motivate the publication of high quality materials datasets that encode transferable information, independent of industrial or academic interest in the particular labels, and encourage further development and application of transfer learning methods to materials informatics problems.
[{'version': 'v1', 'created': 'Thu, 2 Nov 2017 12:54:51 GMT'}]
2017-11-15
Eric Gossett, Cormac Toher, Corey Oses, Olexandr Isayev, Fleur Legrain, Frisco Rose, Eva Zurek, Jes\'us Carrete, Natalio Mingo, Alexander Tropsha, Stefano Curtarolo
AFLOW-ML: A RESTful API for machine-learning predictions of materials properties
null
null
null
cond-mat.mtrl-sci physics.comp-ph
Machine learning approaches, enabled by the emergence of comprehensive databases of materials properties, are becoming a fruitful direction for materials analysis. As a result, a plethora of models have been constructed and trained on existing data to predict properties of new systems. These powerful methods allow researchers to target studies only at interesting materials $\unicode{x2014}$ neglecting the non-synthesizable systems and those without the desired properties $\unicode{x2014}$ thus reducing the amount of resources spent on expensive computations and/or time-consuming experimental synthesis. However, using these predictive models is not always straightforward. Often, they require a panoply of technical expertise, creating barriers for general users. AFLOW-ML (AFLOW $\underline{\mathrm{M}}$achine $\underline{\mathrm{L}}$earning) overcomes the problem by streamlining the use of the machine learning methods developed within the AFLOW consortium. The framework provides an open RESTful API to directly access the continuously updated algorithms, which can be transparently integrated into any workflow to retrieve predictions of electronic, thermal and mechanical properties. These types of interconnected cloud-based applications are envisioned to be capable of further accelerating the adoption of machine learning methods into materials development.
[{'version': 'v1', 'created': 'Wed, 29 Nov 2017 09:35:46 GMT'}]
2017-11-30
Ruijin Cang, Hechao Li, Hope Yao, Yang Jiao, Yi Ren
Improving Direct Physical Properties Prediction of Heterogeneous Materials from Imaging Data via Convolutional Neural Network and a Morphology-Aware Generative Model
null
null
null
physics.comp-ph cond-mat.mtrl-sci
Direct prediction of material properties from microstructures through statistical models has shown to be a potential approach to accelerating computational material design with large design spaces. However, statistical modeling of highly nonlinear mappings defined on high-dimensional microstructure spaces is known to be data-demanding. Thus, the added value of such predictive models diminishes in common cases where material samples (in forms of 2D or 3D microstructures) become costly to acquire either experimentally or computationally. To this end, we propose a generative machine learning model that creates an arbitrary amount of artificial material samples with negligible computation cost, when trained on only a limited amount of authentic samples. The key contribution of this work is the introduction of a morphology constraint to the training of the generative model, that enforces the resultant artificial material samples to have the same morphology distribution as the authentic ones. We show empirically that the proposed model creates artificial samples that better match with the authentic ones in material property distributions than those generated from a state-of-the-art Markov Random Field model, and thus is more effective at improving the prediction performance of a predictive structure-property model.
[{'version': 'v1', 'created': 'Thu, 7 Dec 2017 06:49:29 GMT'}]
2017-12-12
Kristof T. Sch\"utt, Huziel E. Sauceda, Pieter-Jan Kindermans, Alexandre Tkatchenko, Klaus-Robert M\"uller
SchNet - a deep learning architecture for molecules and materials
null
10.1063/1.5019779
null
physics.chem-ph cond-mat.mtrl-sci
Deep learning has led to a paradigm shift in artificial intelligence, including web, text and image search, speech recognition, as well as bioinformatics, with growing impact in chemical physics. Machine learning in general and deep learning in particular is ideally suited for representing quantum-mechanical interactions, enabling to model nonlinear potential-energy surfaces or enhancing the exploration of chemical compound space. Here we present the deep learning architecture SchNet that is specifically designed to model atomistic systems by making use of continuous-filter convolutional layers. We demonstrate the capabilities of SchNet by accurately predicting a range of properties across chemical space for \emph{molecules and materials} where our model learns chemically plausible embeddings of atom types across the periodic table. Finally, we employ SchNet to predict potential-energy surfaces and energy-conserving force fields for molecular dynamics simulations of small molecules and perform an exemplary study of the quantum-mechanical properties of C$_{20}$-fullerene that would have been infeasible with regular ab initio molecular dynamics.
[{'version': 'v1', 'created': 'Sun, 17 Dec 2017 13:55:03 GMT'}, {'version': 'v2', 'created': 'Wed, 7 Mar 2018 12:42:19 GMT'}, {'version': 'v3', 'created': 'Thu, 22 Mar 2018 11:12:43 GMT'}]
2018-04-18
Maxim Ziatdinov, Ondrej Dyck, Artem Maksov, Xufan Li, Xiahan Sang, Kai Xiao, Raymond R. Unocic, Rama Vasudevan, Stephen Jesse, Sergei V. Kalinin
Deep Learning of Atomically Resolved Scanning Transmission Electron Microscopy Images: Chemical Identification and Tracking Local Transformations
ACS Nano, 2017, 11 (12), pp 12742-12752
10.1021/acsnano.7b07504
null
cond-mat.mtrl-sci
Recent advances in scanning transmission electron and scanning probe microscopies have opened exciting opportunities in probing the materials structural parameters and various functional properties in real space with angstrom-level precision. This progress has been accompanied by an exponential increase in the size and quality of datasets produced by microscopic and spectroscopic experimental techniques. These developments necessitate adequate methods for extracting relevant physical and chemical information from the large datasets, for which a priori information on the structures of various atomic configurations and lattice defects is limited or absent. Here we demonstrate an application of deep neural networks to extract information from atomically resolved images including location of the atomic species and type of defects. We develop a 'weakly-supervised' approach that uses information on the coordinates of all atomic species in the image, extracted via a deep neural network, to identify a rich variety of defects that are not part of an initial training set. We further apply our approach to interpret complex atomic and defect transformation, including switching between different coordination of silicon dopants in graphene as a function of time, formation of peculiar silicon dimer with mixed 3-fold and 4-fold coordination, and the motion of molecular 'rotor'. This deep learning based approach resembles logic of a human operator, but can be scaled leading to significant shift in the way of extracting and analyzing information from raw experimental data.
[{'version': 'v1', 'created': 'Wed, 17 Jan 2018 20:45:52 GMT'}]
2018-01-19
Jacob Madsen, Pei Liu, Jens Kling, Jakob Birkedal Wagner, Thomas Willum Hansen, Ole Winther, Jakob Schi{\o}tz
A deep learning approach to identify local structures in atomic-resolution transmission electron microscopy images
Adv. Theory Simul. 1, 1800037 (2018)
10.1002/adts.201800037
null
cond-mat.mtrl-sci
Recording atomic-resolution transmission electron microscopy (TEM) images is becoming increasingly routine. A new bottleneck is then analyzing this information, which often involves time-consuming manual structural identification. We have developed a deep learning-based algorithm for recognition of the local structure in TEM images, which is stable to microscope parameters and noise. The neural network is trained entirely from simulation but is capable of making reliable predictions on experimental images. We apply the method to single sheets of defected graphene, and to metallic nanoparticles on an oxide support.
[{'version': 'v1', 'created': 'Thu, 8 Feb 2018 18:57:20 GMT'}, {'version': 'v2', 'created': 'Fri, 9 Feb 2018 10:43:02 GMT'}]
2018-09-13
Rama K. Vasudevan, Nouamane Laanait, Erik M. Ferragut, Kai Wang, David B. Geohegan, Kai Xiao, Maxim A. Ziatdinov, Stephen Jesse, Ondrej E. Dyck, Sergei V. Kalinin
Mapping mesoscopic phase evolution during e-beam induced transformations via deep learning of atomically resolved images
null
null
null
cond-mat.mtrl-sci
Understanding transformations under electron beam irradiation requires mapping the structural phases and their evolution in real time. To date, this has mostly been a manual endeavor comprising of difficult frame-by-frame analysis that is simultaneously tedious and prone to error. Here, we turn towards the use of deep convolutional neural networks (DCNN) to automatically determine the Bravais lattice symmetry present in atomically-resolved images. A DCNN is trained to identify the Bravais lattice class given a 2D fast Fourier transform of the input image. Monte-Carlo dropout is used for determining the prediction probability, and results are shown for both simulated and real atomically-resolved images from scanning tunneling microscopy and scanning transmission electron microscopy. A reduced representation of the final layer output allows to visualize the separation of classes in the DCNN and agrees with physical intuition. We then apply the trained network to electron beam-induced transformations in WS2, which allows tracking and determination of growth rate of voids. These results are novel in two ways: (1) It shows that DCNNs can be trained to recognize diffraction patterns, which is markedly different from the typical "real image" cases, and (2) it provides a method with in-built uncertainty quantification, allowing the real-time analysis of phases present in atomically resolved images.
[{'version': 'v1', 'created': 'Wed, 28 Feb 2018 16:27:35 GMT'}]
2018-03-01
B.D. Conduit, N.G. Jones, H.J. Stone, G.J. Conduit
Probabilistic design of a molybdenum-base alloy using a neural network
Scripta Materialia 146, 82 (2018)
null
null
cond-mat.mtrl-sci cs.LG physics.comp-ph
An artificial intelligence tool is exploited to discover and characterize a new molybdenum-base alloy that is the most likely to simultaneously satisfy targets of cost, phase stability, precipitate content, yield stress, and hardness. Experimental testing demonstrates that the proposed alloy fulfils the computational predictions, and furthermore the physical properties exceed those of other commercially available Mo-base alloys for forging-die applications.
[{'version': 'v1', 'created': 'Fri, 2 Mar 2018 15:11:49 GMT'}]
2018-03-05
Youngjun Cho, Nadia Bianchi-Berthouze, Nicolai Marquardt and Simon J. Julier
Deep Thermal Imaging: Proximate Material Type Recognition in the Wild through Deep Learning of Spatial Surface Temperature Patterns
null
10.1145/3173574.3173576
null
cs.CV cond-mat.mtrl-sci cs.HC cs.LG
We introduce Deep Thermal Imaging, a new approach for close-range automatic recognition of materials to enhance the understanding of people and ubiquitous technologies of their proximal environment. Our approach uses a low-cost mobile thermal camera integrated into a smartphone to capture thermal textures. A deep neural network classifies these textures into material types. This approach works effectively without the need for ambient light sources or direct contact with materials. Furthermore, the use of a deep learning network removes the need to handcraft the set of features for different materials. We evaluated the performance of the system by training it to recognise 32 material types in both indoor and outdoor environments. Our approach produced recognition accuracies above 98% in 14,860 images of 15 indoor materials and above 89% in 26,584 images of 17 outdoor materials. We conclude by discussing its potentials for real-time use in HCI applications and future directions.
[{'version': 'v1', 'created': 'Tue, 6 Mar 2018 17:29:08 GMT'}]
2018-03-28
B.D. Conduit, N.G. Jones, H.J. Stone, and G.J. Conduit
Design of a nickel-base superalloy using a neural network
Materials & Design 131, 358 (2017)
null
null
cond-mat.mtrl-sci cs.LG physics.comp-ph
A new computational tool has been developed to model, discover, and optimize new alloys that simultaneously satisfy up to eleven physical criteria. An artificial neural network is trained from pre-existing materials data that enables the prediction of individual material properties both as a function of composition and heat treatment routine, which allows it to optimize the material properties to search for the material with properties most likely to exceed a target criteria. We design a new polycrystalline nickel-base superalloy with the optimal combination of cost, density, gamma' phase content and solvus, phase stability, fatigue life, yield stress, ultimate tensile strength, stress rupture, oxidation resistance, and tensile elongation. Experimental data demonstrates that the proposed alloy fulfills the computational predictions, possessing multiple physical properties, particularly oxidation resistance and yield stress, that exceed existing commercially available alloys.
[{'version': 'v1', 'created': 'Thu, 8 Mar 2018 11:04:58 GMT'}]
2018-03-09
Artem Maksov, Ondrej Dyck, Kai Wang, Kai Xiao, David B. Geohegan, Bobby G. Sumpter, Rama K. Vasudevan, Stephen Jesse, Sergei V. Kalinin, Maxim Ziatdinov
Deep Learning Analysis of Defect and Phase Evolution During Electron Beam Induced Transformations in WS2
npj Computational Materials 5, Article number: 12 (2019)
10.1038/s41524-019-0152-9
null
cond-mat.mtrl-sci
Understanding elementary mechanisms behind solid-state phase transformations and reactions is the key to optimizing desired functional properties of many technologically relevant materials. Recent advances in scanning transmission electron microscopy (STEM) allow the real-time visualization of solid-state transformations in materials, including those induced by an electron beam and temperature, with atomic resolution. However, despite the ever-expanding capabilities for high-resolution data acquisition, the inferred information about kinetics and thermodynamics of the process and single defect dynamics and interactions is minima, due to the inherent limitations of manual ex-situ analysis of the collected volumes of data. To circumvent this problem, we developed a deep learning framework for dynamic STEM imaging that is trained to find the structures (defects) that break a crystal lattice periodicity and apply it for mapping solid state reactions and transformations in layered WS2 doped with Mo. This framework allows extracting thousands of lattice defects from raw STEM data (single images and movies) in a matter of seconds, which are then classified into different categories using unsupervised clustering methods. We further expanded our framework to extract parameters of diffusion for the sulfur vacancies and analyzed transition probabilities associated with switching between different configurations of defect complexes consisting of Mo dopant and sulfur vacancy, providing insight into point defect dynamics and reactions. This approach is universal and its application to beam induced reactions allows mapping chemical transformation pathways in solids at the atomic level.
[{'version': 'v1', 'created': 'Wed, 14 Mar 2018 16:16:21 GMT'}, {'version': 'v2', 'created': 'Thu, 15 Mar 2018 05:29:36 GMT'}, {'version': 'v3', 'created': 'Thu, 16 Aug 2018 06:54:26 GMT'}]
2019-02-05
Xiaolin Li, Yichi Zhang, He Zhao, Craig Burkhart, L Catherine Brinson, Wei Chen
A Transfer Learning Approach for Microstructure Reconstruction and Structure-property Predictions
null
null
null
cond-mat.mtrl-sci cs.CE physics.comp-ph
Stochastic microstructure reconstruction has become an indispensable part of computational materials science, but ongoing developments are specific to particular material systems. In this paper, we address this generality problem by presenting a transfer learning-based approach for microstructure reconstruction and structure-property predictions that is applicable to a wide range of material systems. The proposed approach incorporates an encoder-decoder process and feature-matching optimization using a deep convolutional network. For microstructure reconstruction, model pruning is implemented in order to study the correlation between the microstructural features and hierarchical layers within the deep convolutional network. Knowledge obtained in model pruning is then leveraged in the development of a structure-property predictive model to determine the network architecture and initialization conditions. The generality of the approach is demonstrated numerically for a wide range of material microstructures with geometrical characteristics of varying complexity. Unlike previous approaches that only apply to specific material systems or require a significant amount of prior knowledge in model selection and hyper-parameter tuning, the present approach provides an off-the-shelf solution to handle complex microstructures, and has the potential of expediting the discovery of new materials.
[{'version': 'v1', 'created': 'Tue, 8 May 2018 00:01:48 GMT'}]
2018-05-09
Y. Liu, Q. M. Sun, Dr. W. H. Lu, Dr. H. L. Wang, Y. Sun, Z. T. Wang, X. Lu, Prof. K. Y. Zeng
General Resolution Enhancement Method in Atomic Force Microscopy (AFM) Using Deep Learning
null
null
null
physics.data-an cond-mat.mtrl-sci
This paper develops a resolution enhancement method for post-processing the images from Atomic Force Microscopy (AFM). This method is based on deep learning neural networks in the AFM topography measurements. In this study, a very deep convolution neural network is developed to derive the high-resolution topography image from the low-resolution topography image. The AFM measured images from various materials are tested in this study. The derived high-resolution AFM images are comparable with the experimental measured high-resolution images measured at the same locations. The results suggest that this method can be developed as a general post-processing method for AFM image analysis.
[{'version': 'v1', 'created': 'Tue, 11 Sep 2018 07:09:14 GMT'}]
2018-09-12
Yuan Dong, Chuhan Wu, Chi Zhang, Yingda Liu, Jianlin Cheng and Jian Lin
Deep Learning Bandgaps of Topologically Doped Graphene
null
null
null
cond-mat.mtrl-sci physics.comp-ph
Manipulation of material properties via precise doping affords enormous tunable phenomena to explore. Recent advance shows that in the atomic and nano scales topological states of dopants play crucial roles in determining their properties. However, such determination is largely unknown due to the incredible size of topological states. Here, we present a case study of developing deep learning algorithms to predict bandgaps of boron-nitrogen pair doped graphene with arbitrary dopant topologies. A material descriptor system that enables to correlate structures with the bandgaps was developed for convolutional neuron networks (CNNs). Bandgaps calculated by the ab initio calculations and the corresponding structures were fed as input datasets to train VGG16 convolutional network, residual convolutional network, and concatenate convolutional network. Then these trained CNNs were used to predict bandgaps of doped graphene with various dopant topologies. All of them afford great prediction accuracy, showing square of the coefficient of correlation (R2) of > 90% and root-mean-square errors of ~ 0.1 eV for the predicted bandgaps. They are much better than those predicted by a shallow machine learning method - support vector machine. The transfer learning was further performed by leveraging data generated from smaller systems to improve the prediction for large systems. Success of this work provides a cornerstone for future investigation of topologically doped graphene and other 2D materials. Moreover, given ubiquitous existence of topologies in materials, this work will stimulate widespread interests in applying deep learning algorithms to topological design of materials crossing atomic, nano-, meso-, and macro- scales.
[{'version': 'v1', 'created': 'Fri, 28 Sep 2018 05:02:49 GMT'}]
2018-10-01
Soumya Sanyal, Janakiraman Balachandran, Naganand Yadati, Abhishek Kumar, Padmini Rajagopalan, Suchismita Sanyal, Partha Talukdar
MT-CGCNN: Integrating Crystal Graph Convolutional Neural Network with Multitask Learning for Material Property Prediction
null
null
null
cs.LG cond-mat.mtrl-sci stat.ML
Developing accurate, transferable and computationally inexpensive machine learning models can rapidly accelerate the discovery and development of new materials. Some of the major challenges involved in developing such models are, (i) limited availability of materials data as compared to other fields, (ii) lack of universal descriptor of materials to predict its various properties. The limited availability of materials data can be addressed through transfer learning, while the generic representation was recently addressed by Xie and Grossman [1], where they developed a crystal graph convolutional neural network (CGCNN) that provides a unified representation of crystals. In this work, we develop a new model (MT-CGCNN) by integrating CGCNN with transfer learning based on multi-task (MT) learning. We demonstrate the effectiveness of MT-CGCNN by simultaneous prediction of various material properties such as Formation Energy, Band Gap and Fermi Energy for a wide range of inorganic crystals (46774 materials). MT-CGCNN is able to reduce the test error when employed on correlated properties by upto 8%. The model prediction has lower test error compared to CGCNN, even when the training data is reduced by 10%. We also demonstrate our model's better performance through prediction of end user scenario related to metal/non-metal classification. These results encourage further development of machine learning approaches which leverage multi-task learning to address the aforementioned challenges in the discovery of new materials. We make MT-CGCNN's source code available to encourage reproducible research.
[{'version': 'v1', 'created': 'Wed, 14 Nov 2018 06:13:29 GMT'}]
2018-11-15
Kevin Ryczko, David Strubbe, Isaac Tamblyn
Deep Learning and Density Functional Theory
Phys. Rev. A 100, 022512 (2019)
10.1103/PhysRevA.100.022512
null
cond-mat.mtrl-sci physics.comp-ph
We show that deep neural networks can be integrated into, or fully replace, the Kohn-Sham density functional theory scheme for multi-electron systems in simple harmonic oscillator and random external potentials with no feature engineering. We first show that self-consistent charge densities calculated with different exchange-correlation functionals can be used as input to an extensive deep neural network to make predictions for correlation, exchange, external, kinetic and total energies simultaneously. Additionally, we show that one can also make all of the same predictions with the external potential rather than the self-consistent charge density, which allows one to circumvent the Kohn-Sham scheme altogether. We then show that a self-consistent charge density found from a non-local exchange-correlation functional can be used to make energy predictions for a semi-local exchange-correlation functional. Lastly, we use a deep convolutional inverse graphics network to predict the charge density given an external potential for different exchange-correlation functionals and assess the viability of the predicted charge densities. This work shows that extensive deep neural networks are generalizable and transferable given the variability of the potentials (maximum total energy range $\approx100$ Ha), because they require no feature engineering, and because they can scale to an arbitrary system size with an $\mathcal{O}(N)$ computational cost.
[{'version': 'v1', 'created': 'Wed, 21 Nov 2018 20:03:01 GMT'}, {'version': 'v2', 'created': 'Wed, 24 Feb 2021 15:13:46 GMT'}]
2021-02-25
Rahul Singh, Aayush Sharma, Onur Rauf Bingol, Aditya Balu, Ganesh Balasubramanian, Duane D. Johnson and Soumik Sarkar
3D Deep Learning with voxelized atomic configurations for modeling atomistic potentials in complex solid-solution alloys
null
null
null
cond-mat.mtrl-sci cs.LG physics.comp-ph stat.ML
The need for advanced materials has led to the development of complex, multi-component alloys or solid-solution alloys. These materials have shown exceptional properties like strength, toughness, ductility, electrical and electronic properties. Current development of such material systems are hindered by expensive experiments and computationally demanding first-principles simulations. Atomistic simulations can provide reasonable insights on properties in such material systems. However, the issue of designing robust potentials still exists. In this paper, we explore a deep convolutional neural-network based approach to develop the atomistic potential for such complex alloys to investigate materials for insights into controlling properties. In the present work, we propose a voxel representation of the atomic configuration of a cell and design a 3D convolutional neural network to learn the interaction of the atoms. Our results highlight the performance of the 3D convolutional neural network and its efficacy in machine-learning the atomistic potential. We also explore the role of voxel resolution and provide insights into the two bounding box methodologies implemented for voxelization.
[{'version': 'v1', 'created': 'Fri, 23 Nov 2018 23:12:22 GMT'}]
2018-11-27
Tomohiko Konno, Hodaka Kurokawa, Fuyuki Nabeshima, Yuki Sakishita, Ryo Ogawa, Iwao Hosako, Atsutaka Maeda
Deep Learning Model for Finding New Superconductors
Phys. Rev. B 103, 014509 (2021)
10.1103/PhysRevB.103.014509
null
cs.LG cond-mat.mtrl-sci cond-mat.supr-con cs.CL physics.comp-ph
Exploration of new superconductors still relies on the experience and intuition of experts and is largely a process of experimental trial and error. In one study, only 3% of the candidate materials showed superconductivity. Here, we report the first deep learning model for finding new superconductors. We introduced the method named "reading periodic table" which represented the periodic table in a way that allows deep learning to learn to read the periodic table and to learn the law of elements for the purpose of discovering novel superconductors that are outside the training data. It is recognized that it is difficult for deep learning to predict something outside the training data. Although we used only the chemical composition of materials as information, we obtained an $R^{2}$ value of 0.92 for predicting $T_\text{c}$ for materials in a database of superconductors. We also introduced the method named "garbage-in" to create synthetic data of non-superconductors that do not exist. Non-superconductors are not reported, but the data must be required for deep learning to distinguish between superconductors and non-superconductors. We obtained three remarkable results. The deep learning can predict superconductivity for a material with a precision of 62%, which shows the usefulness of the model; it found the recently discovered superconductor CaBi2 and another one Hf0.5Nb0.2V2Zr0.3, neither of which is in the superconductor database; and it found Fe-based high-temperature superconductors (discovered in 2008) from the training data before 2008. These results open the way for the discovery of new high-temperature superconductor families. The candidate materials list, data, and method are openly available from the link https://github.com/tomo835g/Deep-Learning-to-find-Superconductors.
[{'version': 'v1', 'created': 'Mon, 3 Dec 2018 05:30:34 GMT'}, {'version': 'v2', 'created': 'Mon, 3 Jun 2019 07:22:53 GMT'}, {'version': 'v3', 'created': 'Sun, 3 Nov 2019 14:29:01 GMT'}, {'version': 'v4', 'created': 'Thu, 14 Jan 2021 14:36:38 GMT'}]
2021-01-20
Myungjoon Kim, Byung Chul Yeo, Sang Soo Han, Donghun Kim
Slab Graph Convolutional Neural Network for Discovery of N2 Electroreduction Catalysts
null
null
null
cond-mat.mtrl-sci
The catalyst development for N2 electroreduction reaction (NRR) with low onset potential and high Faradaic efficiency is highly desired, but remains challenging. Machine learning (ML) recently emerged as a complementary tool to accelerate material discovery; however a ML model for NRR has yet to be developed. Here, we develop and report slab-graph convolutional neural network (SGCNN), an accurate and flexible ML model that is applicable to catalytic surface reactions. With the self-accumulated database of 2,699 surface calculations, SGCNN predict binding energies, ranging over 8 eV, of five key adsorbates (*H, *N2, *N2H, *NH, *NH2) related to NRR performance with the mean-absolute-error of only 0.23eV. Unlike previously available models, SGCNN avoids using ab initio level inputs, instead is solely based on elemental properties that are all readily available in Periodic-Table-of-Elements; true accelerations can be realized. t-distributed stochastic neighbor embedding (t-SNE) analysis reveals that binary intermetallics of averaged d-electron occupation between 4 and 5 could potentially lower the onset potential in N2 electroreduction.
[{'version': 'v1', 'created': 'Fri, 7 Dec 2018 08:50:31 GMT'}, {'version': 'v2', 'created': 'Wed, 27 Mar 2019 01:07:09 GMT'}]
2019-03-28
Shweta Mehta, Sheena Agarwal, and Kavita Joshi
Combining DFT with ML to study size specific interactions between metal clusters and adsorbates
null
null
null
cond-mat.mtrl-sci
To date, density functional theory (DFT) is one of the most accurate and yet practical theory to gain insight about materials properties. Although successful, the computational cost is the main hurdle even today. A way out is combining DFT with machine learning (ML) to reduce the computational cost without compromising accuracy. However, the success of this approach hinges on the correctness of the descriptors. In the present work, we demonstrate that, based on {\it only} interatomic distances as descriptors, our ML model predicts interaction energy between an adsorbate and Al cluster with absolute mean error (AME) $\sim$ 0.05 eV (or less) and reproduces the PES experienced by an incoming atom. Our extensive DFT calculations reveal that atoms experiencing identical environment within a cluster have identical interaction energy patterns. Further, we demonstrate that our model is not specific to Al clusters, and could be applied to clusters of different elements as well. Its application to compute PES experienced by various test atoms and molecules in the vicinity of different clusters proves the transferability of the model not just to clusters of different elements but also to various molecules. The descriptors chosen are invariant to rotation, translation, and permutation yet very simple to compute is one of the most crucial points of the present work.
[{'version': 'v1', 'created': 'Wed, 12 Dec 2018 13:21:40 GMT'}, {'version': 'v2', 'created': 'Wed, 2 Jan 2019 05:22:03 GMT'}, {'version': 'v3', 'created': 'Tue, 9 Apr 2019 10:00:00 GMT'}, {'version': 'v4', 'created': 'Thu, 18 Apr 2019 06:46:13 GMT'}]
2019-04-19
Tian Xie, Arthur France-Lanord, Yanming Wang, Yang Shao-Horn, Jeffrey C. Grossman
Graph Dynamical Networks for Unsupervised Learning of Atomic Scale Dynamics in Materials
Nat. Commun. 10, 2667 (2019)
10.1038/s41467-019-10663-6
null
cond-mat.mtrl-sci cs.LG physics.chem-ph
Understanding the dynamical processes that govern the performance of functional materials is essential for the design of next generation materials to tackle global energy and environmental challenges. Many of these processes involve the dynamics of individual atoms or small molecules in condensed phases, e.g. lithium ions in electrolytes, water molecules in membranes, molten atoms at interfaces, etc., which are difficult to understand due to the complexity of local environments. In this work, we develop graph dynamical networks, an unsupervised learning approach for understanding atomic scale dynamics in arbitrary phases and environments from molecular dynamics simulations. We show that important dynamical information can be learned for various multi-component amorphous material systems, which is difficult to obtain otherwise. With the large amounts of molecular dynamics data generated everyday in nearly every aspect of materials design, this approach provides a broadly useful, automated tool to understand atomic scale dynamics in material systems.
[{'version': 'v1', 'created': 'Mon, 18 Feb 2019 23:17:27 GMT'}, {'version': 'v2', 'created': 'Wed, 22 May 2019 20:58:39 GMT'}]
2019-07-11
Nouamane Laanait and Qian He and Albina Y. Borisevich
Reconstruction of 3-D Atomic Distortions from Electron Microscopy with Deep Learning
null
null
null
cond-mat.mtrl-sci cs.LG
Deep learning has demonstrated superb efficacy in processing imaging data, yet its suitability in solving challenging inverse problems in scientific imaging has not been fully explored. Of immense interest is the determination of local material properties from atomically-resolved imaging, such as electron microscopy, where such information is encoded in subtle and complex data signatures, and whose recovery and interpretation necessitate intensive numerical simulations subject to the requirement of near-perfect knowledge of the experimental setup. We demonstrate that an end-to-end deep learning model can successfully recover 3-dimensional atomic distortions of a variety of oxide perovskite materials from a single 2-dimensional experimental scanning transmission electron (STEM) micrograph, in the process resolving a longstanding question in the recovery of 3-D atomic distortions from STEM experiments. Our results indicate that deep learning is a promising approach to efficiently address unsolved inverse problems in scientific imaging and to underpin novel material investigations at atomic resolution.
[{'version': 'v1', 'created': 'Tue, 19 Feb 2019 03:31:53 GMT'}]
2019-02-20
Mohammad Rashidi, Jeremiah Croshaw, Kieran Mastel, Marcus Tamura, Hedieh Hosseinzadeh, and Robert A. Wolkow
Deep Learning-Guided Surface Characterization for Autonomous Hydrogen Lithography
Mach. Learn.: Sci. Technol. 1 025001 (2020)
10.1088/2632-2153/ab6d5e
null
cond-mat.mtrl-sci
As the development of atom scale devices transitions from novel, proof-of-concept demonstrations to state-of-the-art commercial applications, automated assembly of such devices must be implemented. Here we present an automation method for the identification of defects prior to atomic fabrication via hydrogen lithography using deep learning. We trained a convolutional neural network to locate and differentiate between surface features of the technologically relevant hydrogen-terminated silicon surface imaged using a scanning tunneling microscope. Once the positions and types of surface features are determined, the predefined atomic structures are patterned in a defect-free area. By training the network to differentiate between common defects we are able to avoid charged defects as well as edges of the patterning terraces. Augmentation with previously developed autonomous tip shaping and patterning modules allows for atomic scale lithography with minimal user intervention.
[{'version': 'v1', 'created': 'Sat, 23 Feb 2019 17:37:28 GMT'}, {'version': 'v2', 'created': 'Fri, 11 Oct 2019 18:57:15 GMT'}]
2020-03-26
Dongsun Yoo, Kyuhyun Lee, Wonseok Jeong, Satoshi Watanabe, Seungwu Han
Atomic energy mapping of neural network potential
Phys. Rev. Materials 3, 093802 (2019)
10.1103/PhysRevMaterials.3.093802
null
physics.comp-ph cond-mat.mtrl-sci physics.chem-ph
We show that the intelligence of the machine-learning potential arises from its ability to infer the reference atomic-energy function from a given set of total energies. By utilizing invariant points in the feature space at which the atomic energy has a fixed reference value, we examine the atomic energy mapping of neural network potentials. Through a series of examples on Si, we demonstrate that the neural network potential is vulnerable to 'ad hoc' mapping in which the total energy appears to be trained accurately while the atomic energy mapping is incorrect in spite of its capability. We show that the energy mapping can be improved by choosing the training set carefully and monitoring the atomic energy at the invariant points during the training procedure.
[{'version': 'v1', 'created': 'Mon, 11 Mar 2019 15:21:01 GMT'}]
2019-09-05
Sandeep Madireddy, Ding-Wen Chung, Troy Loeffler, Subramanian K.R.S. Sankaranarayanan, David N. Seidman, Prasanna Balaprakash, and Olle Heinonen
Phase Segmentation in Atom-Probe Tomography Using Deep Learning-Based Edge Detection
null
null
null
cond-mat.mtrl-sci physics.comp-ph
Atom-probe tomography (APT) facilitates nano- and atomic-scale characterization and analysis of microstructural features. Specifically, APT is well suited to study the interfacial properties of granular or heterophase systems. Traditionally, the identification of the interface between, for precipitate and matrix phases, in APT data has been obtained either by extracting iso-concentration surfaces based on a user-supplied concentration value or by manually perturbing the concentration value until the iso-concentration surface qualitatively matches the interface. These approaches are subjective, not scalable, and may lead to inconsistencies due to local composition inhomogeneities. We propose a digital image segmentation approach based on deep neural networks that transfer learned knowledge from natural images to automatically segment the data obtained from APT into different phases. This approach not only provides an efficient way to segment the data and extract interfacial properties but does so without the need for expensive interface labeling for training the segmentation model. We consider here a system with a precipitate phase in a matrix and with three different interface modalities---layered, isolated, and interconnected---that are obtained for different relative geometries of the precipitate phase. We demonstrate the accuracy of our segmentation approach through qualitative visualization of the interfaces, as well as through quantitative comparisons with proximity histograms obtained by using more traditional approaches.
[{'version': 'v1', 'created': 'Wed, 10 Apr 2019 20:53:33 GMT'}]
2019-04-12
Jutta Rogal, Elia Schneider, Mark E. Tuckerman
Neural network based path collective variables for enhanced sampling of phase transformations
Phys. Rev. Lett. 123, 245701 (2019)
10.1103/PhysRevLett.123.245701
null
cond-mat.mtrl-sci cond-mat.stat-mech physics.comp-ph
We propose a rigorous construction of a 1D path collective variable to sample structural phase transformations in condensed matter. The path collective variable is defined in a space spanned by global collective variables that serve as classifiers derived from local structural units. A reliable identification of local structural environments is achieved by employing a neural network based classification. The 1D path collective variable is subsequently used together with enhanced sampling techniques to explore the complex migration of a phase boundary during a solid-solid phase transformation in molybdenum.
[{'version': 'v1', 'created': 'Sat, 4 May 2019 18:05:54 GMT'}]
2022-12-09
Liang Li, Mindren Lu, and Maria K. Y. Chan
A Deep Learning Model for Atomic Structures Prediction Using X-ray Absorption Spectroscopic Data
null
null
null
physics.comp-ph cond-mat.mtrl-sci
A deep neural network (DNN) model consisting of two hidden layers was proposed for predicting the immediate environments of specific atoms based on X-ray absorption near-edge spectra (XANES). The output layer of the DNN can be adjusted to form a classifier or regressor, to predict the local and overall coordination environments, respectively. Using Li3FeO3.5 as a model system, it was demonstrated that the prediction accuracy of the DNN classifier is higher than 98%, and the predictions of the DNN regressor also showed notable agreement with the ground truth. Therefore, despite its simplicity, this DNN architecture can be expected to be generally capable of predicting the structural properties of various systems. Fine tuning of the hyperparameters, bias-variance tradeoff, and strategies to enrich the versatility of the model were also discussed.
[{'version': 'v1', 'created': 'Fri, 10 May 2019 04:08:40 GMT'}]
2019-05-13
Emi Minamitani, Masayoshi Ogura, Satoshi Watanabe
Simulating lattice thermal conductivity in semiconducting materials using high-dimensional neural network potential
null
10.7567/1882-0786/ab36bc
null
cond-mat.mtrl-sci physics.comp-ph
We demonstrate that a high-dimensional neural network potential (HDNNP) can predict the lattice thermal conductivity of semiconducting materials with an accuracy comparable to that of density functional theory (DFT) calculation. After a training procedure based on the force, the root mean square error between the forces predicted by the HDNNP and DFT is less than 40 meV/{\AA}. As typical examples, we present the results for Si and GaN bulk crystals. The deviation from the thermal conductivity calculated using DFT is within 1% at 200 to 500 K for Si and within 5.4% at 200 to 1000 K for GaN.
[{'version': 'v1', 'created': 'Tue, 21 May 2019 09:13:33 GMT'}]
2019-08-16
Brian Gallagher, Matthew Rever, Donald Loveland, T. Nathan Mundhenk, Brock Beauchamp, Emily Robertson, Golam G. Jaman, Anna M. Hiszpanski, and T. Yong-Jin Han
Predicting Compressive Strength of Consolidated Molecular Solids Using Computer Vision and Deep Learning
null
10.1016/j.matdes.2020.108541
null
physics.comp-ph cond-mat.mtrl-sci
We explore the application of computer vision and machine learning (ML) techniques to predict material properties (e.g. compressive strength) based on SEM images. We show that it's possible to train ML models to predict materials performance based on SEM images alone, demonstrating this capability on the real-world problem of predicting uniaxially compressed peak stress of consolidated molecular solids samples. Our image-based ML approach reduces mean absolute percent error (MAPE) by an average of 24% over baselines representative of the current state-of-the-practice (i.e., domain-expert's analysis and correlation). We compared two complementary approaches to this problem: (1) a traditional ML approach, random forest (RF), using state-of-the-art computer vision features and (2) an end-to-end deep learning (DL) approach, where features are learned automatically from raw images. We demonstrate the complementarity of these approaches, showing that RF performs best in the "small data" regime in which many real-world scientific applications reside (up to 24% lower RMSE than DL), whereas DL outpaces RF in the "big data" regime, where abundant training samples are available (up to 24% lower RMSE than RF). Finally, we demonstrate that models trained using machine learning techniques are capable of discovering and utilizing informative crystal attributes previously underutilized by domain experts.
[{'version': 'v1', 'created': 'Wed, 5 Jun 2019 16:49:00 GMT'}, {'version': 'v2', 'created': 'Sat, 9 Nov 2019 04:52:14 GMT'}, {'version': 'v3', 'created': 'Fri, 28 Feb 2020 01:54:33 GMT'}]
2020-03-02
Cheol Woo Park, Chris Wolverton
Developing an improved Crystal Graph Convolutional Neural Network framework for accelerated materials discovery
Phys. Rev. Materials 4, 063801 (2020)
10.1103/PhysRevMaterials.4.063801
null
physics.comp-ph cond-mat.mtrl-sci physics.data-an
The recently proposed crystal graph convolutional neural network (CGCNN) offers a highly versatile and accurate machine learning (ML) framework by learning material properties directly from graph-like representations of crystal structures ("crystal graphs"). Here, we develop an improved variant of the CGCNN model (iCGCNN) that outperforms the original by incorporating information of the Voronoi tessellated crystal structure, explicit 3-body correlations of neighboring constituent atoms, and an optimized chemical representation of interatomic bonds in the crystal graphs. We demonstrate the accuracy of the improved framework in two distinct illustrations: First, when trained/validated on 180,000/20,000 density functional theory (DFT) calculated thermodynamic stability entries taken from the Open Quantum Materials Database (OQMD) and evaluated on a separate test set of 230,000 entries, iCGCNN achieves a predictive accuracy that is significantly improved, i.e., 20% higher than that of the original CGCNN. Second, when used to assist high-throughput search for materials in the ThCr2Si2 structure-type, iCGCNN exhibited a success rate of 31% which is 310 times higher than an undirected high-throughput search and 2.4 times higher than that of the original CGCNN. Using both CGCNN and iCGCNN, we screened 132,600 compounds with elemental decorations of the ThCr2Si2 prototype crystal structure and identified a total of 97 new unique stable compounds by performing 757 DFT calculations, accelerating the computational time of the high-throughput search by a factor of 130. Our results suggest that the iCGCNN can be used to accelerate high-throughput discoveries of new materials by quickly and accurately identifying crystalline compounds with properties of interest.
[{'version': 'v1', 'created': 'Wed, 12 Jun 2019 17:47:43 GMT'}]
2020-07-01
Linfeng Zhang, Mohan Chen, Xifan Wu, Han Wang, Weinan E, Roberto Car
Deep neural network for the dielectric response of insulators
Phys. Rev. B 102, 041121 (2020)
10.1103/PhysRevB.102.041121
null
physics.comp-ph cond-mat.mtrl-sci physics.chem-ph
We introduce a deep neural network to model in a symmetry preserving way the environmental dependence of the centers of the electronic charge. The model learns from ab-initio density functional theory, wherein the electronic centers are uniquely assigned by the maximally localized Wannier functions. When combined with the Deep Potential model of the atomic potential energy surface, the scheme predicts the dielectric response of insulators for trajectories inaccessible to direct ab-initio simulation. The scheme is non-perturbative and can capture the response of a mutating chemical environment. We demonstrate the approach by calculating the infrared spectra of liquid water at standard conditions, and of ice under extreme pressure, when it transforms from a molecular to an ionic crystal.
[{'version': 'v1', 'created': 'Thu, 27 Jun 2019 04:44:07 GMT'}, {'version': 'v2', 'created': 'Tue, 2 Jul 2019 14:45:58 GMT'}, {'version': 'v3', 'created': 'Sat, 7 Sep 2019 05:11:51 GMT'}, {'version': 'v4', 'created': 'Mon, 3 Feb 2020 11:58:17 GMT'}, {'version': 'v5', 'created': 'Tue, 9 Jun 2020 22:38:05 GMT'}]
2020-07-29
Ruggero Lot, Franco Pellegrini, Yusuf Shaidu, Emine Kucukbenli
PANNA: Properties from Artificial Neural Network Architectures
null
10.1016/j.cpc.2020.107402
null
physics.comp-ph cond-mat.mtrl-sci
Prediction of material properties from first principles is often a computationally expensive task. Recently, artificial neural networks and other machine learning approaches have been successfully employed to obtain accurate models at a low computational cost by leveraging existing example data. Here, we present a software package "Properties from Artificial Neural Network Architectures" (PANNA) that provides a comprehensive toolkit for creating neural network models for atomistic systems. Besides the core routines for neural network training, it includes data parser, descriptor builder and force-field generator suitable for integration within molecular dynamics packages. PANNA offers a variety of activation and cost functions, regularization methods, as well as the possibility of using fully-connected networks with custom size for each atomic species. PANNA benefits from the optimization and hardware-flexibility of the underlying TensorFlow engine which allows it to be used on multiple CPU/GPU/TPU systems, making it possible to develop and optimize neural network models based on large datasets.
[{'version': 'v1', 'created': 'Sat, 6 Jul 2019 00:42:46 GMT'}]
2020-07-15
Kirk Swanson, Shubhendu Trivedi, Joshua Lequieu, Kyle Swanson, Risi Kondor
Deep Learning for Automated Classification and Characterization of Amorphous Materials
null
null
null
cond-mat.soft cond-mat.dis-nn cond-mat.mtrl-sci cs.LG stat.ML
It is difficult to quantify structure-property relationships and to identify structural features of complex materials. The characterization of amorphous materials is especially challenging because their lack of long-range order makes it difficult to define structural metrics. In this work, we apply deep learning algorithms to accurately classify amorphous materials and characterize their structural features. Specifically, we show that convolutional neural networks and message passing neural networks can classify two-dimensional liquids and liquid-cooled glasses from molecular dynamics simulations with greater than 0.98 AUC, with no a priori assumptions about local particle relationships, even when the liquids and glasses are prepared at the same inherent structure energy. Furthermore, we demonstrate that message passing neural networks surpass convolutional neural networks in this context in both accuracy and interpretability. We extract a clear interpretation of how message passing neural networks evaluate liquid and glass structures by using a self-attention mechanism. Using this interpretation, we derive three novel structural metrics that accurately characterize glass formation. The methods presented here provide us with a procedure to identify important structural features in materials that could be missed by standard techniques and give us a unique insight into how these neural networks process data.
[{'version': 'v1', 'created': 'Tue, 10 Sep 2019 17:49:04 GMT'}]
2019-09-11
Giovanni Drera, Chahan M. Kropf, Luigi Sangaletti
Deep neural network for X-ray photoelectron spectroscopy data analysis
null
null
null
cond-mat.dis-nn cond-mat.mtrl-sci physics.comp-ph
In this work, we characterize the performance of a deep convolutional neural network designed to detect and quantify chemical elements in experimental X-ray photoelectron spectroscopy data. Given the lack of a reliable database in literature, in order to train the neural network we computed a large ($>$100 k) dataset of synthetic spectra, based on randomly generated materials covered with a layer of adventitious carbon. The trained net performs as good as standard methods on a test set of $\approx$ 500 well characterized experimental X-ray photoelectron spectra. Fine details about the net layout, the choice of the loss function and the quality assessment strategies are presented and discussed. Given the synthetic nature of the training set, this approach could be applied to the automatization of any photoelectron spectroscopy system, without the need of experimental reference spectra and with a low computational effort.
[{'version': 'v1', 'created': 'Thu, 12 Sep 2019 09:28:21 GMT'}]
2019-09-13
Mingjian Wen and Ellad B. Tadmor
Hybrid neural network potential for multilayer graphene
Phys. Rev. B 100, 195419 (2019)
10.1103/PhysRevB.100.195419
null
cond-mat.mtrl-sci physics.comp-ph
Monolayer and multilayer graphene are promising materials for applications such as electronic devices, sensors, energy generation and storage, and medicine. In order to perform large-scale atomistic simulations of the mechanical and thermal behavior of graphene-based devices, accurate interatomic potentials are required. Here, we present a new interatomic potential for multilayer graphene structures referred to as "hNN--Gr$_x$." This hybrid potential employs a neural network to describe short-range interactions and a theoretically-motivated analytical term to model long-range dispersion. The potential is trained against a large dataset of monolayer graphene, bilayer graphene, and graphite configurations obtained from ab initio total-energy calculations based on density functional theory (DFT). The potential provides accurate energy and forces for both intralayer and interlayer interactions, correctly reproducing DFT results for structural, energetic, and elastic properties such as the equilibrium layer spacing, interlayer binding energy, elastic moduli, and phonon dispersions to which it was not fit. The potential is used to study the effect of vacancies on thermal conductivity in monolayer graphene and interlayer friction in bilayer graphene. The potential is available through the OpenKIM interatomic potential repository at \url{https://openkim.org}.
[{'version': 'v1', 'created': 'Mon, 23 Sep 2019 03:01:10 GMT'}, {'version': 'v2', 'created': 'Sat, 16 Nov 2019 21:41:07 GMT'}]
2019-11-27
Anton S. Bochkarev, Ambroise van Roekeghem, Stefano Mossa, Natalio Mingo
Anharmonic Thermodynamics of Vacancies Using a Neural Network Potential
null
10.1103/PhysRevMaterials.3.093803
null
cond-mat.mtrl-sci
Lattice anharmonicity is thought to strongly affect vacancy concentrations in metals at high temperatures. It is however non-trivial to account for this effect directly using density functional theory (DFT). Here we develop a deep neural network potential for aluminum that overcomes the limitations inherent to DFT, and we use it to obtain accurate anharmonic vacancy formation free energies as a function of temperature. While confirming the important role of anharmonicity at high temperatures, the calculation unveils a markedly nonlinear behavior of the vacancy formation entropy and shows that the vacancy formation free energy only violates Arrhenius law at temperatures above 600 K, in contrast with previous DFT calculations.
[{'version': 'v1', 'created': 'Mon, 23 Sep 2019 09:52:30 GMT'}]
2019-10-02
Nouamane Laanait, Joshua Romero, Junqi Yin, M. Todd Young, Sean Treichler, Vitalii Starchenko, Albina Borisevich, Alex Sergeev, Michael Matheson
Exascale Deep Learning for Scientific Inverse Problems
null
null
null
cs.LG cond-mat.mtrl-sci cs.DC physics.comp-ph stat.ML
We introduce novel communication strategies in synchronous distributed Deep Learning consisting of decentralized gradient reduction orchestration and computational graph-aware grouping of gradient tensors. These new techniques produce an optimal overlap between computation and communication and result in near-linear scaling (0.93) of distributed training up to 27,600 NVIDIA V100 GPUs on the Summit Supercomputer. We demonstrate our gradient reduction techniques in the context of training a Fully Convolutional Neural Network to approximate the solution of a longstanding scientific inverse problem in materials imaging. The efficient distributed training on a dataset size of 0.5 PB, produces a model capable of an atomically-accurate reconstruction of materials, and in the process reaching a peak performance of 2.15(4) EFLOPS$_{16}$.
[{'version': 'v1', 'created': 'Tue, 24 Sep 2019 19:40:59 GMT'}]
2019-09-26
Shenghong Ju, Ryo Yoshida, Chang Liu, Kenta Hongo, Terumasa Tadano, Junichiro Shiomi
Exploring diamond-like lattice thermal conductivity crystals via feature-based transfer learning
Phys. Rev. Materials 5, 053801 (2021)
10.1103/PhysRevMaterials.5.053801
null
cond-mat.mtrl-sci physics.comp-ph
Ultrahigh lattice thermal conductivity materials hold great importance since they play a critical role in the thermal management of electronic and optical devices. Models using machine learning can search for materials with outstanding higher-order properties like thermal conductivity. However, the lack of sufficient data to train a model is a serious hurdle. Herein we show that big data can complement small data for accurate predictions when lower-order feature properties available in big data are selected properly and applied to transfer learning. The connection between the crystal information and thermal conductivity is directly built with a neural network by transferring descriptors acquired through a pre-trained model for the feature property. Successful transfer learning shows the ability of extrapolative prediction and reveals descriptors for lattice anharmonicity. Transfer learning is employed to screen over 60000 compounds to identify novel crystals that can serve as alternatives to diamond.
[{'version': 'v1', 'created': 'Wed, 25 Sep 2019 00:10:13 GMT'}]
2021-05-19
Pietro D'Antuono and Michele Ciavarella
Mean stress effect on Ga{\ss}ner curves interpreted as shifted W\"ohler curves
null
null
null
physics.app-ph cond-mat.mtrl-sci
A criterion for the mean stress effect correction in the shift factor approach for variable amplitude life prediction is presented for both smooth and notched specimens. The criterion is applied to the simple idea proposed by the authors in a previous note that Ga{\ss}ner curves can be interpreted as shifted W\"ohler curves. The mean stress correction used has been proposed by Smith, Watson and Topper and, more in general, by Walker. By applying the correction, a new expression for the shift factor G is obtained and, through the application of the theory of the critical distances in its point variant, surprisingly G is demonstrated to be valid for both smooth and notched geometries since it does not seem to depend on the geometry, but only on the fatigue exponent and the loading history. Finally, a comparison with the SAE Keyhole test program data is added to substantiate the findings.
[{'version': 'v1', 'created': 'Sun, 29 Sep 2019 17:28:27 GMT'}]
2019-10-01
Tarak K Patra, Troy D. Loeffler, Henry Chan, Mathew J. Cherukara, Badri Narayanan and Subramanian K.R.S. Sankaranarayanan
A coarse-grained deep neural network model for liquid water
null
10.1063/1.5116591
null
physics.comp-ph cond-mat.mtrl-sci physics.chem-ph
We introduce a coarse-grained deep neural network model (CG-DNN) for liquid water that utilizes 50 rotational and translational invariant coordinates, and is trained exclusively against energies of ~30,000 bulk water configurations. Our CG-DNN potential accurately predicts both the energies and molecular forces of water; within 0.9 meV/molecule and 54 meV/angstrom of a reference (coarse-grained bond-order potential) model. The CG-DNN water model also provides good prediction of several structural, thermodynamic, and temperature dependent properties of liquid water, with values close to that obtained from the reference model. More importantly, CG-DNN captures the well-known density anomaly of liquid water observed in experiments. Our work lays the groundwork for a scheme where existing empirical water models can be utilized to develop fully flexible neural network framework that can subsequently be trained against sparse data from high-fidelity albeit expensive beyond-DFT calculations.
[{'version': 'v1', 'created': 'Tue, 1 Oct 2019 08:32:01 GMT'}, {'version': 'v2', 'created': 'Mon, 14 Oct 2019 16:28:45 GMT'}]
2020-01-08
Ari Frankel, Kousuke Tachida, Reese Jones
Prediction of the evolution of the stress field of polycrystals undergoing elastic-plastic deformation with a hybrid neural network model
null
null
null
physics.comp-ph cond-mat.mtrl-sci
Crystal plasticity theory is often employed to predict the mesoscopic states of polycrystalline metals, and is well-known to be costly to simulate. Using a neural network with convolutional layers encoding correlations in time and space, we were able to predict the evolution of the stress field given only the initial microstructure and external loading. In comparison to our recent work we were able to predict not only the spatial average of the stress response but the field itself. We show that the stress fields and their rates are in high fidelity with the crystal plasticity data and have no visible artifacts. Furthermore the distribution stress throughout the elastic to fully plastic transition match the truth provided by held out crystal plasticity data. Lastly we demonstrate the efficacy of the trained model in material characterization and optimization tasks.
[{'version': 'v1', 'created': 'Tue, 8 Oct 2019 02:16:45 GMT'}]
2019-10-09
Ryan Jacobs, Tam Mayeshiba, Ben Afflerbach, Luke Miles, Max Williams, Matthew Turner, Raphael Finkel, Dane Morgan
The Materials Simulation Toolkit for Machine Learning (MAST-ML): an automated open source toolkit to accelerate data-driven materials research
Computational Materials Science, 176, 2020
10.1016/j.commatsci.2020.109544
null
physics.comp-ph cond-mat.mtrl-sci
As data science and machine learning methods are taking on an increasingly important role in the materials research community, there is a need for the development of machine learning software tools that are easy to use (even for nonexperts with no programming ability), provide flexible access to the most important algorithms, and codify best practices of machine learning model development and evaluation. Here, we introduce the Materials Simulation Toolkit for Machine Learning (MAST-ML), an open source Python-based software package designed to broaden and accelerate the use of machine learning in materials science research. MAST-ML provides predefined routines for many input setup, model fitting, and post-analysis tasks, as well as a simple structure for executing a multi-step machine learning model workflow. In this paper, we describe how MAST-ML is used to streamline and accelerate the execution of machine learning problems. We walk through how to acquire and run MAST-ML, demonstrate how to execute different components of a supervised machine learning workflow via a customized input file, and showcase a number of features and analyses conducted automatically during a MAST-ML run. Further, we demonstrate the utility of MAST-ML by showcasing examples of recent materials informatics studies which used MAST-ML to formulate and evaluate various machine learning models for an array of materials applications. Finally, we lay out a vision of how MAST-ML, together with complementary software packages and emerging cyberinfrastructure, can advance the rapidly growing field of materials informatics, with a focus on producing machine learning models easily, reproducibly, and in a manner that facilitates model evolution and improvement in the future.
[{'version': 'v1', 'created': 'Mon, 14 Oct 2019 17:20:56 GMT'}]
2020-06-26
Jeffrey M. Ede
Deep Learning Supersampled Scanning Transmission Electron Microscopy
null
null
null
eess.IV cond-mat.mtrl-sci cs.CV
Compressed sensing can increase resolution, and decrease electron dose and scan time of electron microscope point-scan systems with minimal information loss. Building on a history of successful deep learning applications in compressed sensing, we have developed a two-stage multiscale generative adversarial network to supersample scanning transmission electron micrographs with point-scan coverage reduced to 1/16, 1/25, ..., 1/100 px. We propose a novel non-adversarial learning policy to train a unified generator for multiple coverages and introduce an auxiliary network to homogenize prioritization of training data with varied signal-to-noise ratios. This achieves root mean square errors of 3.23% and 4.54% at 1/16 px and 1/100 px coverage, respectively; within 1% of errors for networks trained for each coverage individually. Detailed error distributions are presented for unified and individual coverage generators, including errors per output pixel. In addition, we present a baseline one-stage network for a single coverage and investigate numerical precision for web serving. Source code, training data, and pretrained models are publicly available at https://github.com/Jeffrey-Ede/DLSS-STEM
[{'version': 'v1', 'created': 'Wed, 23 Oct 2019 11:30:25 GMT'}, {'version': 'v2', 'created': 'Fri, 25 Oct 2019 09:39:05 GMT'}]
2019-10-28
Divya Kaushik, Utkarsh Singh, Upasana Sahu, Indu Sreedevi and Debanjan Bhowmik
Comparing domain wall synapse with other Non Volatile Memory devices for on-chip learning in Analog Hardware Neural Network
null
null
null
physics.app-ph cond-mat.mtrl-sci cs.NE
Resistive Random Access Memory (RRAM) and Phase Change Memory (PCM) devices have been popularly used as synapses in crossbar array based analog Neural Network (NN) circuit to achieve more energy and time efficient data classification compared to conventional computers. Here we demonstrate the advantages of recently proposed spin orbit torque driven Domain Wall (DW) device as synapse compared to the RRAM and PCM devices with respect to on-chip learning (training in hardware) in such NN. Synaptic characteristic of DW synapse, obtained by us from micromagnetic modeling, turns out to be much more linear and symmetric (between positive and negative update) than that of RRAM and PCM synapse. This makes design of peripheral analog circuits for on-chip learning much easier in DW synapse based NN compared to that for RRAM and PCM synapses. We next incorporate the DW synapse as a Verilog-A model in the crossbar array based NN circuit we design on SPICE circuit simulator. Successful on-chip learning is demonstrated through SPICE simulations on the popular Fisher's Iris dataset. Time and energy required for learning turn out to be orders of magnitude lower for DW synapse based NN circuit compared to that for RRAM and PCM synapse based NN circuits.
[{'version': 'v1', 'created': 'Mon, 28 Oct 2019 19:25:21 GMT'}]
2019-10-30
Chengqiang Lu, Qi Liu, Qiming Sun, Chang-Yu Hsieh, Shengyu Zhang, Liang Shi, and Chee-Kong Lee
Deep Learning for Optoelectronic Properties of Organic Semiconductors
J. Phys. Chem. C 2020, 124, 13, 7048
null
null
physics.chem-ph cond-mat.mtrl-sci physics.comp-ph
Atomistic modeling of energetic disorder in organic semiconductors (OSCs) and its effects on the optoelectronic properties of OSCs requires a large number of excited-state electronic-structure calculations, a computationally daunting task for many OSC applications. In this work, we advocate the use of deep learning to address this challenge and demonstrate that state-of-the-art deep neural networks (DNNs) are capable of predicting the electronic properties of OSCs at an accuracy comparable with the quantum chemistry methods used for generating training data. We extensively investigate the performances of four recent DNNs (deep tensor neural network, SchNet, message passing neural network, and multilevel graph convolutional neural network) in predicting various electronic properties of an important class of OSCs, i.e., oligothiophenes (OTs), including their HOMO and LUMO energies, excited-state energies and associated transition dipole moments. We find that SchNet shows the best performance for OTs of different sizes (from bithiophene to sexithiophene), achieving average prediction errors in the range of 20-80meV compared to the results from (time-dependent) density functional theory. We show that SchNet also consistently outperforms shallow feed-forward neural networks, especially in difficult cases with large molecules or limited training data. We further show that SchNet could predict the transition dipole moment accurately, a task previously known to be difficult for feed-forward neural networks, and we ascribe the relatively large errors in transition dipole prediction seen for some OT configurations to the charge-transfer character of their excited states. Finally, we demonstrate the effectiveness of SchNet by modeling the UV-Vis absorption spectra of OTs in dichloromethane and a good agreement is observed between the calculated and experimental spectra.
[{'version': 'v1', 'created': 'Tue, 29 Oct 2019 21:42:02 GMT'}]
2021-05-10
Ka-Ming Tam, Nicholas Walker, Samuel Kellar, Mark Jarrell
Interatomic Potential in a Simple Dense Neural Network Representation
null
null
null
physics.comp-ph cond-mat.mtrl-sci cond-mat.stat-mech
Simulations at the atomic scale provide a direct and effective way to understand the mechanical properties of materials. In the regime of classical mechanics, simulations for the thermodynamic properties of metals and alloys can be done by either solving the equations of motion or performing Monte Carlo sampling. The key component for an accurate simulation of such physical systems to produce faithful physical quantities is the use of an appropriate potential or a force field. In this paper, we explore the use of methods from the realm of machine learning to overcome and bypass difficulties encountered when fitting potentials for atomic systems. Particularly, we will show that classical potentials can be represented by a dense neural network with good accuracy.
[{'version': 'v1', 'created': 'Mon, 4 Nov 2019 17:51:13 GMT'}]
2019-11-05
Anh D. Phan, Cuong V. Nguyen, Pham T. Linh, Tran V. Huynh, Vu D. Lam, and Anh-Tuan Le
Deep Learning for The Inverse Design of Mid-infrared Graphene Plasmons
null
null
null
physics.app-ph cond-mat.mes-hall cond-mat.mtrl-sci physics.optics
We theoretically investigate the plasmonic properties of mid-infrared graphene-based metamaterials and apply deep learning of a neural network for the inverse design. These artificial structures have square periodic arrays of graphene plasmonic resonators deposited on dielectric thin films. Optical spectra vary significantly with changes in structural parameters. Our numerical results are in accordance with previous experiments. Then, the theoretical approach is employed to generate data for training and testing deep neural networks. By merging the pre-trained neural network with the inverse network, we implement calculations for inverse design of the graphene-based metameterials. We also discuss the limitation of the data-driven approach.
[{'version': 'v1', 'created': 'Thu, 28 Nov 2019 07:36:31 GMT'}, {'version': 'v2', 'created': 'Wed, 19 Feb 2020 06:17:32 GMT'}]
2020-02-20
So Takamoto, Satoshi Izumi, Ju Li
TeaNet: universal neural network interatomic potential inspired by iterative electronic relaxations
Computational Materials Science 207 (2022) 111280
10.1016/j.commatsci.2022.111280
null
physics.comp-ph cond-mat.mtrl-sci cs.LG stat.ML
A universal interatomic potential for an arbitrary set of chemical elements is urgently needed in computational materials science. Graph convolution neural network (GCN) has rich expressive power, but previously was mainly employed to transport scalars and vectors, not rank $\ge 2$ tensors. As classic interatomic potentials were inspired by tight-binding electronic relaxation framework, we want to represent this iterative propagation of rank $\ge 2$ tensor information by GCN. Here we propose an architecture called the tensor embedded atom network (TeaNet) where angular interaction is translated into graph convolution through the incorporation of Euclidean tensors, vectors and scalars. By applying the residual network (ResNet) architecture and training with recurrent GCN weights initialization, a much deeper (16 layers) GCN was constructed, whose flow is similar to an iterative electronic relaxation. Our traning dataset is generated by density functional theory calculation of mostly chemically and structurally randomized configurations. We demonstrate that arbitrary structures and reactions involving the first 18 elements on the periodic table (H to Ar) can be realized satisfactorily by TeaNet, including C-H molecular structures, metals, amorphous SiO${}_2$, and water, showing surprisingly good performance (energy mean absolute error 19 meV/atom) and robustness for arbitrary chemistries involving elements from H to Ar.
[{'version': 'v1', 'created': 'Mon, 2 Dec 2019 08:47:16 GMT'}, {'version': 'v2', 'created': 'Sun, 10 Oct 2021 18:34:11 GMT'}]
2022-03-17
Ruiyang Li, Eungkyu Lee, Tengfei Luo
A Unified Deep Neural Network Potential Capable of Predicting Thermal Conductivity of Silicon in Different Phases
null
null
null
cond-mat.mtrl-sci physics.comp-ph
Molecular dynamics simulations have been extensively used to predict thermal properties, but simulating different phases with similar precision using a unified force field is often difficult, due to the lack of accurate and transferrable interatomistic potential fields. As a result, this issue has become a major barrier to predicting the phase change of materials and their transport properties with atomistic-level modeling techniques. Recently, machine learning based algorithms have emerged as promising tools to develop accurate potentials for molecular dynamics simulations. In this work, we approach the problem of predicting the thermal conductivity of silicon in different phases by performing molecular dynamics simulations with a deep neural network potential. This neural network potential is trained with ab-initio data of silicon in the crystalline, liquid and amorphous phases. The accuracy of our potential is first validated through reproducing the atomistic structures during the phase transition, where other empirical potentials usually fail. The thermal conductivity of different phases is then calculated, showing a good agreement with the experimental results and ab-initio calculation results. Our work shows that a unified neural network-based potential can be a promising tool for studying phase change and thermal transport of materials with high accuracy.
[{'version': 'v1', 'created': 'Tue, 10 Dec 2019 23:17:49 GMT'}]
2019-12-12
James P. Horwath, Dmitri N. Zakharov, Remi Megret, Eric A. Stach
Understanding Important Features of Deep Learning Models for Transmission Electron Microscopy Image Segmentation
null
null
null
eess.IV cond-mat.mtrl-sci cs.LG
Cutting edge deep learning techniques allow for image segmentation with great speed and accuracy. However, application to problems in materials science is often difficult since these complex models may have difficultly learning physical parameters. In situ electron microscopy provides a clear platform for utilizing automated image analysis. In this work we consider the case of studying coarsening dynamics in supported nanoparticles, which is important for understanding e.g. the degradation of industrial catalysts. By systematically studying dataset preparation, neural network architecture, and accuracy evaluation, we describe important considerations in applying deep learning to physical applications, where generalizable and convincing models are required.
[{'version': 'v1', 'created': 'Thu, 12 Dec 2019 16:52:23 GMT'}]
2019-12-13
R. Ravinder, Karthikeya H. Sreedhara, Suresh Bishnoi, Hargun Singh Grover, Mathieu Bauchy, Jayadeva, Hariprasad Kodamana, N. M. Anoop Krishnan
Deep Learning Aided Rational Design of Oxide Glasses
null
null
null
cond-mat.mtrl-sci cond-mat.dis-nn physics.data-an
Despite the extensive usage of oxide glasses for a few millennia, the composition-property relationships in these materials still remain poorly understood. While empirical and physics-based models have been used to predict properties, these remain limited to a few select compositions or a series of glasses. Designing new glasses requires a priori knowledge of how the composition of a glass dictates its properties such as stiffness, density, or processability. Thus, accelerated design of glasses for targeted applications remain impeded due to the lack of universal composition-property models. Herein, using deep learning, we present a methodology for the rational design of oxide glasses. Exploiting a large dataset of glasses comprising of up to 37 oxide components and more than 100,000 glass compositions, we develop high-fidelity deep neural networks for the prediction of eight properties that enable the design of glasses, namely, density, Young's modulus, shear modulus, hardness, glass transition temperature, thermal expansion coefficient, liquidus temperature, and refractive index. These models are by far the most extensive models developed as they cover the entire range of human-made glass compositions. We demonstrate that the models developed here exhibit excellent predictability, ensuring close agreement with experimental observations. Using these models, we develop a series of new design charts, termed as glass selection charts. These charts enable the rational design of functional glasses for targeted applications by identifying unique compositions that satisfy two or more constraints, on both compositions and properties, simultaneously. The generic design approach presented herein could catalyze machine-learning assisted materials design and discovery for a large class of materials including metals, ceramics, and proteins.
[{'version': 'v1', 'created': 'Wed, 25 Dec 2019 03:04:59 GMT'}]
2019-12-30
Reza Rashetnia and Mohammad Pour-Ghaz
Deep learning surrogate interacting Markov chain Monte Carlo based full wave inversion scheme for properties of materials quantification
null
null
null
cond-mat.mtrl-sci cs.LG stat.ML
Full Wave Inversion (FWI) imaging scheme has many applications in engineering, geoscience and medical sciences. In this paper, a surrogate deep learning FWI approach is presented to quantify properties of materials using stress waves. Such inverse problems, in general, are ill-posed and nonconvex, especially in cases where the solutions exhibit shocks, heterogeneity, discontinuities, or large gradients. The proposed approach is proven efficient to obtain global minima responses in these cases. This approach is trained based on random sampled set of material properties and sampled trials around local minima, therefore, it requires a forward simulation can handle high heterogeneity, discontinuities and large gradients. High resolution Kurganov-Tadmor (KT) central finite volume method is used as forward wave propagation operator. Using the proposed framework, material properties of 2D media are quantified for several different situations. The results demonstrate the feasibility of the proposed method for estimating mechanical properties of materials with high accuracy using deep learning approaches.
[{'version': 'v1', 'created': 'Mon, 16 Dec 2019 19:43:51 GMT'}]
2020-01-08
Chia-Hao Lee (1), Abid Khan (2 and 3), Di Luo (2 and 3), Tatiane P. Santos (1), Chuqiao Shi (1), Blanka E. Janicek (1), Sangmin Kang (4), Wenjuan Zhu (4), Nahil A. Sobh (5), Andr\'e Schleife (1, 6 and 7), Bryan K. Clark (2), Pinshane Y. Huang (1 and 6) ((1) Department of Materials Science and Engineering, (2) Department of Physics, (3) These authors contributed equally to this work, (4) Department of Electrical and Computer Engineering, (5) Beckman Institute for Advanced Science and Technology, (6) Materials Research Laboratory, (7) National Center for Supercomputing Applications)
Deep Learning Enabled Strain Mapping of Single-Atom Defects in 2D Transition Metal Dichalcogenides with Sub-picometer Precision
null
10.1021/acs.nanolett.0c00269
null
cond-mat.mtrl-sci cond-mat.mes-hall
2D materials offer an ideal platform to study the strain fields induced by individual atomic defects, yet challenges associated with radiation damage have so-far limited electron microscopy methods to probe these atomic-scale strain fields. Here, we demonstrate an approach to probe single-atom defects with sub-picometer precision in a monolayer 2D transition metal dichalcogenide, WSe$_{2-2x}$Te$_{2x}$. We utilize deep learning to mine large datasets of aberration-corrected scanning transmission electron microscopy images to locate and classify point defects. By combining hundreds of images of nominally identical defects, we generate high signal-to-noise class-averages which allow us to measure 2D atomic coordinates with up to 0.3 pm precision. Our methods reveal that Se vacancies introduce complex, oscillating strain fields in the WSe$_{2-2x}$Te$_{2x}$ lattice which cannot be explained by continuum elastic theory. These results indicate the potential impact of computer vision for the development of high-precision electron microscopy methods for beam-sensitive materials.
[{'version': 'v1', 'created': 'Wed, 22 Jan 2020 19:05:53 GMT'}]
2020-04-29
Saaketh Desai, Samuel Temple Reeve, James F. Belak
Implementing a neural network interatomic model with performance portability for emerging exascale architectures
null
null
null
physics.comp-ph cond-mat.mtrl-sci
The two main thrusts of computational science are more accurate predictions and faster calculations; to this end, the zeitgeist in molecular dynamics (MD) simulations is pursuing machine learned and data driven interatomic models, e.g. neural network potentials, and novel hardware architectures, e.g. GPUs. Current implementations of neural network potentials are orders of magnitude slower than traditional interatomic models and while looming exascale computing offers the ability to run large, accurate simulations with these models, achieving portable performance for MD with new and varied exascale hardware requires rethinking traditional algorithms, using novel data structures, and library solutions. We re-implement a neural network interatomic model in CabanaMD, an MD proxy application, built on libraries developed for performance portability. Our implementation shows significantly improved on-node scaling in this complex kernel as compared to a current LAMMPS implementation, across both strong and weak scaling. Our single-source solution results in improved performance in many cases, with thread-scalability enabling simulations up to 21 million atoms on a single CPU node and 2 million atoms on a single GPU. We also explore parallelism and data layout choices (using flexible data structures called AoSoAs) and their effect on performance, seeing up to ~25% and ~10% improvements in performance on a GPU simply by choosing the right level of parallelism and data layout, respectively.
[{'version': 'v1', 'created': 'Fri, 31 Jan 2020 20:49:30 GMT'}, {'version': 'v2', 'created': 'Tue, 4 Feb 2020 19:22:38 GMT'}, {'version': 'v3', 'created': 'Fri, 21 Feb 2020 18:54:46 GMT'}]
2020-02-24
Maxim Ziatdinov, Udi Fuchs, James H.G. Owen, John N. Randall, Sergei V. Kalinin
Robust multi-scale multi-feature deep learning for atomic and defect identification in Scanning Tunneling Microscopy on H-Si(100) 2x1 surface
null
null
null
cond-mat.mtrl-sci physics.app-ph
The nature of the atomic defects on the hydrogen passivated Si (100) surface is analyzed using deep learning and scanning tunneling microscopy (STM). A robust deep learning framework capable of identifying atomic species, defects, in the presence of non-resolved contaminates, step edges, and noise is developed. The automated workflow, based on the combination of several networks for image assessment, atom-finding and defect finding, is developed to perform the analysis at different levels of description and is deployed on an operational STM platform. This is further extended to unsupervised classification of the extracted defects using the mean-shift clustering algorithm, which utilizes features automatically engineered from the combined output of neural networks. This combined approach allows the identification of localized and extended defects on the topographically non-uniform surfaces or real materials. Our approach is universal in nature and can be applied to other surfaces for building comprehensive libraries of atomic defects in quantum materials.
[{'version': 'v1', 'created': 'Tue, 11 Feb 2020 22:18:28 GMT'}]
2020-02-19
Christopher M. Andolina, Philip Williamson, and Wissam A. Saidi
Optimization and Validation of a Deep Learning CuZr Atomistic Potential: Robust Applications for Crystalline and Amorphous Phases with near-DFT Accuracy
J. Chem. Phys. 152, 154701 (2020)
10.1063/5.0005347
null
cond-mat.mtrl-sci
We show that a deep-learning neural network potential (DP) based on density functional theory (DFT) calculations can well describe Cu-Zr materials, an example of a binary alloy system that can coexist in several ordered intermetallics and as an amorphous phase. The complex phase diagram for Cu-Zr makes it a challenging system for traditional atomistic force-fields that fail to describe well the different properties and phases. Instead, we show that a DP approach using a large database with ~300k configurations can render results generally on par with DFT. The training set includes configurations of pristine and bulk elementary metals and intermetallics in the liquid and solid phases in addition to slab and amorphous configurations. The DP model was validated by comparing bulk properties such as lattice constants, elastic constants, bulk moduli, phonon spectra, surface energies to DFT values for identical structures. Further, we contrast the DP results with values obtained using well-established two embedded atom method potentials. Overall, our DP potential provides near DFT accuracy for the different Cu-Zr phases but with a fraction of its computational cost, thus enabling accurate computations of realistic atomistic models especially for the amorphous phase.
[{'version': 'v1', 'created': 'Mon, 17 Feb 2020 04:15:31 GMT'}]
2020-04-29
Chongze Hu, Yunxing Zuo, Chi Chen, Shyue Ping Ong, Jian Luo
Genetic Algorithm-Guided Deep Learning of Grain Boundary Diagrams: Addressing the Challenge of Five Degrees of Freedom
null
null
null
cond-mat.mtrl-sci physics.comp-ph
Grain boundaries (GBs) often control the processing and properties of polycrystalline materials. Here, a potentially transformative research is represented by constructing GB property diagrams as functions of temperature and bulk composition, also called "complexion diagrams," as a general materials science tool on par with phase diagrams. However, a GB has five macroscopic (crystallographic) degrees of freedom (DOFs). It is essentially a "mission impossible" to construct property diagrams for GBs as a function of five DOFs by either experiments or modeling. Herein, we combine isobaric semi-grand-canonical ensemble hybrid Monte Carlo and molecular dynamics (hybrid MC/MD) simulations with a genetic algorithm (GA) and deep neural network (DNN) models to tackle this grand challenge. The DNN prediction is ~108 faster than atomistic simulations, thereby enabling the construction of the property diagrams for millions of distinctly different GBs of five DOFs. Notably, excellent prediction accuracies have been achieved for not only symmetric-tilt and twist GBs, but also asymmetric-tilt and mixed tilt-twist GBs; the latter are more complex and much less understood, but they are ubiquitous and often limit the performance properties of real polycrystals as the weak links. The data-driven prediction of GB properties as function of temperature, bulk composition, and five crystallographic DOFs (i.e., in a 7D space) opens a new paradigm.
[{'version': 'v1', 'created': 'Tue, 25 Feb 2020 02:43:36 GMT'}]
2020-02-26
Xiaoyu Sun, Nathaniel J. Krakauer, Alexander Politowicz, Wei-Ting Chen, Qiying Li, Zuoyi Li, Xianjia Shao, Alfred Sunaryo, Mingren Shen, James Wang, Dane Morgan
Assessing Graph-based Deep Learning Models for Predicting Flash Point
Mol. Inf. 2020, 39, 1900101
10.1002/minf.201900101
null
physics.comp-ph cond-mat.mtrl-sci cs.LG
Flash points of organic molecules play an important role in preventing flammability hazards and large databases of measured values exist, although millions of compounds remain unmeasured. To rapidly extend existing data to new compounds many researchers have used quantitative structure-property relationship (QSPR) analysis to effectively predict flash points. In recent years graph-based deep learning (GBDL) has emerged as a powerful alternative method to traditional QSPR. In this paper, GBDL models were implemented in predicting flash point for the first time. We assessed the performance of two GBDL models, message-passing neural network (MPNN) and graph convolutional neural network (GCNN), by comparing methods. Our result shows that MPNN both outperforms GCNN and yields slightly worse but comparable performance with previous QSPR studies. The average R2 and Mean Absolute Error (MAE) scores of MPNN are, respectively, 2.3% lower and 2.0 K higher than previous comparable studies. To further explore GBDL models, we collected the largest flash point dataset to date, which contains 10575 unique molecules. The optimized MPNN gives a test data R2 of 0.803 and MAE of 17.8 K on the complete dataset. We also extracted 5 datasets from our integrated dataset based on molecular types (acids, organometallics, organogermaniums, organosilicons, and organotins) and explore the quality of the model in these classes.against 12 previous QSPR studies using more traditional
[{'version': 'v1', 'created': 'Wed, 26 Feb 2020 06:10:12 GMT'}]
2020-02-28
Brian DeCost, Jason Hattrick-Simpers, Zachary Trautt, Aaron Kusne, Eva Campo and Martin Green
Scientific AI in materials science: a path to a sustainable and scalable paradigm
null
null
null
cond-mat.mtrl-sci physics.comp-ph
Recently there has been an ever-increasing trend in the use of machine learning (ML) and artificial intelligence (AI) methods by the materials science, condensed matter physics, and chemistry communities. This perspective article identifies key scientific, technical, and social opportunities that the materials community must prioritize to consistently develop and leverage Scientific AI to provide a credible path towards the advancement of current materials-limited technologies. Here we highlight the intersections of these opportunities with a series of proposed paths forward. The opportunities are roughly sorted from scientific/technical (e.g., development of robust, physically meaningful multiscale material representations) to social (e.g., promoting an AI-ready workforce). The proposed paths forward range from developing new infrastructure and capabilities to deploying them in industry and academia. We provide a brief introduction to AI in materials science and engineering, followed by detailed discussions of each of the opportunities and paths forward.
[{'version': 'v1', 'created': 'Wed, 18 Mar 2020 20:59:05 GMT'}]
2020-03-20
Chunyang Wang, Guanglei Ding, Yitong Liu, Huolin L. Xin
0.71-{\AA} resolution electron tomography enabled by deep learning aided information recovery
null
null
null
cond-mat.mtrl-sci eess.IV physics.app-ph physics.ins-det
Electron tomography, as an important 3D imaging method, offers a powerful method to probe the 3D structure of materials from the nano- to the atomic-scale. However, as a grant challenge, radiation intolerance of the nanoscale samples and the missing-wedge-induced information loss and artifacts greatly hindered us from obtaining 3D atomic structures with high fidelity. Here, for the first time, by combining generative adversarial models with state-of-the-art network architectures, we demonstrate the resolution of electron tomography can be improved to 0.71 angstrom which is the highest three-dimensional imaging resolution that has been reported thus far. We also show it is possible to recover the lost information and remove artifacts in the reconstructed tomograms by only acquiring data from -50 to +50 degrees (44% reduction of dosage compared to -90 to +90 degrees full tilt series). In contrast to conventional methods, the deep learning model shows outstanding performance for both macroscopic objects and atomic features solving the long-standing dosage and missing-wedge problems in electron tomography. Our work provides important guidance for the application of machine learning methods to tomographic imaging and sheds light on its applications in other 3D imaging techniques.
[{'version': 'v1', 'created': 'Fri, 27 Mar 2020 07:16:30 GMT'}]
2020-03-30
Haotong Liang, Valentin Stanev, A. Gilad Kusne, Ichiro Takeuchi
CRYSPNet: Crystal Structure Predictions via Neural Network
Phys. Rev. Materials 4, 123802 (2020)
10.1103/PhysRevMaterials.4.123802
null
cond-mat.mtrl-sci stat.ML
Structure is the most basic and important property of crystalline solids; it determines directly or indirectly most materials characteristics. However, predicting crystal structure of solids remains a formidable and not fully solved problem. Standard theoretical tools for this task are computationally expensive and at times inaccurate. Here we present an alternative approach utilizing machine learning for crystal structure prediction. We developed a tool called Crystal Structure Prediction Network (CRYSPNet) that can predict the Bravais lattice, space group, and lattice parameters of an inorganic material based only on its chemical composition. CRYSPNet consists of a series of neural network models, using as inputs predictors aggregating the properties of the elements constituting the compound. It was trained and validated on more than 100,000 entries from the Inorganic Crystal Structure Database. The tool demonstrates robust predictive capability and outperforms alternative strategies by a large margin. Made available to the public (at https://github.com/AuroraLHT/cryspnet), it can be used both as an independent prediction engine or as a method to generate candidate structures for further computational and/or experimental validation.
[{'version': 'v1', 'created': 'Tue, 31 Mar 2020 16:05:18 GMT'}]
2021-01-04
Sehyun Chun, Sidhartha Roy, Yen Thi Nguyen, Joseph B. Choi, H.S. Udaykumar, Stephen S. Baek
Deep learning for synthetic microstructure generation in a materials-by-design framework for heterogeneous energetic materials
null
10.1038/s41598-020-70149-0
null
cond-mat.mtrl-sci cs.LG
The sensitivity of heterogeneous energetic (HE) materials (propellants, explosives, and pyrotechnics) is critically dependent on their microstructure. Initiation of chemical reactions occurs at hot spots due to energy localization at sites of porosities and other defects. Emerging multi-scale predictive models of HE response to loads account for the physics at the meso-scale, i.e. at the scale of statistically representative clusters of particles and other features in the microstructure. Meso-scale physics is infused in machine-learned closure models informed by resolved meso-scale simulations. Since microstructures are stochastic, ensembles of meso-scale simulations are required to quantify hot spot ignition and growth and to develop models for microstructure-dependent energy deposition rates. We propose utilizing generative adversarial networks (GAN) to spawn ensembles of synthetic heterogeneous energetic material microstructures. The method generates qualitatively and quantitatively realistic microstructures by learning from images of HE microstructures. We show that the proposed GAN method also permits the generation of new morphologies, where the porosity distribution can be controlled and spatially manipulated. Such control paves the way for the design of novel microstructures to engineer HE materials for targeted performance in a materials-by-design framework.
[{'version': 'v1', 'created': 'Sun, 5 Apr 2020 16:58:31 GMT'}]
2023-05-12
Arash Rabbani, Masoud Babaei, Reza Shams, Ying Da Wang, Traiwit Chung
DeePore: a deep learning workflow for rapid and comprehensive characterization of porous materials
Advances in Water Resources, 2020, 103787
10.1016/j.advwatres.2020.103787
null
cond-mat.mtrl-sci cs.LG
DeePore is a deep learning workflow for rapid estimation of a wide range of porous material properties based on the binarized micro-tomography images. By combining naturally occurring porous textures we generated 17700 semi-real 3-D micro-structures of porous geo-materials with size of 256^3 voxels and 30 physical properties of each sample are calculated using physical simulations on the corresponding pore network models. Next, a designed feed-forward convolutional neural network (CNN) is trained based on the dataset to estimate several morphological, hydraulic, electrical, and mechanical characteristics of the porous material in a fraction of a second. In order to fine-tune the CNN design, we tested 9 different training scenarios and selected the one with the highest average coefficient of determination (R^2) equal to 0.885 for 1418 testing samples. Additionally, 3 independent synthetic images as well as 3 realistic tomography images have been tested using the proposed method and results are compared with pore network modelling and experimental data, respectively. Tested absolute permeabilities had around 13 % relative error compared to the experimental data which is noticeable considering the accuracy of the direct numerical simulation methods such as Lattice Boltzmann and Finite Volume. The workflow is compatible with any physical size of the images due to its dimensionless approach and can be used to characterize large-scale 3-D images by averaging the model outputs for a sliding window that scans the whole geometry.
[{'version': 'v1', 'created': 'Sun, 3 May 2020 08:46:09 GMT'}, {'version': 'v2', 'created': 'Sat, 10 Oct 2020 09:06:32 GMT'}]
2020-10-13
Claudia Mangold, Shunda Chen, Giuseppe Barbalinardo, Joerg Behler, Pascal Pochet, Konstantinos Termentzidis, Yang Han, Laurent Chaput, David Lacroix, Davide Donadio
Transferability of neural network potentials for varying stoichiometry: phonons and thermal conductivity of Mn$_x$Ge$_y$ compounds
Journal of Applied Physics 127, 244901 (2020)
10.1063/5.0009550
null
cond-mat.mtrl-sci
Germanium manganese compounds exhibit a variety of stable and metastable phases with different stoichiometry. These materials entail interesting electronic, magnetic and thermal properties both in their bulk form and as heterostructures. Here we develop and validate a transferable machine learning potential, based on the high-dimensional neural network formalism, to enable the study of Mn$_x$Ge$_y$ materials over a wide range of compositions. We show that a neural network potential fitted on a minimal training set reproduces successfully the structural and vibrational properties and the thermal conductivity of systems with different local chemical environments, and it can be used to predict phononic effects in nanoscale heterostructures.
[{'version': 'v1', 'created': 'Tue, 19 May 2020 17:15:00 GMT'}]
2020-07-02
George S. Baggs, Paul Guerrier, Andrew Loeb, Jason C. Jones
Automated Copper Alloy Grain Size Evaluation Using a Deep-learning CNN
null
null
null
cs.CV cond-mat.mtrl-sci cs.LG stat.ML
Moog Inc. has automated the evaluation of copper (Cu) alloy grain size using a deep-learning convolutional neural network (CNN). The proof-of-concept automated image acquisition and batch-wise image processing offers the potential for significantly reduced labor, improved accuracy of grain evaluation, and decreased overall turnaround times for approving Cu alloy bar stock for use in flight critical aircraft hardware. A classification accuracy of 91.1% on individual sub-images of the Cu alloy coupons was achieved. Process development included minimizing the variation in acquired image color, brightness, and resolution to create a dataset with 12300 sub-images, and then optimizing the CNN hyperparameters on this dataset using statistical design of experiments (DoE). Over the development of the automated Cu alloy grain size evaluation, a degree of "explainability" in the artificial intelligence (XAI) output was realized, based on the decomposition of the large raw images into many smaller dataset sub-images, through the ability to explain the CNN ensemble image output via inspection of the classification results from the individual smaller sub-images.
[{'version': 'v1', 'created': 'Wed, 20 May 2020 13:13:38 GMT'}]
2020-05-21
Samad Hajinazar, Aidan Thorn, Ernesto D. Sandoval, Saba Kharabadze, Aleksey N. Kolmogorov
MAISE: Construction of neural network interatomic models and evolutionary structure optimization
null
10.1016/j.cpc.2020.107679
null
physics.comp-ph cond-mat.mtrl-sci
Module for ab initio structure evolution (MAISE) is an open-source package for materials modeling and prediction. The code's main feature is an automated generation of neural network (NN) interatomic potentials for use in global structure searches. The systematic construction of Behler-Parrinello-type NN models approximating ab initio energy and forces relies on two approaches introduced in our recent studies. An evolutionary sampling scheme for generating reference structures improves the NNs' mapping of regions visited in unconstrained searches, while a stratified training approach enables the creation of standardized NN models for multiple elements. A more flexible NN architecture proposed here expands the applicability of the stratified scheme for an arbitrary number of elements. The full workflow in the NN development is managed with a customizable 'MAISE-NET' wrapper written in Python. The global structure optimization capability in MAISE is based on an evolutionary algorithm applicable for nanoparticles, films, and bulk crystals. A multitribe extension of the algorithm allows for an efficient simultaneous optimization of nanoparticles in a given size range. Implemented structure analysis functions include fingerprinting with radial distribution functions and finding space groups with the SPGLIB tool. This work overviews MAISE's available features, constructed models, and confirmed predictions.
[{'version': 'v1', 'created': 'Mon, 25 May 2020 14:15:39 GMT'}, {'version': 'v2', 'created': 'Tue, 15 Sep 2020 02:25:40 GMT'}]
2020-10-26
Tarak K Patra, Troy D. Loeffler and Subramanian K R S Sankaranarayanan
Accelerating Copolymer Inverse Design using AI Gaming algorithm
null
null
null
cond-mat.soft cond-mat.mes-hall cond-mat.mtrl-sci
There exists a broad class of sequencing problems, for example, in proteins and polymers that can be formulated as a heuristic search algorithm that involve decision making akin to a computer game. AI gaming algorithms such as Monte Carlo tree search (MCTS) gained prominence after their exemplary performance in the computer Go game and are decision trees aimed at identifying the path (moves) that should be taken by the policy to reach the final winning or optimal solution. Major challenges in inverse sequencing problems are that the materials search space is extremely vast and property evaluation for each sequence is computationally demanding. Reaching an optimal solution by minimizing the total number of evaluations in a given design cycle is therefore highly desirable. We demonstrate that one can adopt this approach for solving the sequencing problem by developing and growing a decision tree, where each node in the tree is a candidate sequence whose fitness is directly evaluated by molecular simulations. We interface MCTS with MD simulations and use a representative example of designing a copolymer compatibilizer, where the goal is to identify sequence specific copolymers that lead to zero interfacial energy between two immiscible homopolymers. We apply the MCTS algorithm to polymer chain lengths varying from 10-mer to 30-mer, wherein the overall search space varies from 210 (1024) to 230 (~1 billion). In each case, we identify a target sequence that leads to zero interfacial energy within a few hundred evaluations demonstrating the scalability and efficiency of MCTS in exploring practical materials design problems with exceedingly vast chemical/material search space. Our MCTS-MD framework can be easily extended to several other polymer and protein inverse design problems, in particular, for cases where sequence-property data is either unavailable and/or is resource intensive.
[{'version': 'v1', 'created': 'Mon, 1 Jun 2020 21:27:55 GMT'}]
2020-06-08
Troy D Loeffler, Sukriti Manna, Tarak K Patra, Henry Chan, Badri Narayanan, and Subramanian Sankaranarayanan
Active Learning A Neural Network Model For Gold Clusters \& Bulk From Sparse First Principles Training Data
null
10.1002/cctc.202000774
null
physics.comp-ph cond-mat.mtrl-sci
Small metal clusters are of fundamental scientific interest and of tremendous significance in catalysis. These nanoscale clusters display diverse geometries and structural motifs depending on the cluster size; a knowledge of this size-dependent structural motifs and their dynamical evolution has been of longstanding interest. Classical MD typically employ predefined functional forms which limits their ability to capture such complex size-dependent structural and dynamical transformation. Neural Network (NN) based potentials represent flexible alternatives and in principle, well-trained NN potentials can provide high level of flexibility, transferability and accuracy on-par with the reference model used for training. A major challenge, however, is that NN models are interpolative and requires large quantities of training data to ensure that the model adequately samples the energy landscape both near and far-from-equilibrium. Here, we introduce an active learning (AL) scheme that trains a NN model on-the-fly with minimal amount of first-principles based training data. Our AL workflow is initiated with a sparse training dataset (1 to 5 data points) and is updated on-the-fly via a Nested Ensemble Monte Carlo scheme that iteratively queries the energy landscape in regions of failure and updates the training pool to improve the network performance. Using a representative system of gold clusters, we demonstrate that our AL workflow can train a NN with ~500 total reference calculations. Our NN predictions are within 30 meV/atom and 40 meV/\AA of the reference DFT calculations. Moreover, our AL-NN model also adequately captures the various size-dependent structural and dynamical properties of gold clusters in excellent agreement with DFT calculations and available experiments.
[{'version': 'v1', 'created': 'Fri, 5 Jun 2020 20:44:30 GMT'}]
2020-07-21
Henry Chan, Youssef S.G. Nashed, Saugat Kandel, Stephan Hruszkewycz, Subramanian Sankaranarayanan, Ross J. Harder, Mathew J. Cherukara
Real-time 3D Nanoscale Coherent Imaging via Physics-aware Deep Learning
null
10.1063/5.0031486
null
eess.IV cond-mat.mtrl-sci cs.LG physics.app-ph
Phase retrieval, the problem of recovering lost phase information from measured intensity alone, is an inverse problem that is widely faced in various imaging modalities ranging from astronomy to nanoscale imaging. The current process of phase recovery is iterative in nature. As a result, the image formation is time-consuming and computationally expensive, precluding real-time imaging. Here, we use 3D nanoscale X-ray imaging as a representative example to develop a deep learning model to address this phase retrieval problem. We introduce 3D-CDI-NN, a deep convolutional neural network and differential programming framework trained to predict 3D structure and strain solely from input 3D X-ray coherent scattering data. Our networks are designed to be "physics-aware" in multiple aspects; in that the physics of x-ray scattering process is explicitly enforced in the training of the network, and the training data are drawn from atomistic simulations that are representative of the physics of the material. We further refine the neural network prediction through a physics-based optimization procedure to enable maximum accuracy at lowest computational cost. 3D-CDI-NN can invert a 3D coherent diffraction pattern to real-space structure and strain hundreds of times faster than traditional iterative phase retrieval methods, with negligible loss in accuracy. Our integrated machine learning and differential programming solution to the phase retrieval problem is broadly applicable across inverse problems in other application areas.
[{'version': 'v1', 'created': 'Tue, 16 Jun 2020 18:35:32 GMT'}]
2024-06-12
Marco Eckhoff, Florian Sch\"onewald, Marcel Risch, Cynthia A. Volkert, Peter E. Bl\"ochl, J\"org Behler
Closing the gap between theory and experiment for lithium manganese oxide spinels using a high-dimensional neural network potential
Phys. Rev. B 102, 174102 (2020)
10.1103/PhysRevB.102.174102
null
cond-mat.mtrl-sci
Many positive electrode materials in lithium ion batteries include transition metals which are difficult to describe by electronic structure methods like density functional theory (DFT) due to the presence of multiple oxidation states. A prominent example is the lithium manganese oxide spinel Li$_x$Mn$_2$O$_4$ with $0\leq x\leq2$. While DFT, employing the local hybrid functional PBE0r, provides a reliable description, the need for extended computer simulations of large structural models remains a significant challenge. Here, we close this gap by constructing a DFT-based high-dimensional neural network potential (HDNNP) providing accurate energies and forces at a fraction of the computational costs. As different oxidation states and the resulting Jahn-Teller distortions represent a new level of complexity for HDNNPs, the potential is carefully validated by performing X-ray diffraction experiments. We demonstrate that the HDNNP provides atomic level details and is able to predict a series of properties like the lattice parameters and expansion with increasing Li content or temperature, the orthorhombic to cubic transition, the lithium diffusion barrier, and the phonon frequencies. We show that for understanding these properties access to large time and length scales as enabled by the HDNNP is essential to close the gap between theory and experiment.
[{'version': 'v1', 'created': 'Wed, 1 Jul 2020 08:44:44 GMT'}, {'version': 'v2', 'created': 'Fri, 2 Oct 2020 16:40:30 GMT'}]
2020-11-11
Haotian Feng and Pavana Prabhakar
Difference-Based Deep Learning Framework for Stress Predictions in Heterogeneous Media
null
10.1016/j.compstruct.2021.113957
null
physics.app-ph cond-mat.mtrl-sci cs.LG
Stress analysis of heterogeneous media, like composite materials, using Finite Element Analysis (FEA) has become commonplace in design and analysis. However, determining stress distributions in heterogeneous media using FEA can be computationally expensive in situations like optimization and multi-scaling. To address this, we utilize Deep Learning for developing a set of novel Difference-based Neural Network (DiNN) frameworks based on engineering and statistics knowledge to determine stress distribution in heterogeneous media, for the first time, with special focus on discontinuous domains that manifest high stress concentrations. The novelty of our approach is that instead of directly using several FEA model geometries and stresses as inputs for training a Neural Network, as typically done previously, we focus on highlighting the differences in stress distribution between different input samples for improving the accuracy of prediction in heterogeneous media. We evaluate the performance of DiNN frameworks by considering different types of geometric models that are commonly used in the analysis of composite materials, including volume fraction and spatial randomness. Results show that the DiNN structures significantly enhance the accuracy of stress prediction compared to existing structures, especially for composite models with random volume fraction when localized high stress concentrations are present.
[{'version': 'v1', 'created': 'Wed, 1 Jul 2020 00:18:14 GMT'}, {'version': 'v2', 'created': 'Wed, 15 Jul 2020 03:30:14 GMT'}, {'version': 'v3', 'created': 'Mon, 29 Mar 2021 12:01:43 GMT'}]
2021-04-22
Mart\'in Leandro Paleico, J\"org Behler
Global Optimization of Copper Clusters at the ZnO(10-10) Surface Using a DFT-based Neural Network Potential and Genetic Algorithms
null
null
null
physics.chem-ph cond-mat.mtrl-sci physics.comp-ph
The determination of the most stable structures of metal clusters supported at solid surfaces by computer simulations represents a formidable challenge due to the complexity of the potential-energy surface. Here we combine a high-dimensional neural network potential, which allows to predict the energies and forces of a large number of structures with first-principles accuracy, with a global optimization scheme employing genetic algorithms. This very efficient setup is used to identify the global minima and low-energy local minima for a series of copper clusters containing between four and ten atoms adsorbed at the ZnO(10$\bar{1}$0) surface. A series of structures with common structural features resembling the Cu(111) and Cu(110) surfaces at the metal-oxide interface has been identified, and the geometries of the emerging clusters are characterized in detail. We demonstrate that the frequently employed approximation of a frozen substrate surface in global optimization can result in missing the most relevant structures.
[{'version': 'v1', 'created': 'Mon, 13 Jul 2020 15:50:20 GMT'}]
2020-07-14
Ryan Cohn (1) and Elizabeth Holm (1) ((1) Department of Materials Science and Engineering, Carnegie Mellon University, Pittsburgh, PA, USA)
Unsupervised machine learning via transfer learning and k-means clustering to classify materials image data
null
10.1007/s40192-021-00205-8
null
cond-mat.mtrl-sci cs.LG eess.IV
Unsupervised machine learning offers significant opportunities for extracting knowledge from unlabeled data sets and for achieving maximum machine learning performance. This paper demonstrates how to construct, use, and evaluate a high performance unsupervised machine learning system for classifying images in a popular microstructural dataset. The Northeastern University Steel Surface Defects Database includes micrographs of six different defects observed on hot-rolled steel in a format that is convenient for training and evaluating models for image classification. We use the VGG16 convolutional neural network pre-trained on the ImageNet dataset of natural images to extract feature representations for each micrograph. After applying principal component analysis to extract signal from the feature descriptors, we use k-means clustering to classify the images without needing labeled training data. The approach achieves $99.4\% \pm 0.16\%$ accuracy, and the resulting model can be used to classify new images without retraining This approach demonstrates an improvement in both performance and utility compared to a previous study. A sensitivity analysis is conducted to better understand the influence of each step on the classification performance. The results provide insight toward applying unsupervised machine learning techniques to problems of interest in materials science.
[{'version': 'v1', 'created': 'Thu, 16 Jul 2020 14:36:04 GMT'}]
2021-04-13
Joohwi Lee and Ryoji Asahi
Transfer learning for materials informatics using crystal graph convolutional neural network
null
10.1016/j.commatsci.2021.110314
null
cond-mat.mtrl-sci physics.comp-ph
For successful applications of machine learning in materials informatics, it is necessary to overcome the inaccuracy of predictions ascribed to insufficient amount of data. In this study, we propose a transfer learning using a crystal graph convolutional neural network (TL-CGCNN). Herein, TL-CGCNN is pretrained with big data such as formation energies for crystal structures, and then used for predicting target properties with relatively small data. We confirm that TL-CGCNN can improve predictions of various properties such as bulk moduli, dielectric constants, and quasiparticle band gaps, which are computationally demanding, to construct big data for materials. Moreover, we quantitatively observe that the prediction of properties in target models via TL-CGCNN becomes more accurate with an increase in size of training dataset in pretrained models. Finally, we confirm that TL-CGCNN is superior to other regression methods in the predictions of target properties, which suffer from small amount of data. Therefore, we conclude that TL-CGCNN is promising along with compiling big data for materials that are easy to accumulate and relevant to the target properties.
[{'version': 'v1', 'created': 'Mon, 20 Jul 2020 08:27:57 GMT'}, {'version': 'v2', 'created': 'Tue, 18 Aug 2020 01:51:07 GMT'}, {'version': 'v3', 'created': 'Tue, 10 Nov 2020 02:38:18 GMT'}, {'version': 'v4', 'created': 'Fri, 29 Jan 2021 08:05:58 GMT'}]
2021-02-01
Zachary D. McClure and Alejandro H. Strachan
Expanding materials selection via transfer learning for high-temperature oxide selection
null
null
null
cond-mat.mtrl-sci
Materials with higher operating temperatures than today's state of the art can improve system performance in several applications and enable new technologies. Under most scenarios, a protective oxide scale with high melting temperatures and thermodynamic stability as well as low ionic diffusivity is required. Thus, the design of high-temperature systems would benefit from knowledge of these properties and related ones for all known oxides. While some properties of interest are known for many oxides (e.g. elastic constants exist for over 1,000 oxides), melting temperature is known for a relatively small subset. The determination of melting temperatures is time consuming and costly, both experimentally and computationally, thus we use data science tools to develop predictive models from the existing data. The relatively small number of available melting temperature values precludes the use of standard tools; therefore, we use a multi-step approach based on sequential learning where surrogate data from first-principles calculations is leveraged to develop models using small datasets. We use these models to predict the desired properties for nearly 11,000 oxides and quantify uncertainties in the space.
[{'version': 'v1', 'created': 'Mon, 20 Jul 2020 16:45:59 GMT'}]
2020-07-27
Shreshth A. Malik, Rhys E. A. Goodall, Alpha A. Lee
Materials Graph Transformer predicts the outcomes of inorganic reactions with reliable uncertainties
Chemistry of Materials 2021 33 (2), 616-624
10.1021/acs.chemmater.0c03885
null
physics.comp-ph cond-mat.mtrl-sci
A common bottleneck for materials discovery is synthesis. While recent methodological advances have resulted in major improvements in the ability to predicatively design novel materials, researchers often still rely on trial-and-error approaches for determining synthesis procedures. In this work, we develop a model that predicts the major product of solid-state reactions. The cardinal feature of this approach is the construction of fixed-length, learned representations of reactions. Precursors are represented as nodes on a `reaction graph', and message-passing operations between nodes are used to embody the interactions between precursors in the reaction mixture. Through an ablation study, it is shown that this framework not only outperforms less physically-motivated baseline methods but also more reliably assesses the uncertainty in its predictions.
[{'version': 'v1', 'created': 'Thu, 30 Jul 2020 21:39:58 GMT'}, {'version': 'v2', 'created': 'Thu, 3 Sep 2020 17:08:03 GMT'}]
2021-01-27
Anran Wei, Jie Xiong, Weidong Yang, Fenglin Guo
Identifying the elastic isotropy of architectured materials based on deep learning method
null
10.1016/j.eml.2021.101173
null
physics.app-ph cond-mat.dis-nn cond-mat.mtrl-sci
With the achievement on the additive manufacturing, the mechanical properties of architectured materials can be precisely designed by tailoring microstructures. As one of the primary design objectives, the elastic isotropy is of great significance for many engineering applications. However, the prevailing experimental and numerical methods are normally too costly and time-consuming to determine the elastic isotropy of architectured materials with tens of thousands of possible microstructures in design space. The quick mechanical characterization is thus desired for the advanced design of architectured materials. Here, a deep learning-based approach is developed as a portable and efficient tool to identify the elastic isotropy of architectured materials directly from the images of their representative microstructures with arbitrary component distributions. The measure of elastic isotropy for architectured materials is derived firstly in this paper to construct a database with associated images of microstructures. Then a convolutional neural network is trained with the database. It is found that the convolutional neural network shows good performance on the isotropy identification. Meanwhile, it exhibits enough robustness to maintain the performance under fluctuated material properties in practical fabrications. Moreover, the well-trained convolutional neural network can be successfully transferred among different types of architectured materials, including two-phase composites and porous materials, which greatly enhance the efficiency of the deep learning-based approach. This study can give new inspirations on the fast mechanical characterization for the big-data driven design of architectured materials.
[{'version': 'v1', 'created': 'Sun, 2 Aug 2020 10:16:28 GMT'}, {'version': 'v2', 'created': 'Sun, 9 Aug 2020 03:15:22 GMT'}]
2021-04-15
Shusuke Kasamatsu, Yuichi Motoyama, Kazuyoshi Yoshimi, Ushio Matsumoto, Akihide Kuwabara, and Takafumi Ogawa
Facilitating {\it ab initio} configurational sampling of multicomponent solids using an on-lattice neural network model and active learning
null
10.1063/5.0096645
null
physics.comp-ph cond-mat.mtrl-sci
We propose a scheme for {\it ab initio} configurational sampling in multicomponent crystalline solids using Behler-Parinello type neural network potentials (NNPs) in an unconventional way: the NNPs are trained to predict the energies of relaxed structures from the perfect lattice with configurational disorder instead of the usual way of training to predict energies as functions of continuous atom coordinates. An active learning scheme is employed to obtain a training set containing configurations of thermodynamic relevance. This enables bypassing of the structural relaxation procedure which is necessary when applying conventional NNP approaches to the lattice configuration problem. The idea is demonstrated on the calculation of the temperature dependence of the degree of A/B site inversion in three spinel oxides, MgAl$_2$O$_4$, ZnAl$_2$O$_4$, and MgGa$_2$O$_4$. The present scheme may serve as an alternative to cluster expansion for `difficult' systems, e.g., complex bulk or interface systems with many components and sublattices that are relevant to many technological applications today.
[{'version': 'v1', 'created': 'Thu, 6 Aug 2020 11:07:32 GMT'}, {'version': 'v2', 'created': 'Wed, 20 Apr 2022 12:33:53 GMT'}]
2024-06-19
Koji Shimizu, Elvis F. Arguelles, Wenwen Li, Yasunobu Ando, Emi Minamitani, and Satoshi Watanabe
Phase stability of Au-Li binary systems studied using neural network potential
Phys. Rev. B 103, 094112 (2021)
10.1103/PhysRevB.103.094112
null
cond-mat.mtrl-sci
The miscibility of Au and Li exhibits a potential application as an adhesion layer and electrode material in secondary batteries. Here, to explore alloying properties, we constructed a neural network potential (NNP) of Au-Li binary systems based on density functional theory (DFT) calculations. To accelerate construction of NNPs, we proposed an efficient and inexpensive method of structural dataset generation. The predictions by the constructed NNP on lattice parameters and phonon properties agree well with those obtained by DFT calculations. We also investigated the mixing energy of Au$_{1-x}$Li$_{x}$ with fine composition grids, showing excellent agreement with DFT verifications. We found the existence of various compositions with structures on and slightly above the convex hull, which can explain the lack of consensus on the Au-Li stable phases in previous studies. Moreover, we newly found Au$_{0.469}$Li$_{0.531}$ as a stable phase, which has never been reported elsewhere. Finally, we examined the alloying process starting from the phase separated structure to the complete mixing phase. We found that when multiple adjacent Au atoms dissolved into Li, the alloying of the entire Au/Li interface started from the dissolved region. This paper demonstrates the applicability of NNPs toward miscible phases and provides the understanding of the alloying mechanism.
[{'version': 'v1', 'created': 'Wed, 12 Aug 2020 03:57:09 GMT'}]
2021-03-31
Mohammadreza Karamad, Rishikesh Magar, Yuting Shi, Samira Siahrostami, Ian D. Gates and Amir Barati Farimani
Orbital Graph Convolutional Neural Network for Material Property Prediction
Phys. Rev. Materials 4, 093801 (2020)
10.1103/PhysRevMaterials.4.093801
null
physics.comp-ph cond-mat.mtrl-sci cs.LG
Material representations that are compatible with machine learning models play a key role in developing models that exhibit high accuracy for property prediction. Atomic orbital interactions are one of the important factors that govern the properties of crystalline materials, from which the local chemical environments of atoms is inferred. Therefore, to develop robust machine learningmodels for material properties prediction, it is imperative to include features representing such chemical attributes. Here, we propose the Orbital Graph Convolutional Neural Network (OGCNN), a crystal graph convolutional neural network framework that includes atomic orbital interaction features that learns material properties in a robust way. In addition, we embedded an encoder-decoder network into the OGCNN enabling it to learn important features among basic atomic (elemental features), orbital-orbital interactions, and topological features. We examined the performance of this model on a broad range of crystalline material data to predict different properties. We benchmarked the performance of the OGCNN model with that of: 1) the crystal graph convolutional neural network (CGCNN), 2) other state-of-the-art descriptors for material representations including Many-body Tensor Representation (MBTR) and the Smooth Overlap of Atomic Positions (SOAP), and 3) other conventional regression machine learning algorithms where different crystal featurization methods have been used. We find that OGCNN significantly outperforms them. The OGCNN model with high predictive accuracy can be used to discover new materials among the immense phase and compound spaces of materials
[{'version': 'v1', 'created': 'Fri, 14 Aug 2020 15:22:22 GMT'}]
2020-10-06
Kaiqi Yang, Yifan Cao, Youtian Zhang, Ming Tang, Daniel Aberg, Babak Sadigh, Fei Zhou
Self-Supervised Learning and Prediction of Microstructure Evolution with Recurrent Neural Networks
null
null
null
cond-mat.mtrl-sci physics.comp-ph
Microstructural evolution is a key aspect of understanding and exploiting the structure-property-performance relation of materials. Modeling microstructure evolution usually relies on coarse-grained simulations with evolution principles described by partial differential equations (PDEs). Here we demonstrate that convolutional recurrent neural networks can learn the underlying physical rules and replace PDE-based simulations in the prediction of microstructure phenomena. Neural nets are trained by self-supervised learning with image sequences from simulations of several common processes, including plane wave propagation, grain growth, spinodal decomposition and dendritic crystal growth. The trained networks can accurately predict both short-term local dynamics and long-term statistical properties of microstructures and is capable of extrapolating beyond the training datasets in spatiotemporal domains and configurational and parametric spaces. Such a data-driven approach offers significant advantages over PDE-based simulations in time stepping efficiency and offers a useful alternative especially when the material parameters or governing PDEs are not well determined.
[{'version': 'v1', 'created': 'Mon, 17 Aug 2020 22:46:12 GMT'}, {'version': 'v2', 'created': 'Sun, 30 Aug 2020 01:09:41 GMT'}]
2020-09-01
Rongzhi Dong, Yabo Dan, Xiang Li, Jianjun Hu
Inverse Design of Composite Metal Oxide Optical Materials based on Deep Transfer Learning
Computational Materials Science (2020): 110166
10.1016/j.commatsci.2020.110166
null
cond-mat.mtrl-sci
Optical materials with special optical properties are widely used in a broad span of technologies, from computer displays to solar energy utilization leading to large dataset accumulated from years of extensive materials synthesis and optical characterization. Previously, machine learning models have been developed to predict the optical absorption spectrum from a materials characterization image or vice versa. Herein we propose TLOpt, a transfer learning based inverse optical materials design algorithm for suggesting material compositions with a desired target light absorption spectrum. Our approach is based on the combination of a deep neural network model and global optimization algorithms including a genetic algorithm and Bayesian optimization. A transfer learning strategy is employed to solve the small dataset issue in training the neural network predictor of optical absorption spectrum using the Magpie materials composition descriptor. Our extensive experiments show that our algorithm can inverse design the materials composition with stoichiometry with high accuracy.
[{'version': 'v1', 'created': 'Mon, 24 Aug 2020 18:00:14 GMT'}]
2020-11-26
Juhyeok Lee and Chaehwa Jeong and Yongsoo Yang
Single-atom level determination of 3-dimensional surface atomic structure via neural network-assisted atomic electron tomography
Nature Communications 12, 1962 (2021)
10.1038/s41467-021-22204-1
null
cond-mat.mtrl-sci
Functional properties of nanomaterials strongly depend on their surface atomic structure, but they often become largely different from their bulk structure, exhibiting surface reconstructions and relaxations. However, most of the surface characterization methods are either limited to 2-dimensional measurements or not reaching to true 3D atomic-scale resolution, and single-atom level determination of the 3D surface atomic structure for general 3D nanomaterials still remains elusive. Here we show the measurement of 3D atomic structure of a Pt nanoparticle at 15 pm precision, aided by a deep learning-based missing data retrieval. The surface atomic structure was reliably measured, and we find that <100> and <111> facets contribute differently to the surface strain, resulting in anisotropic strain distribution as well as compressive support boundary effect. The capability of single-atom level surface characterization will not only deepen our understanding of the functional properties of nanomaterials but also open a new door for fine tailoring of their performance.
[{'version': 'v1', 'created': 'Thu, 27 Aug 2020 10:08:28 GMT'}]
2021-10-01
Tsz Wai Ko, Jonas A. Finkler, Stefan Goedecker and J\"org Behler
A Fourth-Generation High-Dimensional Neural Network Potential with Accurate Electrostatics Including Non-local Charge Transfer
null
10.1038/s41467-020-20427-2
null
cond-mat.mtrl-sci physics.chem-ph physics.comp-ph
Machine learning potentials have become an important tool for atomistic simulations in many fields, from chemistry via molecular biology to materials science. Most of the established methods, however, rely on local properties and are thus unable to take global changes in the electronic structure into account, which result from long-range charge transfer or different charge states. In this work we overcome this limitation by introducing a fourth-generation high-dimensional neural network potential that combines a charge equilibration scheme employing environment-dependent atomic electronegativities with accurate atomic energies. The method, which is able to correctly describe global charge distributions in arbitrary systems, yields much improved energies and substantially extends the applicability of modern machine learning potentials. This is demonstrated for a series of systems representing typical scenarios in chemistry and materials science that are incorrectly described by current methods, while the fourth-generation neural network potential is in excellent agreement with electronic structure calculations.
[{'version': 'v1', 'created': 'Mon, 14 Sep 2020 14:43:31 GMT'}]
2021-03-17
G.P. Purja Pun, V. Yamakov, J. Hickman, E. H. Glaessgen, Y. Mishin
Development of a general-purpose machine-learning interatomic potential for aluminum by the physically-informed neural network method
Phys. Rev. Materials 4, 113807 (2020)
10.1103/PhysRevMaterials.4.113807
null
physics.comp-ph cond-mat.mtrl-sci
Abstract Interatomic potentials constitute the key component of large-scale atomistic simulations of materials. The recently proposed physically-informed neural network (PINN) method combines a high-dimensional regression implemented by an artificial neural network with a physics-based bond-order interatomic potential applicable to both metals and nonmetals. In this paper, we present a modified version of the PINN method that accelerates the potential training process and further improves the transferability of PINN potentials to unknown atomic environments. As an application, a modified PINN potential for Al has been developed by training on a large database of electronic structure calculations. The potential reproduces the reference first-principles energies within 2.6 meV per atom and accurately predicts a wide spectrum of physical properties of Al. Such properties include, but are not limited to, lattice dynamics, thermal expansion, energies of point and extended defects, the melting temperature, the structure and dynamic properties of liquid Al, the surface tensions of the liquid surface and the solid-liquid interface, and the nucleation and growth of a grain boundary crack. Computational efficiency of PINN potentials is also discussed.
[{'version': 'v1', 'created': 'Mon, 14 Sep 2020 15:59:28 GMT'}, {'version': 'v2', 'created': 'Mon, 26 Oct 2020 17:11:19 GMT'}, {'version': 'v3', 'created': 'Mon, 9 Nov 2020 12:17:48 GMT'}]
2020-11-25
Jeffrey M. Ede
Review: Deep Learning in Electron Microscopy
null
10.1088/2632-2153/abd614
null
eess.IV cond-mat.mtrl-sci cs.CV cs.LG
Deep learning is transforming most areas of science and technology, including electron microscopy. This review paper offers a practical perspective aimed at developers with limited familiarity. For context, we review popular applications of deep learning in electron microscopy. Afterwards, we discuss hardware and software needed to get started with deep learning and interface with electron microscopes. We then review neural network components, popular architectures, and their optimization. Finally, we discuss future directions of deep learning in electron microscopy.
[{'version': 'v1', 'created': 'Thu, 17 Sep 2020 14:23:55 GMT'}, {'version': 'v2', 'created': 'Fri, 18 Sep 2020 14:38:31 GMT'}, {'version': 'v3', 'created': 'Mon, 5 Oct 2020 07:22:38 GMT'}, {'version': 'v4', 'created': 'Tue, 3 Nov 2020 15:38:31 GMT'}, {'version': 'v5', 'created': 'Sun, 27 Dec 2020 22:14:24 GMT'}, {'version': 'v6', 'created': 'Thu, 31 Dec 2020 18:28:54 GMT'}, {'version': 'v7', 'created': 'Mon, 8 Mar 2021 10:12:04 GMT'}]
2021-03-09
Mahmudul Islam, Md Shajedul Hoque Thakur, Satyajit Mojumder and Mohammad Nasim Hasan
Extraction of Material Properties through Multi-fidelity Deep Learning from Molecular Dynamics Simulation
null
10.1016/j.commatsci.2020.110187
null
physics.comp-ph cond-mat.dis-nn cond-mat.mtrl-sci
Simulation of reasonable timescales for any long physical process using molecular dynamics (MD) is a major challenge in computational physics. In this study, we have implemented an approach based on multi-fidelity physics informed neural network (MPINN) to achieve long-range MD simulation results over a large sample space with significantly less computational cost. The fidelity of our present multi-fidelity study is based on the integration timestep size of MD simulations. While MD simulations with larger timestep produce results with lower level of accuracy, it can provide enough computationally cheap training data for MPINN to learn an accurate relationship between these low-fidelity results and high-fidelity MD results obtained using smaller simulation timestep. We have performed two benchmark studies, involving one and two component LJ systems, to determine the optimum percentage of high-fidelity training data required to achieve accurate results with high computational saving. The results show that important system properties such as system energy per atom, system pressure and diffusion coefficients can be determined with high accuracy while saving 68% computational costs. Finally, as a demonstration of the applicability of our present methodology in practical MD studies, we have studied the viscosity of argon-copper nanofluid and its variation with temperature and volume fraction by MD simulation using MPINN. Then we have compared them with numerous previous studies and theoretical models. Our results indicate that MPINN can predict accurate nanofluid viscosity at a wide range of sample space with significantly small number of MD simulations. Our present methodology is the first implementation of MPINN in conjunction with MD simulation for predicting nanoscale properties. This can pave pathways to investigate more complex engineering problems that demand long-range MD simulations.
[{'version': 'v1', 'created': 'Mon, 28 Sep 2020 07:32:24 GMT'}, {'version': 'v2', 'created': 'Tue, 29 Sep 2020 03:14:41 GMT'}, {'version': 'v3', 'created': 'Sat, 14 Nov 2020 22:44:06 GMT'}]
2020-12-08
End of preview. Expand in Data Studio

Dataset Card for "Materials-Informatics"

Dataset Name: Materials-Informatics

Dataset Owner: cs-mubashir

Language: English

Size: ~600+ entries

Last Updated: May 2025

Source: Extracted from arxiv dataset research repository

Dataset Summary

The Materials-Informatics dataset is a curated collection of research papers from arxiv repository focusing on the intersection of artificial intelligence (AI) and materials science and engineering (MSE). Each entry provides metadata and descriptive information about a research paper, including its title, authors, abstract, keywords, publication year, material types, AI techniques used, and application domains.

This dataset aims to serve as a valuable resource for researchers and practitioners working at the convergence of machine learning, deep learning, and materials discovery/design. It can be used for tasks like information retrieval, scientific NLP, trend analysis, paper classification, and LLM fine-tuning for domain-specific tasks.

Downloads last month
1