Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 7-Day Trial for You or Your Team.

Learn More →

Machine learning bandgaps of double perovskites

Machine learning bandgaps of double perovskites www.nature.com/scientificreports OPEN Machine learning bandgaps of double perovskites 1 2 1 2 3 G. Pilania , A. Mannodi-Kanakkithodi , B. P. Uberuaga , R. Ramprasad , J. E. Gubernatis & T. Lookman received: 29 October 2015 The ability to make rapid and accurate predictions on bandgaps of double perovskites is of much a ccepted: 07 December 2015 practical interest for a range of applications. While quantum mechanical computations for high-fidelity Published: 19 January 2016 bandgaps are enormously computation-time intensive and thus impractical in high throughput studies, informatics-based statistical learning approaches can be a promising alternative. Here we demonstrate a systematic feature-engineering approach and a robust learning framework for efficient and accurate predictions of electronic bandgaps of double perovskites. After evaluating a set of more than 1.2 million features, we identify lowest occupied Kohn-Sham levels and elemental electronegativities of the constituent atomic species as the most crucial and relevant predictors. The developed models are validated and tested using the best practices of data science and further analyzed to rationalize their prediction performance. In the recent past, high throughput explorations of enormous chemical spaces have significantly aided the rational 1–4 materials design and discovery process . Massive open-access databases of computed/predicted materials prop- 5–7 erties (including electronic structure, thermodynamic and structural properties) are now available . Materials scientists are currently looking at efficient ways to extract knowledge and mine trends out of materials big-data . As a result, well-established statistical techniques of machine learning (ML) are gradually making inroads into materials science . These methods of data-science and information theory have already met phenomenal success 10 11 12 13 14 in the fields of cheminformatics , game theory , pattern recognition , artificial intelligence , event forecasting etc. and are now being customized for materials informatics to help identify next generation materials break- 15,16 throughs and process optimizations . Given past knowledge—in terms of high quality data on a given property of interest for a limited set of mate- rial candidates within a well defined chemical space—informatics based statistical learning approaches lead to efficient pathways to make high-fidelity predictions on new compounds within the target chemical space. Some 17,18 recent examples of materials’ property predictions using informatics include predictions of molecular and 19–22 23 24,25 26–28 periodic systems’ properties , transition states , potentials , structure classifications , dielectric proper- 2 29 30,31 ties , self-consistent solutions for quantum mechanics and predictions of bandgaps . In this contribution, we aim to build a validated statistical learning model for a specific class of complex oxides, namely the double perovskites. The double perovskite structure, shown in Fig. 1a, is represented by the chemical formula AA’BB’O ; where A and A’ cations are generally of larger radii and have a 12-fold coordination, while the relatively smaller B and B’ metal ions occupy six-fold coordinated positions in oxygen octahedra. The A-site ions typically have + 1, + 2 or + 3 nominal charge states, while the charge state of the B-site cations is gov- erned by the overall charge neutrality of the system. We consider the double perovskite oxides as a material-class of interest owing to both the chemical flexibility made available by the perovskite framework in accommodating a broad spectrum of atomic substitutions, and the vastness of compositional and configurational space spanned by the double perovskites . The ability to rapidly and accurately predict bandgaps of double perovskites is of much interest for a range of applications that require materials with pre-specified constraints on bandgaps, for instance, scintillation , 34 35 photovoltaics and catalysis , to name a few. Local and semi-local functionals used within density functional theory—the current workhorse for electronic structure computations—have a well known deficiency of severely 36 37 underestimating the bandgaps. More advanced methods such as the GW approach and hybrid functionals are enormously computation-time intensive and are thus impractical in high throughput studies aimed at screening Materials Science and Technology Division, Los Alamos National Laboratory, Los Alamos 87545, NM, USA. Department of Materials Science & Engineering and Institute of Materials Science, University of Connecticut, Storrs, 06269 CT USA. Theoretical Division, Los Alamos National Laboratory, Los Alamos 87545, NM, USA. Correspondence and requests for materials should be addressed to G.P. (email: [email protected]) Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 1 www.nature.com/scientificreports/ Figure 1. (a) Double perovskite crystal structure with rocksalt ordering of both A- and B-site cations. Oxygen octahedral coordination around the B-site cations is explicitly shown. (b) Chemical space of the double perovskite oxides explored in the present study. Cations appearing at the A-site and/or the B-site are highlighted. (c) Matrix plot of the double perovskites bandgaps in the database used in the present study. The abscissa and ordinate represent the A–B cation pairs of the constituent single perovskites. The matrix diagonal, shown with gray circles, represents the 53 single perovskites which were not included in the database. White circles represent the 72 compositions with either zero or negligible bandgaps, which were also not included in the database. (d) Histogram of GLLB-SC bandgaps of 1306 double perovskites used in the development of the ML model presented here. promising candidates with targeted bandgaps. This is one of the most important reasons for seeking to develop a statistical (or machine) learning model where one can use easily accessible attributes (also referred to as features) of a material to directly predict its bandgap, in an efficient yet accurate manner. Our primary goal is to develop a validated and predictive model that establishes a mathematical relationship (or mapping) between the bandgap of material i residing in the predefined chemical space, and an Ω -dimensional (Ω − D) feature vector f of the mate- rial i. Here, the Ω − D vector f (also referred to as a descriptor) is composed of Ω different features and uniquely describes the material i. It is also desirable to have a model which is both simple (i.e., with sufficiently small Ω ) and reasonably accurate. Here, we report a feature-selection (i.e., how to find an optimal Ω − D feature vector) and learning framework (i.e., how to establish the mapping between the bandgaps and feature vectors) for efficient and accurate predictions of electronic bandgaps of double perovskites. Results We start by describing the details of our double perovskite bandgap dataset that was used to train, validate and test the prediction performance of the ML models developed here. The dataset used here came from the Computational Materials Repository (CMR) . The double perovskite structures reported in this dataset were obtained by combining 53 stable cubic perovskite oxides which were found to have a finite bandgap in a previous 38,39 screening based on single perovskites . These 53 parent single perovskites contained fourteen different A-site cations (viz. Ag, Ba, Ca, Cs, K, La, Li, Mg, Na, Pb, Rb, Sr, Tl and Y) and ten B-site cations (viz. Al, Hf, Nb, Sb, Sc, Si, Ta, Ti, V, Zr). Four cations (Ga, Ge, In and Sn) were found to appear on either A- or B-sites. The chemical space spanned by these compounds is captured in Fig. 1b. A total of C = 1378 unique double perovskites are possible by combining the 53 stable single cubic per- ovskite oxides, when taken pairwise. However, out of these systems, 72 double perovskites are metallic (or have a very small bandgap <01 . eV) and are not included in the database. These systems are depicted in Fig. 1c as off-diagonal circles. The CMR dataset reports the electronic bandgaps of the remaining1306 unique double perovskites. Depending on the nature of the cations, various types of cation ordering can consequently arise in the double perovskites . For a doubly substituted AA’BB’O -type perovskite, there are three common ways in which cations at each of the two sublattices can order, leading to a total of nine different ordered arrangements. Specifically, A and A’ (and B and B’) cations can order in layered, columnar, or rocksalt arrangements. The most commonly Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 2 www.nature.com/scientificreports/ Figure 2. Overall workflow for the statistical learning model. Schematic presents the details of the CMR double perovskite bandgap database and outlines the workflow adopted for the primary and compound feature selection, leading to a cross-validated and tested nonlinear regression model for bandgap predictions. observed type of ordering for the B-site sublattice is the one in which the cations alternate in all three dimensions, mimicking a rocksalt-type ordered sublattice to effectively accommodate any local strain arising due to size mis - match of the two cations. Less frequently, the B-site cations may form a layered order, where they alternate only in one direction and form continuous layers in the other two normal cartesian axes. Rarely, however, a columnar order may take place, where the two different cations alternate in two orthogonal directions, but form a continu - ous column along the third direction. The CMR database reports the bandgaps of all the double perovskites with the rocksalt ordering of cations at both the A- and the B-sites . e r Th eported bandgaps ( cf. Fig. 1c,d) are computed using density functional theory (DFT) as implemented 42 43 in the GPAW code with the Gritsenko, van Leeuwen, van Lenthe and Baerends potential (GLLB) , further opti- mized for solids (-SC) by Kuisma et al. . The GLLB functional has an inbuilt prescription for the evaluation of the derivative discontinuity , which is added back to the Kohn-Sham bandgap to correct for the bandgap under- estimation within conventional DFT. In fact, the GLLB-SC bandgaps for several single metal oxides have been found in excellent agreement with the corresponding experimental values (cf. Supplementary Information) . Furthermore, the GLLB-SC functional was recently tested against the more advanced and demanding eigenvalue-self-consistent GW approach and has been shown to give good agreement for the bandgap of 20 ran- domly chosen systems forming an unconventional set of ternary and quaternary compounds taken from from the Materials Project database . Finally, we would also like to note that despite its significantly low computational cost compared to the GW approach, the GLLB-SC functional is about twice as expensive as compared to a con- ventional DFT calculation employing a local or semi-local functional. Any ML method, targeted towards learning a prespecified material property, relies on two main ingredients: the learning algorithm itself and a numerical representation (in form of a descriptor) of the materials in the learning (or training) dataset. Identification of an appropriate and most suitable fingerprint for a given prediction task is one of the central challenges, at present being actively pursued by the community. The specific choice of this numerical representation is entirely application dependent and a number of proposals in terms of high-level 26,47,48 features (e.g., d-band center, elemental electronegativities and ionization potentials, valence orbital radii) , 49 19 topological features , atomic radial distribution functions , compositional, configurational and motif based 2,18,25 fingerprints have been made. Regardless of the specific choice pursued, the representations are expected to satisfy certain basic requirements such as invariance with respect to transformations of the material such as translation, rotation, and permutation of like elements. Additionally, it is also desirable that the fingerprinting be chemically intuitive and physically meaningful. The overall workflow adopted in the present study, combining an effective feature search strategy with a state-of-the-art statistical learning method, is outlined in Fig. 2. Our systematic approach starts with identifica - tion of seven atomic (or elemental) features for each of the cation species forming the double perovskite structure. These elemental features ( viz. Pauling’s electronegativity (χ), ionization potential (I), highest occupied atomic level (h), lowest unoccupied atomic level (l), and s-, p- and d- valence orbital radii r , r and r of isolated neutral s p d atoms) are easily accessible and physically intuitive attributes of the constituent atoms at the A- and B-sites. While χ, I, h and l naturally form the energy scales relevant towards prediction of the bandgap, the valence orbital radii 50,51 were included based on their excellent performance exhibited in classification of AB binary solids . Taking these seven elemental features for each of the four atoms, occupying either A- or B-site, forms our starting set of twenty eight elemental features. Further details on the feature set are provided in the Methods section. It is also worthwhile to note at this point that the double perovskite structure under investigation is invariant with respect to swapping of the two A-site cations as well as the two B-site cations, i.e., AA’BB’O , A’ABB’O and 6 6 AA’B’BO are all identical systems. To incorporate this structural symmetry into the model, we symmetrize the above-mentioned 28 elemental features such that they reflect the underlying symmetry of the A- and B-site sub - lattices of the double perovskite crystal structure. This is achieved by taking the absolute sum and ff + AA′ absolute difference for each pair of elemental features f and f , representing the two A-site cations. ff − A A′ AA′ Features for the B-site cations were also transformed in a similar fashion. For convenience of notation, ff ± AA′ Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 3 www.nature.com/scientificreports/ Figure 3. Pearson correlation map for features. A graphical representation of the Pearson correlation matrix 1,16 for the downselected primary (labelled as Pi with i ∈ in the upper-right part) and compound (labelled as 1,16 Ci with i ∈ in the lower-left part) features is presented. Blue and red colors indicate positive and negative correlations, respectively; the lighter the tone used, the less significant the corresponding correlation. The filled fraction of the circle in each of the pie charts corresponds to the absolute value of the associated Pearson correlation coeficient. ± ± and ff ± are henceforth represented by f and f , respectively. Building such symmetry at the feature BB′ A B level ensures that the deemed ML model would predict identical bandgaps for symmetry unique systems, irre- spective of any specific labeling of the two A- and two B-site atoms. This set of 28 symmetrized features thus achieved is hereaer r ft eferred to as primary features. At this point, we adopt a two-fold route for feature selection. While the primary features can directly be used in a statistical learning model, we also consider a large set of conjunctive—or compound—features built in a con- trolled manner to allow for non-linearity at the feature level. The compound feature set is built in the following 1/2 2 3 x way: 6 prototypical functions, namely, x, x , x , x , ln(1 + x), and e , with x being one of the 28 primary features, were considered. This immediately generates 168 features. Simply multiplying these features of single functions taken either two or three at a time leads to additional 16,464 and 1,229,312 features, respectively. This approach thus provides us with 1,245,944 compound features, each of which is a function involving up to 3 primary fea- tures. Finally, a least absolute shrinkage and selection operator (LASSO)-based model selection is employed to downselect a set of 40 compound features, which are deemed most relevant towards prediction of the bandgaps. We note here that this strategy of creating a large number of initial compound features and down-selecting to the most relevant ones using LASSO has recently been successfully employed to identify new crystal structure classifiers . A LASSO-based formalism has also been employed to identify lower-dimensional representations of alloy cluster expansions . Next, the primary features and downselected compound features are subjected to a Pearson correlation filter (cf. Fig. 2) to remove features that exhibit a high correlation with the other features in each set. e Th cut-off of the Pearson correlation filter was adjusted such that only 16 features in each set survive. Tests showed that selecting more than 16 features does not lead to any improvements in the out-of-sample prediction accuracy of ML mod- els. A Pearson correlation map showing the correlation for each pair of the primary or compound features is presented in Fig. 3. The above sets of 16-primary and 16-compound features were subsequently used separately to construct all possible Ω -dimensional (or Ω − D) descriptors (i.e., taking Ω features at a time), where Ω was varied from 1 to 16. This leads to 2 − 1(= 65535) total possible descriptors to be tested for the primary and the compound feature sets. Since testing and evaluating prediction performance of such a large number of descriptors using non-linear statistical learning models (such as kernel ridge regression or KRR) is a highly computation-time intensive task, we resort here to a cross-validated linear least square fit (LLSF) model instead. A training set consisting of 90% of the whole dataset was used to fit a linear model, and the rest 10% was used as a test set to evaluate root mean squared (rms) error and coefficient of determination ( R ) of the fit. To take into account variability of the models, average test set rms error and average test set R over 100 different bootstrap runs were used to rank the linear models. The LLSF performance of the best Ω − D descriptors for a given Ω ∈ [1, 16] is presented in Fig. 4a. We find that for any given Ω , the descriptors with the compound features perform much better than those formed using the primary features. Certainly, this boost in performance can be attributed to the additional flexibility imparted Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 4 www.nature.com/scientificreports/ Figure 4. Prediction performance of the developed linear and non-linear learning models. (a) Computed test set rms errors and coefficient of determinations ( R ) for the best Ω − D models, with Ω ranging from 1 through 16, in linear least square fit (LLSF) models which are built on either primary or compound features. The prediction performance reported here is computed as an average over 100 different runs, each with randomly selected 90% training and 10% test sets. Representative parity plots comparing the DFT-computed bandgaps against the KRR predicted bandgaps, for (b) the 4-D primary descriptor, (c) the 16-D primary descriptor and (d) the 16-D compound descriptor. Histograms of the test set average rms error, computed over 400 different runs, are also presented as insets in the last three panels. by the non-linear functions in the compound features. Furthermore, a compound feature can effectively have a combination of up to three functions of primary features. We also note that going beyond a 10-D descriptor does not improve the prediction performance in either case (cf. Fig. 4a). For instance, the average rms error for the 16-D descriptor formed with compound features is 0.786 eV, while that for the 10-D descriptor is 0.792 eV. The average rms errors for the corresponding descriptors with primary features are 0.971 eV and 0.973 eV, respectively. Performance of the best primary and compound Ω − D descriptors (with Ω ∈[1, 16]) identified above was 53,54 then reassessed in a Kernel ridge regression (KRR) model—a state-of-the-art ML method capable of handling complex non-linear relationships—which has recently been shown to be promising for prediction of a diverse set 2,17,55–57 of materials properties . Based on the principle of similarity, the KRR method first uses a distance measure 2 th th such as the Euclidean norm in the descriptor space (i.e., , for i and j compounds in the training set) ‖ff − ‖ ij to quantify (dis)similarity between materials; the property to be predicted is then computed as a linear combina- tion of the kernel (e.g., a gaussian kernel used in the present case) distance k (, ij)= exp2 −− ‖‖ ff / σ () ij functions of materials of interest and the training set materials. Therefore, constructing descriptors in which materials have a small distance when their property of interest is similar is of particular importance for the learn- ing process. Further details on the KRR learning model are provided in Methods. Results obtained using the cross-validated KRR models are presented in Table 1 and the identities of the best Ω − D descriptors are provided in the Supplementary Information. For each descriptor, the average rms error and average R on training and test sets are reported. The average was taken over 100 different KRR runs, in each of which a 90% training set and a 10% test set were randomly selected. Not surprisingly, it is seen that both the learning and prediction accuracies grow with the descriptor dimensionality (and complexity). Interestingly, Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 5 www.nature.com/scientificreports/ Performance of KRR ML models with primary Performance of KRR ML models with descriptors compound descriptors Training set Training set Descriptor (90%) Test set (10%) Descriptor (90%) Test set (10%) rms rms rms rms error error error error 2 2 2 2 Dimension (eV) R (eV) R Dimension (eV) R (eV) R 1-D 0.959 0.634 1.056 0.546 1-D 0.954 0.638 1.056 0.545 2-D 1.059 0.554 1.114 0.493 2-D 0.888 0.686 0.879 0.685 3-D 0.689 0.811 0.867 0.692 3-D 0.774 0.762 0.777 0.753 4-D 0.306 0.963 0.501 0.897 4-D 0.716 0.796 0.742 0.775 5-D 0.270 0.971 0.529 0.885 5-D 0.566 0.872 0.637 0.834 6-D 0.302 0.968 0.521 0.889 6-D 0.502 0.900 0.594 0.855 7-D 0.356 0.949 0.510 0.892 7-D 0.426 0.928 0.554 0.875 8-D 0.281 0.969 0.429 0.925 8-D 0.437 0.924 0.563 0.870 9-D 0.219 0.981 0.406 0.932 9-D 0.399 0.937 0.550 0.876 10-D 0.139 0.992 0.397 0.935 10-D 0.300 0.964 0.484 0.904 11-D 0.178 0.987 0.413 0.930 11-D 0.249 0.975 0.455 0.915 12-D 0.074 0.998 0.393 0.936 12-D 0.222 0.980 0.451 0.917 13-D 0.137 0.993 0.365 0.944 13-D 0.187 0.986 0.418 0.928 14-D 0.110 0.995 0.397 0.935 14-D 0.169 0.989 0.413 0.930 15-D 0.087 0.997 0.377 0.941 15-D 0.144 0.992 0.376 0.942 16-D 0.080 0.997 0.371 0.939 16-D 0.132 0.993 0.360 0.947 Table 1. Prediction performance of the KRR models. Computed training set and test set rms errors and coefficient of determinations (R ) for the best Ω − D models, with Ω ∈ [1, 16], in 5-fold cross-validated KRR models built on top primary and compound descriptors (identified in the LLSF models). The reported rms errors and R are averaged values over 100 different runs, each with randomly selected 90% training and 10% 2 2 2 test sets. Here R =− 1 (− bb )/ (− bb ) with b , b and b being the GLLB-SC computed ∑ ∑ i i,pred avg ii,pred i avg bandgaps, model-predicted bandgaps and average bandgap value in the training or test set. unlike the LLSF model, in the KRR model the performance of a primary descriptor is found to be comparable to that of the corresponding compound descriptor. This is owing to the inclusion of the nonlinearity in the learning model itself, which boosts the performance of the primary descriptors. In light of this observation, going forward with the KRR model, we choose the simpler models with the primary descriptors over the compound descriptors. It is interesting to note that going from the 3-D to the 4-D primary descriptor (cf. Table 1) leads to a significant improvement in the model prediction performance. For instance, the average R on the test set increases from 0.69 to 0.90 and average rms error decreases from ~0.87 eV to ~0.50 eV. Going beyond the 4-D descriptor, how- ever, only results in marginal improvements. For instance, with the 16-D descriptor (containing all of the primary features) the obtained average test set R is ~0.94, only slightly better than that of the 4-D descriptor. Figure 4b–d compare the KRR prediction performance of the 4-D descriptor with the 16-D primary and the 16-D compound descriptor in separate parity plots, using a representative training/test set split. It can be seen that while the train- ing set performance is significantly better in the KRR models with the higher dimensional descriptors, the test set performance of those models can be considered comparable (or only slightly better) to that of the 4-D descriptor. Discussion We now examine the individual primary features that combine to give the 4-D primary descriptor. These features + − + − are: l , l , , , i.e., the absolute sum and difference of elemental lowest occupied levels of the two A-site atoms χ χ A A B B and the electronegativities of the two B site atoms. We also note that the two conjugate pairs of these elemental features appear together and none of the primary features with valence orbital radii appears. Furthermore, the descriptor is well balanced with respect to the participation from the features specific to the A-site atoms (i.e., l , − + − l ) and to the B-site atoms (i.e., χ , χ ). In addition to being chemically intuitive, elegant and symmetric, the B B identified descriptor is also simple and easily accessible. It is always desirable to have a ML prediction model built on simpler (i.e., low dimensional) and intuitive descriptors, since with high dimensional complex descriptors there is always a danger of overt fi ting leading to poor model generalizability. Therefore, by preferring the 4-D primary descriptor over the 16-D descriptor, we are trading some model accuracy for model simplicity and better model generalizability. To further test the model’s predictability, we used the cross-validated KRR learning model, trained on a ran- domly selected 90% double perovskite dataset, to predict bandgaps of the original 53 parent single perovskites. We note that for the single perovskites, owing to the constraints A = A’ and B = B’, only two of the four features − − survive (i.e., for all the single perovskites we have l = 0, χ = 0). Figure 5 compares the bandgaps predicted by the model with those computed using DFT with the GLLB-SC functional. Given that the model was never trained on single perovskites and that only two of the four primary features effectively survive for a single perovskite, such a prediction performance is rather remarkable. Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 6 www.nature.com/scientificreports/ Figure 5. Predictions on single perovskites. Prediction performance of the KRR model in predicting the bandgaps of the parent 53 single perovskites, which form all the double perovskites in the database. The model was trained on a randomly selected 90% training set from the double perovskite bandgap database using the 4-D primary descriptor. To gain deeper insight into the model’s remarkable prediction performance, we next construct 2-D contour plots in which dependence of any two of the four features has been marginalized (by considering an averaged value along those particular dimensions, as explained below). We start with a fine 4-D grid in the feature space constituted by the four primary features identified above, while still confining ourselves within the boundaries of the original feature space used to train the KRR model. Each point on this grid then, in principle, represents a descriptor. Next, we use the trained KRR model to make predictions using each of these descriptor points as a model input. For the sake of better representation, we convert the predictions in this 4-D feature space into a 2-D plot by averaging out any given two of the four primary features. This approach allows us to explicitly visualize the dependence of the bandgap along any two pre-specified features, while the dependence of other two features is considered only in an averaged manner. We can now represent this data in a 2-D contour plot. Three out of a total of six such possible plots are shown in Fig. 6, where green and purple regions represent the high- and low-bandgap regions, respectively. The data-points in the double perovskites dataset are also plotted on top for validation, color (or size) coded according to their GLLB-SC bandgaps. Since the dependence of two out of the four features has already been integrated out, one does not expect a quantitative agreement between the contour and scatter plots. However, it can be seen from the figure that the two are in quite good agreement. The green “mountains” on the contour plot are largely occupied by red (large) circles while the purple “valleys” are mostly populated with blue (small) circles. Such feature-property maps provide a pathway towards drawing decision rules (for a targeted functionality) from statistical learning models. Furthermore, while the original model can be used to make quantitative predictions, such simple feature-property maps can be employed as a first-line of screening to make qualitative predictions or devise simple screening criteria for a given property (in our case the bandgap). Finally, we comment about the limitations and domain of applicability of the ML model. The presented model is applicable within the considered chemical space (i.e., aforementioned choices of A- and B-site cations) and to non-magnetic AA’BB’O type perovskites, which can be separated into two charge neutral ABO and A’B’O single 6 3 3 perovskites. Test performance on double perovskite compounds which cannot be decomposed in such a manner was found to be poor, which is not surprising since the learning model was never trained on such compounds. Extending the ML model to such compounds and accounting for other possible A- and B-site cation orderings remains work to be undertaken in future studies. It will also be interesting to check the general applicability of the identified descriptor by employing it to predict the bandgaps of other materials classes, quite distinct from perovskites and related chemistries. In summary, we have presented a robust ML model along with a simple elemental descriptor set for efficient predictions of electronic bandgaps of double perovskites. The proposed optimal descriptor set was identified via searching a large part of feature space that involved more than ~1.2 million descriptors formed by combining simple elemental features such as electronegativities, ionization potentials, electronic energy levels and valence orbital radii of the constituent atomic species. The KRR-based statistical learning model developed here was trained and tested on a database consisting of accurate bandgaps of ~1300 double perovskites computed using the GLLB-SC functional within the framework of density functional theory. One of the most important chem- ical insights that came out of the adopted learning framework is that the bandgap is primarily controlled (and therefore can efficiently be learned) by the lowest occupied energy levels of the A-site elements and electronega - tivities of the B-site elements. The average test set rms error of the cross validated model with only four primary features (i.e., the 4-D primary descriptor) is found to be 0.5 eV, which is further reduced to ~0.37 eV (0.36 eV) Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 7 www.nature.com/scientificreports/ Figure 6. Feature-property maps. (a–c) Two dimensional contour maps showing regions of relatively larger and smaller bandgaps in the feature space. In each of the panels, only two of the four features are explicitly presented, while the dependency of the other two features has been integrated out (see text for details). The data-points in the double perovskites dataset are also plotted on top of the contour plots. The marker size and color of the scattered points represent their bandgaps. with the primary (compound) 16-D descriptor. Out-of-sample prediction performance of the trained model is further demonstrated by its ability to predict bandgaps of several single perovskites. Finally we have shown that the prediction performance of the model can be visually rationalized by constructing several two-dimensional feature-property contour maps. We believe that the ML approach presented here is general and can be applied to any material class in a restricted chemical space with a given crystal structure to make efficient predictions of bandgaps. Such a prediction strategy can be practically useful in an initial screening to identify promising candi- dates in a high throughput manner. Methods Details of feature set. For feature set accumulation, we start from 7 atomic features for each metal atom A in the double perovskite structure. These primary atomic features are Pauling’s electronegativity ( χ), ionization potential (I), highest occupied atomic Kohn-Sham level (h), lowest unoccupied atomic Kohn-Sham level (l), and s-, p- and d- Zunger’s valence orbital radii r , r and r of isolated neutral atoms . The ionization potential s p d 6,59 and Pauling’s electronegativity data were taken from the literature and the highest-occupied lowest-unoccu- pied Kohn-Sham levels of the isolated atomic species were computed using the GGA-PBE exchange-correlation functional . Machine learning model. Within the present similarity-based KRR learning model, the bandgap of a sys- tem in the test set is given by a sum of weighted Gaussians over the entire training set. As a part of the model N i i 2 N 2 training process, the learning is performed by minimizing the expression , (− EE )+ λω ∑ ∑ i=1 gK ,, RR gDFT i=1 i i i with being the KRR estimated bandgap value, the DFT value, and λ a regularization parameter. The E E gK , RR gD , FT −1 explicit solution to this minimization problem is ω =(kI +) λ E , where I is the identity matrix, and gD , FT is the kernel matrix elements of all compounds in the training set. The parameters λ, k =( exp −− ff ) ij 2 ij 2σ σ are determined in an inner loop of fivefold cross validation using a logarithmically scaled fine grid. We note that KRR training and hyperparameter determination were performed only using the training data and the test set samples were never seen by the KRR model during the training procedure. References 1. Curtarolo, S. et al. The high-throughput highway to computational materials design. Nat. Mater. 12, 191–201 (2013). 2. Pilania, G., Wang, C., Jiang, X., Rajasekaran, S. & Ramprasad, R. Accelerating materials property predictions using machine learning. Sci. rep. 3, 2810 (2013). 3. Sharma, V. et al. Rational design of all organic polymer dielectrics. Nat. comm. 5, 4845 (2014). 4. Ceder, G., Hauthier, G., Jain, A. & Ong, S. P. Recharging lithium battery research with first-principles methods. Mater. Res. Soc. Bull. 36, 185–191 (2011). 5. Curtarolo, S. et al. AFLOWLIB.ORG: AFLOWLIB. ORG: A distributed materials properties repository from high-throughput ab initio calculations. Comput. Mater. Sci. 58, 227 (2012). 6. Materials Project - A Materials Genome Approach, http://materialsproject.org/ (accessed: 15th October 2015). 7. Computational Materials Repository https://wiki.fysik.dtu.dk/cmr/ (Documentation) and https://cmr.fysik.dtu.dk/ (accessed: 15th October 2015). 8. Service, R. F. Materials scientists look to a data-intensive future. Science 335, 1434–1435 (2012). 9. Flach, P. Machine Learning: The Art and Science of Algorithms that Make Sense of Data (Cambridge University Press, Cambridge, 2012). 10. Burbidge, R., Trotter, M., Buxton, B. & Holden, S. Drug design by machine learning: support vector machines for pharmaceutical data analysis. Computers & chemistry 26, 5–14 (2001). 11. Jones, N. Quiz-playing computer system could revolutionize research. Nature News (2011), Available at: http://dx.doi.org/10.1038/ news.2011.95. (Accessed: 23rd November 2015). 12. MacLeod, N., Benfield, M. & Culverhouse, P. Time to automate identification. Nature 467, 154–155 (2010). Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 8 www.nature.com/scientificreports/ 13. Abu-Mostafa, Y. S. Machines that Think for Themselves. Sci Am 307, 78–81 (2012). 14. Silver, N. e S Th ignal and the Noise: Why So Many Predictions Fail but Some Don’t (Penguin Press, New York, 2012). 15. Mueller, T., Kusne, A. G. & Ramprasad, R. Machine learning in materials science: Recent progress and emerging applications. Rev. Comput. Chem. (Accepted for publication). 16. Rajan, K. in Informatics for Materials Science and Engineering: Data-driven Discovery for Accelerated Experimentation and Application (ed. Rajan, K.), Ch. 1, 1–16 (Butterworth-Heinemann, Oxford, 2013). 17. Rupp, M., Tkatchenko, A., Muller, K.-R. & von Lilienfeld, O. A. Fast and accurate modeling of molecular atomization energies with machine learning. Phys. Rev. Lett. 108, 058301 (2012). 18. Huan, T. D., Mannodi-Kanakkithodi, A. & Ramprasad, R. Accelerated materials property predictions and design using motif-based fingerprints, Phys. Rev. B 92, 014106 (2015). 19. Schütt, K. T. et al. How to represent crystal structures for machine learning: Towards fast prediction of electronic properties. Phys. Rev. B 89, 205118 (2014). 20. Meredig, B. et al. Combinatorial screening for new materials in unconstrained composition space with machine learning. Phys. Rev. B 89 094104 (2014). 21. Faber, F., Lindmaa, A., von Lilienfeld, O. A. & Armiento, R. Crystal Structure Representations for Machine Learning Models of Formation Energies. Int. J. Quantum. Chem. 115, 1094–1101(2015). 22. Faber, F., Lindmaa, A., von Lilienfeld, O. A. & Armiento, R. Machine Learning Energies of 2 M Elpasolite (ABC2D6) Crystals. http:// arxiv.org/abs/1508.05315 (2015). 23. Pozun, Z. et al. Optimizing transition states via kernel-based machine learning. Chem. Phys. 136, 174101 (2012). 24. Behler, J. Atom-centered symmetry functions for constructing high-dimensional neural network potentials. J. Chem. Phys. 134, 074106 (2011). 25. Botu, V. & Ramprasad, R. Adaptive machine learning framework to accelerate ab initio molecular dynamics, Int. J. Quantum Chem. 115, 1074–1083 (2015). 26. Pilania, G., Gubernatis, J. E. & Lookman, T. Structure classification and melting temperature prediction in octet AB solids via machine learning. Phys. Rev. B 91, 214302 (2015). 27. Pilania, G., Gubernatis, J. E. & Lookman, T. Classification of octet AB-type binary compounds using dynamical charges: A materials informatics perspective. accepted for publication in Sci. Rep. (2015). 28. Pilania, G., Balachandran, P. V., Gubernatis, J. E. & Lookman, T. Predicting the formability of ABO perovskite solids: A machine learning study. Acta Cryst. B 71, 507–513 (2015). 29. Snyder, J. C., Rupp, M., Hansen, K., Müller, K. R. & Burke, K. Finding density functionals with machine learning. Phys. Rev. Lett. 108, 253002 (2012). 30. Lee, J., Seko, A., Shitara, K. & Tanaka, I. Prediction model of band-gap for AX binary compounds by combination of density functional theory calculations and machine learning techniques. arXiv preprint arXiv:1509.00973 (2015). 31. P., Dey et al. Informatics-aided bandgap engineering for solar materials. Com. Mat. Sci. 83, 185–195 (2014). 32. Mitchell, R. H. Perovskites: Modern and Ancient (Almaz Press, Ontario, Canada, 2002). 33. Setyawan, W., Gaume, R. M., Lam, S., Feigelson, R. S. & Curtarolo, S. High-throughput combinatorial database of electronic band structures for inorganic scintillator materials. ACS Comb. Sci. 13, 382–390 (2011). 34. Olivares-Amaya, R. et al. Accelerated computational discovery of high-performance materials for organic photovoltaics by means of cheminformatics. Energy Environ. Sci. 4, 4849 (2011). 35. Chemical Bonding at Surfaces and Interfaces (Eds Nilsson, A., Pettersson, L. G. M. & Nørskov, J. K.) (Elsevier, Amsterdam, The Netherlands, 2008). 36. Hedin, L. New method for calculating the one-particle Green’s function with application to the electron-gas problem. Phys. Rev. 139 A796 (1965). 37. Heyd, J., Scuseria, G. E. & Ernzerhof, M. Hybrid functionals based on a screened Coulomb potential. J. Chem. Phys. 124, 219906 (2006). 38. Castelli, I. E. et al. Computational screening of perovskite metal oxides for optimal solar light capture. Energy Environ. Sci. 5, 5814 (2012). 39. Castelli, I. E., Thygesen, K. S. & Jacobsen, K. W. Bandgap engineering of double perovskites for one-and two-photon water splitting. MRS Proceedings 1523, mrsf12-1523-qq07-06 (2013), doi: 10.1557/opl.2013.450. 40. Vasala, S. & Karppinen, M. A B’B”O perovskites: A review. Prog. Solid State Chem. 43, 1–36 (2015). 2 6 41. Martin, R. Electronic Structure: Basic Theory and Practical Methods (Cambridge University Press, New York, 2004). 42. Mortensen, J. J., Hansen, L. B. & Jacobsen, K. W. Real-space grid implementation of the projector augmented wave method. Phys. Rev. B 71, 35109 (2005). 43. Gritsenko, O., van Leeuwen, R., van Lenthe, E. & Baerends, E. J. Self-consistent approximation to the Kohn-Sham exchange potential. Phys. Rev. A 51, 1944 (1995). 44. Kuisma, M., Ojanen, J., Enkovaara, J. & Rantala, T. T. Kohn-Sham potential with discontinuity for band gap materials. Phys. Rev. B 82, 115106 (2010). 45. Talman, J. D. & Shadwick, W. F. Optimized effective atomic central potential. Phys. Rev. A 14, 36 (1976). 46. Castelli, I. E. et al. New light-harvesting materials using accurate and efficient bandgap calculations. Adv. Energy Mater. 5, 1400915 (2015). 47. Andriotis, A. N. et al. Informatics guided discovery of surface structure-chemistry relationships in catalytic nanoparticles. J. Chem. Phys. 140, 094705 (2014). 48. Dam, H. C., Pham, T. L., Ho, T. B., Nguyen, A. T. & Nguyen, V. C. Data mining for materials design: A computational study of single molecule magnet. J. Chem. Phys. 140, 044101 (2014). 49. Brown, R. D. & Martin, Y. C. The information content of 2D and 3D structural descriptors relevant to ligand-receptor binding. J. Chem. Inf. Comput. Sci. 37, 1 (1997). 50. Zunger, A. Systematization of the stable crystal structure of all AB-type binary compounds. Phys. Rev. B 22, 5839 (1980). 51. Ghiringhelli, L. M., Vybiral, J., Levchenko, S. V., Draxl, C. & Scheffler, M. Big data of materials science: Critical role of the descriptor. Phys. Rev. Lett. 114, 105503 (2015). 52. Nelson, L. J., Hart, G. L., Zhou, F. & Ozoliņš, V. Compressive sensing as a paradigm for building physics models. Phys. Rev. B 87, 035125 (2013). 53. Hastie, T., Tibshirani, R. & Friedman, J. e E Th lements of Statistical Learning: Data Mining, Inference, and Prediction (Springer, New York, 2009). 54. Müller, K.-R., Mika, S., Ratsch, G., Tsuda, K. & Scholkopf, B. An introduction to kernel-based learning algorithms. IEEE Trans Neural Netw 12, 181–201 (2001). 55. Bereau, T., Andrienko, D. & von Lilienfeld, O. A. Transferable atomic multipole machine learning models for small organic molecules. J. Chem. Theory Comput. 11, 3225–3233 (2015). 56. Hansen, K. et al. Assessment and validation of machine learning methods for predicting molecular atomization energies. J. Chem. e Th ory Comput. 9, 3404 (2013). 57. Lopez-Bezanilla, A. & von Lilienfeld, O. A. Modeling electronic quantum transport with machine learning. Phys. Rev. B 89, 235411 (2014). Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 9 www.nature.com/scientificreports/ 58. Perdew, J. P., Burke, K. & Ernzerhof, M. Generalized gradient approximation made simple. Phys. Rev. Lett. 77, 3865 (1996). 59. Lide, D. R. Handbook of Chemistry and Physics (CRC Press, Boston, 2004). Acknowledgements G.P., J.E.G. and T.L. acknowledge support from the LANL LDRD program and B.P.U. acknowledges support by the U.S. Department of Energy, Office of Science, Basic Energy Sciences, Materials Sciences and Engineering Division. Los Alamos National Laboratory is operated by Los Alamos National Security, LLC, for the National Nuclear Security Administration of the (U.S.) Department of Energy under contract DE-AC52-06NA25396. G.P. gratefully acknowledges discussions with Ivano E. Castelli. Author Contributions G.P. designed the study with inputs from T.L., J.E.G., B.P.U. and R.R. A.M.K. and G.P. built the data sets and the machine learning models. All authors analyzed and discussed the results, and wrote the manuscript. Additional Information Supplementary information accompanies this paper at http://www.nature.com/srep Competing financial interests: e a Th uthors declare no competing financial interests. How to cite this article: Pilania, G. et al. Machine learning bandgaps of double perovskites. Sci. Rep. 6, 19375; doi: 10.1038/srep19375 (2016). This work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder to reproduce the material. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 10 http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Scientific Reports Springer Journals

Machine learning bandgaps of double perovskites

Loading next page...
 
/lp/springer-journals/machine-learning-bandgaps-of-double-perovskites-YHEJJ0ZIU7

References (70)

Publisher
Springer Journals
Copyright
Copyright © 2016 by The Author(s)
Subject
Science, Humanities and Social Sciences, multidisciplinary; Science, Humanities and Social Sciences, multidisciplinary; Science, multidisciplinary
eISSN
2045-2322
DOI
10.1038/srep19375
Publisher site
See Article on Publisher Site

Abstract

www.nature.com/scientificreports OPEN Machine learning bandgaps of double perovskites 1 2 1 2 3 G. Pilania , A. Mannodi-Kanakkithodi , B. P. Uberuaga , R. Ramprasad , J. E. Gubernatis & T. Lookman received: 29 October 2015 The ability to make rapid and accurate predictions on bandgaps of double perovskites is of much a ccepted: 07 December 2015 practical interest for a range of applications. While quantum mechanical computations for high-fidelity Published: 19 January 2016 bandgaps are enormously computation-time intensive and thus impractical in high throughput studies, informatics-based statistical learning approaches can be a promising alternative. Here we demonstrate a systematic feature-engineering approach and a robust learning framework for efficient and accurate predictions of electronic bandgaps of double perovskites. After evaluating a set of more than 1.2 million features, we identify lowest occupied Kohn-Sham levels and elemental electronegativities of the constituent atomic species as the most crucial and relevant predictors. The developed models are validated and tested using the best practices of data science and further analyzed to rationalize their prediction performance. In the recent past, high throughput explorations of enormous chemical spaces have significantly aided the rational 1–4 materials design and discovery process . Massive open-access databases of computed/predicted materials prop- 5–7 erties (including electronic structure, thermodynamic and structural properties) are now available . Materials scientists are currently looking at efficient ways to extract knowledge and mine trends out of materials big-data . As a result, well-established statistical techniques of machine learning (ML) are gradually making inroads into materials science . These methods of data-science and information theory have already met phenomenal success 10 11 12 13 14 in the fields of cheminformatics , game theory , pattern recognition , artificial intelligence , event forecasting etc. and are now being customized for materials informatics to help identify next generation materials break- 15,16 throughs and process optimizations . Given past knowledge—in terms of high quality data on a given property of interest for a limited set of mate- rial candidates within a well defined chemical space—informatics based statistical learning approaches lead to efficient pathways to make high-fidelity predictions on new compounds within the target chemical space. Some 17,18 recent examples of materials’ property predictions using informatics include predictions of molecular and 19–22 23 24,25 26–28 periodic systems’ properties , transition states , potentials , structure classifications , dielectric proper- 2 29 30,31 ties , self-consistent solutions for quantum mechanics and predictions of bandgaps . In this contribution, we aim to build a validated statistical learning model for a specific class of complex oxides, namely the double perovskites. The double perovskite structure, shown in Fig. 1a, is represented by the chemical formula AA’BB’O ; where A and A’ cations are generally of larger radii and have a 12-fold coordination, while the relatively smaller B and B’ metal ions occupy six-fold coordinated positions in oxygen octahedra. The A-site ions typically have + 1, + 2 or + 3 nominal charge states, while the charge state of the B-site cations is gov- erned by the overall charge neutrality of the system. We consider the double perovskite oxides as a material-class of interest owing to both the chemical flexibility made available by the perovskite framework in accommodating a broad spectrum of atomic substitutions, and the vastness of compositional and configurational space spanned by the double perovskites . The ability to rapidly and accurately predict bandgaps of double perovskites is of much interest for a range of applications that require materials with pre-specified constraints on bandgaps, for instance, scintillation , 34 35 photovoltaics and catalysis , to name a few. Local and semi-local functionals used within density functional theory—the current workhorse for electronic structure computations—have a well known deficiency of severely 36 37 underestimating the bandgaps. More advanced methods such as the GW approach and hybrid functionals are enormously computation-time intensive and are thus impractical in high throughput studies aimed at screening Materials Science and Technology Division, Los Alamos National Laboratory, Los Alamos 87545, NM, USA. Department of Materials Science & Engineering and Institute of Materials Science, University of Connecticut, Storrs, 06269 CT USA. Theoretical Division, Los Alamos National Laboratory, Los Alamos 87545, NM, USA. Correspondence and requests for materials should be addressed to G.P. (email: [email protected]) Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 1 www.nature.com/scientificreports/ Figure 1. (a) Double perovskite crystal structure with rocksalt ordering of both A- and B-site cations. Oxygen octahedral coordination around the B-site cations is explicitly shown. (b) Chemical space of the double perovskite oxides explored in the present study. Cations appearing at the A-site and/or the B-site are highlighted. (c) Matrix plot of the double perovskites bandgaps in the database used in the present study. The abscissa and ordinate represent the A–B cation pairs of the constituent single perovskites. The matrix diagonal, shown with gray circles, represents the 53 single perovskites which were not included in the database. White circles represent the 72 compositions with either zero or negligible bandgaps, which were also not included in the database. (d) Histogram of GLLB-SC bandgaps of 1306 double perovskites used in the development of the ML model presented here. promising candidates with targeted bandgaps. This is one of the most important reasons for seeking to develop a statistical (or machine) learning model where one can use easily accessible attributes (also referred to as features) of a material to directly predict its bandgap, in an efficient yet accurate manner. Our primary goal is to develop a validated and predictive model that establishes a mathematical relationship (or mapping) between the bandgap of material i residing in the predefined chemical space, and an Ω -dimensional (Ω − D) feature vector f of the mate- rial i. Here, the Ω − D vector f (also referred to as a descriptor) is composed of Ω different features and uniquely describes the material i. It is also desirable to have a model which is both simple (i.e., with sufficiently small Ω ) and reasonably accurate. Here, we report a feature-selection (i.e., how to find an optimal Ω − D feature vector) and learning framework (i.e., how to establish the mapping between the bandgaps and feature vectors) for efficient and accurate predictions of electronic bandgaps of double perovskites. Results We start by describing the details of our double perovskite bandgap dataset that was used to train, validate and test the prediction performance of the ML models developed here. The dataset used here came from the Computational Materials Repository (CMR) . The double perovskite structures reported in this dataset were obtained by combining 53 stable cubic perovskite oxides which were found to have a finite bandgap in a previous 38,39 screening based on single perovskites . These 53 parent single perovskites contained fourteen different A-site cations (viz. Ag, Ba, Ca, Cs, K, La, Li, Mg, Na, Pb, Rb, Sr, Tl and Y) and ten B-site cations (viz. Al, Hf, Nb, Sb, Sc, Si, Ta, Ti, V, Zr). Four cations (Ga, Ge, In and Sn) were found to appear on either A- or B-sites. The chemical space spanned by these compounds is captured in Fig. 1b. A total of C = 1378 unique double perovskites are possible by combining the 53 stable single cubic per- ovskite oxides, when taken pairwise. However, out of these systems, 72 double perovskites are metallic (or have a very small bandgap <01 . eV) and are not included in the database. These systems are depicted in Fig. 1c as off-diagonal circles. The CMR dataset reports the electronic bandgaps of the remaining1306 unique double perovskites. Depending on the nature of the cations, various types of cation ordering can consequently arise in the double perovskites . For a doubly substituted AA’BB’O -type perovskite, there are three common ways in which cations at each of the two sublattices can order, leading to a total of nine different ordered arrangements. Specifically, A and A’ (and B and B’) cations can order in layered, columnar, or rocksalt arrangements. The most commonly Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 2 www.nature.com/scientificreports/ Figure 2. Overall workflow for the statistical learning model. Schematic presents the details of the CMR double perovskite bandgap database and outlines the workflow adopted for the primary and compound feature selection, leading to a cross-validated and tested nonlinear regression model for bandgap predictions. observed type of ordering for the B-site sublattice is the one in which the cations alternate in all three dimensions, mimicking a rocksalt-type ordered sublattice to effectively accommodate any local strain arising due to size mis - match of the two cations. Less frequently, the B-site cations may form a layered order, where they alternate only in one direction and form continuous layers in the other two normal cartesian axes. Rarely, however, a columnar order may take place, where the two different cations alternate in two orthogonal directions, but form a continu - ous column along the third direction. The CMR database reports the bandgaps of all the double perovskites with the rocksalt ordering of cations at both the A- and the B-sites . e r Th eported bandgaps ( cf. Fig. 1c,d) are computed using density functional theory (DFT) as implemented 42 43 in the GPAW code with the Gritsenko, van Leeuwen, van Lenthe and Baerends potential (GLLB) , further opti- mized for solids (-SC) by Kuisma et al. . The GLLB functional has an inbuilt prescription for the evaluation of the derivative discontinuity , which is added back to the Kohn-Sham bandgap to correct for the bandgap under- estimation within conventional DFT. In fact, the GLLB-SC bandgaps for several single metal oxides have been found in excellent agreement with the corresponding experimental values (cf. Supplementary Information) . Furthermore, the GLLB-SC functional was recently tested against the more advanced and demanding eigenvalue-self-consistent GW approach and has been shown to give good agreement for the bandgap of 20 ran- domly chosen systems forming an unconventional set of ternary and quaternary compounds taken from from the Materials Project database . Finally, we would also like to note that despite its significantly low computational cost compared to the GW approach, the GLLB-SC functional is about twice as expensive as compared to a con- ventional DFT calculation employing a local or semi-local functional. Any ML method, targeted towards learning a prespecified material property, relies on two main ingredients: the learning algorithm itself and a numerical representation (in form of a descriptor) of the materials in the learning (or training) dataset. Identification of an appropriate and most suitable fingerprint for a given prediction task is one of the central challenges, at present being actively pursued by the community. The specific choice of this numerical representation is entirely application dependent and a number of proposals in terms of high-level 26,47,48 features (e.g., d-band center, elemental electronegativities and ionization potentials, valence orbital radii) , 49 19 topological features , atomic radial distribution functions , compositional, configurational and motif based 2,18,25 fingerprints have been made. Regardless of the specific choice pursued, the representations are expected to satisfy certain basic requirements such as invariance with respect to transformations of the material such as translation, rotation, and permutation of like elements. Additionally, it is also desirable that the fingerprinting be chemically intuitive and physically meaningful. The overall workflow adopted in the present study, combining an effective feature search strategy with a state-of-the-art statistical learning method, is outlined in Fig. 2. Our systematic approach starts with identifica - tion of seven atomic (or elemental) features for each of the cation species forming the double perovskite structure. These elemental features ( viz. Pauling’s electronegativity (χ), ionization potential (I), highest occupied atomic level (h), lowest unoccupied atomic level (l), and s-, p- and d- valence orbital radii r , r and r of isolated neutral s p d atoms) are easily accessible and physically intuitive attributes of the constituent atoms at the A- and B-sites. While χ, I, h and l naturally form the energy scales relevant towards prediction of the bandgap, the valence orbital radii 50,51 were included based on their excellent performance exhibited in classification of AB binary solids . Taking these seven elemental features for each of the four atoms, occupying either A- or B-site, forms our starting set of twenty eight elemental features. Further details on the feature set are provided in the Methods section. It is also worthwhile to note at this point that the double perovskite structure under investigation is invariant with respect to swapping of the two A-site cations as well as the two B-site cations, i.e., AA’BB’O , A’ABB’O and 6 6 AA’B’BO are all identical systems. To incorporate this structural symmetry into the model, we symmetrize the above-mentioned 28 elemental features such that they reflect the underlying symmetry of the A- and B-site sub - lattices of the double perovskite crystal structure. This is achieved by taking the absolute sum and ff + AA′ absolute difference for each pair of elemental features f and f , representing the two A-site cations. ff − A A′ AA′ Features for the B-site cations were also transformed in a similar fashion. For convenience of notation, ff ± AA′ Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 3 www.nature.com/scientificreports/ Figure 3. Pearson correlation map for features. A graphical representation of the Pearson correlation matrix 1,16 for the downselected primary (labelled as Pi with i ∈ in the upper-right part) and compound (labelled as 1,16 Ci with i ∈ in the lower-left part) features is presented. Blue and red colors indicate positive and negative correlations, respectively; the lighter the tone used, the less significant the corresponding correlation. The filled fraction of the circle in each of the pie charts corresponds to the absolute value of the associated Pearson correlation coeficient. ± ± and ff ± are henceforth represented by f and f , respectively. Building such symmetry at the feature BB′ A B level ensures that the deemed ML model would predict identical bandgaps for symmetry unique systems, irre- spective of any specific labeling of the two A- and two B-site atoms. This set of 28 symmetrized features thus achieved is hereaer r ft eferred to as primary features. At this point, we adopt a two-fold route for feature selection. While the primary features can directly be used in a statistical learning model, we also consider a large set of conjunctive—or compound—features built in a con- trolled manner to allow for non-linearity at the feature level. The compound feature set is built in the following 1/2 2 3 x way: 6 prototypical functions, namely, x, x , x , x , ln(1 + x), and e , with x being one of the 28 primary features, were considered. This immediately generates 168 features. Simply multiplying these features of single functions taken either two or three at a time leads to additional 16,464 and 1,229,312 features, respectively. This approach thus provides us with 1,245,944 compound features, each of which is a function involving up to 3 primary fea- tures. Finally, a least absolute shrinkage and selection operator (LASSO)-based model selection is employed to downselect a set of 40 compound features, which are deemed most relevant towards prediction of the bandgaps. We note here that this strategy of creating a large number of initial compound features and down-selecting to the most relevant ones using LASSO has recently been successfully employed to identify new crystal structure classifiers . A LASSO-based formalism has also been employed to identify lower-dimensional representations of alloy cluster expansions . Next, the primary features and downselected compound features are subjected to a Pearson correlation filter (cf. Fig. 2) to remove features that exhibit a high correlation with the other features in each set. e Th cut-off of the Pearson correlation filter was adjusted such that only 16 features in each set survive. Tests showed that selecting more than 16 features does not lead to any improvements in the out-of-sample prediction accuracy of ML mod- els. A Pearson correlation map showing the correlation for each pair of the primary or compound features is presented in Fig. 3. The above sets of 16-primary and 16-compound features were subsequently used separately to construct all possible Ω -dimensional (or Ω − D) descriptors (i.e., taking Ω features at a time), where Ω was varied from 1 to 16. This leads to 2 − 1(= 65535) total possible descriptors to be tested for the primary and the compound feature sets. Since testing and evaluating prediction performance of such a large number of descriptors using non-linear statistical learning models (such as kernel ridge regression or KRR) is a highly computation-time intensive task, we resort here to a cross-validated linear least square fit (LLSF) model instead. A training set consisting of 90% of the whole dataset was used to fit a linear model, and the rest 10% was used as a test set to evaluate root mean squared (rms) error and coefficient of determination ( R ) of the fit. To take into account variability of the models, average test set rms error and average test set R over 100 different bootstrap runs were used to rank the linear models. The LLSF performance of the best Ω − D descriptors for a given Ω ∈ [1, 16] is presented in Fig. 4a. We find that for any given Ω , the descriptors with the compound features perform much better than those formed using the primary features. Certainly, this boost in performance can be attributed to the additional flexibility imparted Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 4 www.nature.com/scientificreports/ Figure 4. Prediction performance of the developed linear and non-linear learning models. (a) Computed test set rms errors and coefficient of determinations ( R ) for the best Ω − D models, with Ω ranging from 1 through 16, in linear least square fit (LLSF) models which are built on either primary or compound features. The prediction performance reported here is computed as an average over 100 different runs, each with randomly selected 90% training and 10% test sets. Representative parity plots comparing the DFT-computed bandgaps against the KRR predicted bandgaps, for (b) the 4-D primary descriptor, (c) the 16-D primary descriptor and (d) the 16-D compound descriptor. Histograms of the test set average rms error, computed over 400 different runs, are also presented as insets in the last three panels. by the non-linear functions in the compound features. Furthermore, a compound feature can effectively have a combination of up to three functions of primary features. We also note that going beyond a 10-D descriptor does not improve the prediction performance in either case (cf. Fig. 4a). For instance, the average rms error for the 16-D descriptor formed with compound features is 0.786 eV, while that for the 10-D descriptor is 0.792 eV. The average rms errors for the corresponding descriptors with primary features are 0.971 eV and 0.973 eV, respectively. Performance of the best primary and compound Ω − D descriptors (with Ω ∈[1, 16]) identified above was 53,54 then reassessed in a Kernel ridge regression (KRR) model—a state-of-the-art ML method capable of handling complex non-linear relationships—which has recently been shown to be promising for prediction of a diverse set 2,17,55–57 of materials properties . Based on the principle of similarity, the KRR method first uses a distance measure 2 th th such as the Euclidean norm in the descriptor space (i.e., , for i and j compounds in the training set) ‖ff − ‖ ij to quantify (dis)similarity between materials; the property to be predicted is then computed as a linear combina- tion of the kernel (e.g., a gaussian kernel used in the present case) distance k (, ij)= exp2 −− ‖‖ ff / σ () ij functions of materials of interest and the training set materials. Therefore, constructing descriptors in which materials have a small distance when their property of interest is similar is of particular importance for the learn- ing process. Further details on the KRR learning model are provided in Methods. Results obtained using the cross-validated KRR models are presented in Table 1 and the identities of the best Ω − D descriptors are provided in the Supplementary Information. For each descriptor, the average rms error and average R on training and test sets are reported. The average was taken over 100 different KRR runs, in each of which a 90% training set and a 10% test set were randomly selected. Not surprisingly, it is seen that both the learning and prediction accuracies grow with the descriptor dimensionality (and complexity). Interestingly, Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 5 www.nature.com/scientificreports/ Performance of KRR ML models with primary Performance of KRR ML models with descriptors compound descriptors Training set Training set Descriptor (90%) Test set (10%) Descriptor (90%) Test set (10%) rms rms rms rms error error error error 2 2 2 2 Dimension (eV) R (eV) R Dimension (eV) R (eV) R 1-D 0.959 0.634 1.056 0.546 1-D 0.954 0.638 1.056 0.545 2-D 1.059 0.554 1.114 0.493 2-D 0.888 0.686 0.879 0.685 3-D 0.689 0.811 0.867 0.692 3-D 0.774 0.762 0.777 0.753 4-D 0.306 0.963 0.501 0.897 4-D 0.716 0.796 0.742 0.775 5-D 0.270 0.971 0.529 0.885 5-D 0.566 0.872 0.637 0.834 6-D 0.302 0.968 0.521 0.889 6-D 0.502 0.900 0.594 0.855 7-D 0.356 0.949 0.510 0.892 7-D 0.426 0.928 0.554 0.875 8-D 0.281 0.969 0.429 0.925 8-D 0.437 0.924 0.563 0.870 9-D 0.219 0.981 0.406 0.932 9-D 0.399 0.937 0.550 0.876 10-D 0.139 0.992 0.397 0.935 10-D 0.300 0.964 0.484 0.904 11-D 0.178 0.987 0.413 0.930 11-D 0.249 0.975 0.455 0.915 12-D 0.074 0.998 0.393 0.936 12-D 0.222 0.980 0.451 0.917 13-D 0.137 0.993 0.365 0.944 13-D 0.187 0.986 0.418 0.928 14-D 0.110 0.995 0.397 0.935 14-D 0.169 0.989 0.413 0.930 15-D 0.087 0.997 0.377 0.941 15-D 0.144 0.992 0.376 0.942 16-D 0.080 0.997 0.371 0.939 16-D 0.132 0.993 0.360 0.947 Table 1. Prediction performance of the KRR models. Computed training set and test set rms errors and coefficient of determinations (R ) for the best Ω − D models, with Ω ∈ [1, 16], in 5-fold cross-validated KRR models built on top primary and compound descriptors (identified in the LLSF models). The reported rms errors and R are averaged values over 100 different runs, each with randomly selected 90% training and 10% 2 2 2 test sets. Here R =− 1 (− bb )/ (− bb ) with b , b and b being the GLLB-SC computed ∑ ∑ i i,pred avg ii,pred i avg bandgaps, model-predicted bandgaps and average bandgap value in the training or test set. unlike the LLSF model, in the KRR model the performance of a primary descriptor is found to be comparable to that of the corresponding compound descriptor. This is owing to the inclusion of the nonlinearity in the learning model itself, which boosts the performance of the primary descriptors. In light of this observation, going forward with the KRR model, we choose the simpler models with the primary descriptors over the compound descriptors. It is interesting to note that going from the 3-D to the 4-D primary descriptor (cf. Table 1) leads to a significant improvement in the model prediction performance. For instance, the average R on the test set increases from 0.69 to 0.90 and average rms error decreases from ~0.87 eV to ~0.50 eV. Going beyond the 4-D descriptor, how- ever, only results in marginal improvements. For instance, with the 16-D descriptor (containing all of the primary features) the obtained average test set R is ~0.94, only slightly better than that of the 4-D descriptor. Figure 4b–d compare the KRR prediction performance of the 4-D descriptor with the 16-D primary and the 16-D compound descriptor in separate parity plots, using a representative training/test set split. It can be seen that while the train- ing set performance is significantly better in the KRR models with the higher dimensional descriptors, the test set performance of those models can be considered comparable (or only slightly better) to that of the 4-D descriptor. Discussion We now examine the individual primary features that combine to give the 4-D primary descriptor. These features + − + − are: l , l , , , i.e., the absolute sum and difference of elemental lowest occupied levels of the two A-site atoms χ χ A A B B and the electronegativities of the two B site atoms. We also note that the two conjugate pairs of these elemental features appear together and none of the primary features with valence orbital radii appears. Furthermore, the descriptor is well balanced with respect to the participation from the features specific to the A-site atoms (i.e., l , − + − l ) and to the B-site atoms (i.e., χ , χ ). In addition to being chemically intuitive, elegant and symmetric, the B B identified descriptor is also simple and easily accessible. It is always desirable to have a ML prediction model built on simpler (i.e., low dimensional) and intuitive descriptors, since with high dimensional complex descriptors there is always a danger of overt fi ting leading to poor model generalizability. Therefore, by preferring the 4-D primary descriptor over the 16-D descriptor, we are trading some model accuracy for model simplicity and better model generalizability. To further test the model’s predictability, we used the cross-validated KRR learning model, trained on a ran- domly selected 90% double perovskite dataset, to predict bandgaps of the original 53 parent single perovskites. We note that for the single perovskites, owing to the constraints A = A’ and B = B’, only two of the four features − − survive (i.e., for all the single perovskites we have l = 0, χ = 0). Figure 5 compares the bandgaps predicted by the model with those computed using DFT with the GLLB-SC functional. Given that the model was never trained on single perovskites and that only two of the four primary features effectively survive for a single perovskite, such a prediction performance is rather remarkable. Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 6 www.nature.com/scientificreports/ Figure 5. Predictions on single perovskites. Prediction performance of the KRR model in predicting the bandgaps of the parent 53 single perovskites, which form all the double perovskites in the database. The model was trained on a randomly selected 90% training set from the double perovskite bandgap database using the 4-D primary descriptor. To gain deeper insight into the model’s remarkable prediction performance, we next construct 2-D contour plots in which dependence of any two of the four features has been marginalized (by considering an averaged value along those particular dimensions, as explained below). We start with a fine 4-D grid in the feature space constituted by the four primary features identified above, while still confining ourselves within the boundaries of the original feature space used to train the KRR model. Each point on this grid then, in principle, represents a descriptor. Next, we use the trained KRR model to make predictions using each of these descriptor points as a model input. For the sake of better representation, we convert the predictions in this 4-D feature space into a 2-D plot by averaging out any given two of the four primary features. This approach allows us to explicitly visualize the dependence of the bandgap along any two pre-specified features, while the dependence of other two features is considered only in an averaged manner. We can now represent this data in a 2-D contour plot. Three out of a total of six such possible plots are shown in Fig. 6, where green and purple regions represent the high- and low-bandgap regions, respectively. The data-points in the double perovskites dataset are also plotted on top for validation, color (or size) coded according to their GLLB-SC bandgaps. Since the dependence of two out of the four features has already been integrated out, one does not expect a quantitative agreement between the contour and scatter plots. However, it can be seen from the figure that the two are in quite good agreement. The green “mountains” on the contour plot are largely occupied by red (large) circles while the purple “valleys” are mostly populated with blue (small) circles. Such feature-property maps provide a pathway towards drawing decision rules (for a targeted functionality) from statistical learning models. Furthermore, while the original model can be used to make quantitative predictions, such simple feature-property maps can be employed as a first-line of screening to make qualitative predictions or devise simple screening criteria for a given property (in our case the bandgap). Finally, we comment about the limitations and domain of applicability of the ML model. The presented model is applicable within the considered chemical space (i.e., aforementioned choices of A- and B-site cations) and to non-magnetic AA’BB’O type perovskites, which can be separated into two charge neutral ABO and A’B’O single 6 3 3 perovskites. Test performance on double perovskite compounds which cannot be decomposed in such a manner was found to be poor, which is not surprising since the learning model was never trained on such compounds. Extending the ML model to such compounds and accounting for other possible A- and B-site cation orderings remains work to be undertaken in future studies. It will also be interesting to check the general applicability of the identified descriptor by employing it to predict the bandgaps of other materials classes, quite distinct from perovskites and related chemistries. In summary, we have presented a robust ML model along with a simple elemental descriptor set for efficient predictions of electronic bandgaps of double perovskites. The proposed optimal descriptor set was identified via searching a large part of feature space that involved more than ~1.2 million descriptors formed by combining simple elemental features such as electronegativities, ionization potentials, electronic energy levels and valence orbital radii of the constituent atomic species. The KRR-based statistical learning model developed here was trained and tested on a database consisting of accurate bandgaps of ~1300 double perovskites computed using the GLLB-SC functional within the framework of density functional theory. One of the most important chem- ical insights that came out of the adopted learning framework is that the bandgap is primarily controlled (and therefore can efficiently be learned) by the lowest occupied energy levels of the A-site elements and electronega - tivities of the B-site elements. The average test set rms error of the cross validated model with only four primary features (i.e., the 4-D primary descriptor) is found to be 0.5 eV, which is further reduced to ~0.37 eV (0.36 eV) Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 7 www.nature.com/scientificreports/ Figure 6. Feature-property maps. (a–c) Two dimensional contour maps showing regions of relatively larger and smaller bandgaps in the feature space. In each of the panels, only two of the four features are explicitly presented, while the dependency of the other two features has been integrated out (see text for details). The data-points in the double perovskites dataset are also plotted on top of the contour plots. The marker size and color of the scattered points represent their bandgaps. with the primary (compound) 16-D descriptor. Out-of-sample prediction performance of the trained model is further demonstrated by its ability to predict bandgaps of several single perovskites. Finally we have shown that the prediction performance of the model can be visually rationalized by constructing several two-dimensional feature-property contour maps. We believe that the ML approach presented here is general and can be applied to any material class in a restricted chemical space with a given crystal structure to make efficient predictions of bandgaps. Such a prediction strategy can be practically useful in an initial screening to identify promising candi- dates in a high throughput manner. Methods Details of feature set. For feature set accumulation, we start from 7 atomic features for each metal atom A in the double perovskite structure. These primary atomic features are Pauling’s electronegativity ( χ), ionization potential (I), highest occupied atomic Kohn-Sham level (h), lowest unoccupied atomic Kohn-Sham level (l), and s-, p- and d- Zunger’s valence orbital radii r , r and r of isolated neutral atoms . The ionization potential s p d 6,59 and Pauling’s electronegativity data were taken from the literature and the highest-occupied lowest-unoccu- pied Kohn-Sham levels of the isolated atomic species were computed using the GGA-PBE exchange-correlation functional . Machine learning model. Within the present similarity-based KRR learning model, the bandgap of a sys- tem in the test set is given by a sum of weighted Gaussians over the entire training set. As a part of the model N i i 2 N 2 training process, the learning is performed by minimizing the expression , (− EE )+ λω ∑ ∑ i=1 gK ,, RR gDFT i=1 i i i with being the KRR estimated bandgap value, the DFT value, and λ a regularization parameter. The E E gK , RR gD , FT −1 explicit solution to this minimization problem is ω =(kI +) λ E , where I is the identity matrix, and gD , FT is the kernel matrix elements of all compounds in the training set. The parameters λ, k =( exp −− ff ) ij 2 ij 2σ σ are determined in an inner loop of fivefold cross validation using a logarithmically scaled fine grid. We note that KRR training and hyperparameter determination were performed only using the training data and the test set samples were never seen by the KRR model during the training procedure. References 1. Curtarolo, S. et al. The high-throughput highway to computational materials design. Nat. Mater. 12, 191–201 (2013). 2. Pilania, G., Wang, C., Jiang, X., Rajasekaran, S. & Ramprasad, R. Accelerating materials property predictions using machine learning. Sci. rep. 3, 2810 (2013). 3. Sharma, V. et al. Rational design of all organic polymer dielectrics. Nat. comm. 5, 4845 (2014). 4. Ceder, G., Hauthier, G., Jain, A. & Ong, S. P. Recharging lithium battery research with first-principles methods. Mater. Res. Soc. Bull. 36, 185–191 (2011). 5. Curtarolo, S. et al. AFLOWLIB.ORG: AFLOWLIB. ORG: A distributed materials properties repository from high-throughput ab initio calculations. Comput. Mater. Sci. 58, 227 (2012). 6. Materials Project - A Materials Genome Approach, http://materialsproject.org/ (accessed: 15th October 2015). 7. Computational Materials Repository https://wiki.fysik.dtu.dk/cmr/ (Documentation) and https://cmr.fysik.dtu.dk/ (accessed: 15th October 2015). 8. Service, R. F. Materials scientists look to a data-intensive future. Science 335, 1434–1435 (2012). 9. Flach, P. Machine Learning: The Art and Science of Algorithms that Make Sense of Data (Cambridge University Press, Cambridge, 2012). 10. Burbidge, R., Trotter, M., Buxton, B. & Holden, S. Drug design by machine learning: support vector machines for pharmaceutical data analysis. Computers & chemistry 26, 5–14 (2001). 11. Jones, N. Quiz-playing computer system could revolutionize research. Nature News (2011), Available at: http://dx.doi.org/10.1038/ news.2011.95. (Accessed: 23rd November 2015). 12. MacLeod, N., Benfield, M. & Culverhouse, P. Time to automate identification. Nature 467, 154–155 (2010). Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 8 www.nature.com/scientificreports/ 13. Abu-Mostafa, Y. S. Machines that Think for Themselves. Sci Am 307, 78–81 (2012). 14. Silver, N. e S Th ignal and the Noise: Why So Many Predictions Fail but Some Don’t (Penguin Press, New York, 2012). 15. Mueller, T., Kusne, A. G. & Ramprasad, R. Machine learning in materials science: Recent progress and emerging applications. Rev. Comput. Chem. (Accepted for publication). 16. Rajan, K. in Informatics for Materials Science and Engineering: Data-driven Discovery for Accelerated Experimentation and Application (ed. Rajan, K.), Ch. 1, 1–16 (Butterworth-Heinemann, Oxford, 2013). 17. Rupp, M., Tkatchenko, A., Muller, K.-R. & von Lilienfeld, O. A. Fast and accurate modeling of molecular atomization energies with machine learning. Phys. Rev. Lett. 108, 058301 (2012). 18. Huan, T. D., Mannodi-Kanakkithodi, A. & Ramprasad, R. Accelerated materials property predictions and design using motif-based fingerprints, Phys. Rev. B 92, 014106 (2015). 19. Schütt, K. T. et al. How to represent crystal structures for machine learning: Towards fast prediction of electronic properties. Phys. Rev. B 89, 205118 (2014). 20. Meredig, B. et al. Combinatorial screening for new materials in unconstrained composition space with machine learning. Phys. Rev. B 89 094104 (2014). 21. Faber, F., Lindmaa, A., von Lilienfeld, O. A. & Armiento, R. Crystal Structure Representations for Machine Learning Models of Formation Energies. Int. J. Quantum. Chem. 115, 1094–1101(2015). 22. Faber, F., Lindmaa, A., von Lilienfeld, O. A. & Armiento, R. Machine Learning Energies of 2 M Elpasolite (ABC2D6) Crystals. http:// arxiv.org/abs/1508.05315 (2015). 23. Pozun, Z. et al. Optimizing transition states via kernel-based machine learning. Chem. Phys. 136, 174101 (2012). 24. Behler, J. Atom-centered symmetry functions for constructing high-dimensional neural network potentials. J. Chem. Phys. 134, 074106 (2011). 25. Botu, V. & Ramprasad, R. Adaptive machine learning framework to accelerate ab initio molecular dynamics, Int. J. Quantum Chem. 115, 1074–1083 (2015). 26. Pilania, G., Gubernatis, J. E. & Lookman, T. Structure classification and melting temperature prediction in octet AB solids via machine learning. Phys. Rev. B 91, 214302 (2015). 27. Pilania, G., Gubernatis, J. E. & Lookman, T. Classification of octet AB-type binary compounds using dynamical charges: A materials informatics perspective. accepted for publication in Sci. Rep. (2015). 28. Pilania, G., Balachandran, P. V., Gubernatis, J. E. & Lookman, T. Predicting the formability of ABO perovskite solids: A machine learning study. Acta Cryst. B 71, 507–513 (2015). 29. Snyder, J. C., Rupp, M., Hansen, K., Müller, K. R. & Burke, K. Finding density functionals with machine learning. Phys. Rev. Lett. 108, 253002 (2012). 30. Lee, J., Seko, A., Shitara, K. & Tanaka, I. Prediction model of band-gap for AX binary compounds by combination of density functional theory calculations and machine learning techniques. arXiv preprint arXiv:1509.00973 (2015). 31. P., Dey et al. Informatics-aided bandgap engineering for solar materials. Com. Mat. Sci. 83, 185–195 (2014). 32. Mitchell, R. H. Perovskites: Modern and Ancient (Almaz Press, Ontario, Canada, 2002). 33. Setyawan, W., Gaume, R. M., Lam, S., Feigelson, R. S. & Curtarolo, S. High-throughput combinatorial database of electronic band structures for inorganic scintillator materials. ACS Comb. Sci. 13, 382–390 (2011). 34. Olivares-Amaya, R. et al. Accelerated computational discovery of high-performance materials for organic photovoltaics by means of cheminformatics. Energy Environ. Sci. 4, 4849 (2011). 35. Chemical Bonding at Surfaces and Interfaces (Eds Nilsson, A., Pettersson, L. G. M. & Nørskov, J. K.) (Elsevier, Amsterdam, The Netherlands, 2008). 36. Hedin, L. New method for calculating the one-particle Green’s function with application to the electron-gas problem. Phys. Rev. 139 A796 (1965). 37. Heyd, J., Scuseria, G. E. & Ernzerhof, M. Hybrid functionals based on a screened Coulomb potential. J. Chem. Phys. 124, 219906 (2006). 38. Castelli, I. E. et al. Computational screening of perovskite metal oxides for optimal solar light capture. Energy Environ. Sci. 5, 5814 (2012). 39. Castelli, I. E., Thygesen, K. S. & Jacobsen, K. W. Bandgap engineering of double perovskites for one-and two-photon water splitting. MRS Proceedings 1523, mrsf12-1523-qq07-06 (2013), doi: 10.1557/opl.2013.450. 40. Vasala, S. & Karppinen, M. A B’B”O perovskites: A review. Prog. Solid State Chem. 43, 1–36 (2015). 2 6 41. Martin, R. Electronic Structure: Basic Theory and Practical Methods (Cambridge University Press, New York, 2004). 42. Mortensen, J. J., Hansen, L. B. & Jacobsen, K. W. Real-space grid implementation of the projector augmented wave method. Phys. Rev. B 71, 35109 (2005). 43. Gritsenko, O., van Leeuwen, R., van Lenthe, E. & Baerends, E. J. Self-consistent approximation to the Kohn-Sham exchange potential. Phys. Rev. A 51, 1944 (1995). 44. Kuisma, M., Ojanen, J., Enkovaara, J. & Rantala, T. T. Kohn-Sham potential with discontinuity for band gap materials. Phys. Rev. B 82, 115106 (2010). 45. Talman, J. D. & Shadwick, W. F. Optimized effective atomic central potential. Phys. Rev. A 14, 36 (1976). 46. Castelli, I. E. et al. New light-harvesting materials using accurate and efficient bandgap calculations. Adv. Energy Mater. 5, 1400915 (2015). 47. Andriotis, A. N. et al. Informatics guided discovery of surface structure-chemistry relationships in catalytic nanoparticles. J. Chem. Phys. 140, 094705 (2014). 48. Dam, H. C., Pham, T. L., Ho, T. B., Nguyen, A. T. & Nguyen, V. C. Data mining for materials design: A computational study of single molecule magnet. J. Chem. Phys. 140, 044101 (2014). 49. Brown, R. D. & Martin, Y. C. The information content of 2D and 3D structural descriptors relevant to ligand-receptor binding. J. Chem. Inf. Comput. Sci. 37, 1 (1997). 50. Zunger, A. Systematization of the stable crystal structure of all AB-type binary compounds. Phys. Rev. B 22, 5839 (1980). 51. Ghiringhelli, L. M., Vybiral, J., Levchenko, S. V., Draxl, C. & Scheffler, M. Big data of materials science: Critical role of the descriptor. Phys. Rev. Lett. 114, 105503 (2015). 52. Nelson, L. J., Hart, G. L., Zhou, F. & Ozoliņš, V. Compressive sensing as a paradigm for building physics models. Phys. Rev. B 87, 035125 (2013). 53. Hastie, T., Tibshirani, R. & Friedman, J. e E Th lements of Statistical Learning: Data Mining, Inference, and Prediction (Springer, New York, 2009). 54. Müller, K.-R., Mika, S., Ratsch, G., Tsuda, K. & Scholkopf, B. An introduction to kernel-based learning algorithms. IEEE Trans Neural Netw 12, 181–201 (2001). 55. Bereau, T., Andrienko, D. & von Lilienfeld, O. A. Transferable atomic multipole machine learning models for small organic molecules. J. Chem. Theory Comput. 11, 3225–3233 (2015). 56. Hansen, K. et al. Assessment and validation of machine learning methods for predicting molecular atomization energies. J. Chem. e Th ory Comput. 9, 3404 (2013). 57. Lopez-Bezanilla, A. & von Lilienfeld, O. A. Modeling electronic quantum transport with machine learning. Phys. Rev. B 89, 235411 (2014). Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 9 www.nature.com/scientificreports/ 58. Perdew, J. P., Burke, K. & Ernzerhof, M. Generalized gradient approximation made simple. Phys. Rev. Lett. 77, 3865 (1996). 59. Lide, D. R. Handbook of Chemistry and Physics (CRC Press, Boston, 2004). Acknowledgements G.P., J.E.G. and T.L. acknowledge support from the LANL LDRD program and B.P.U. acknowledges support by the U.S. Department of Energy, Office of Science, Basic Energy Sciences, Materials Sciences and Engineering Division. Los Alamos National Laboratory is operated by Los Alamos National Security, LLC, for the National Nuclear Security Administration of the (U.S.) Department of Energy under contract DE-AC52-06NA25396. G.P. gratefully acknowledges discussions with Ivano E. Castelli. Author Contributions G.P. designed the study with inputs from T.L., J.E.G., B.P.U. and R.R. A.M.K. and G.P. built the data sets and the machine learning models. All authors analyzed and discussed the results, and wrote the manuscript. Additional Information Supplementary information accompanies this paper at http://www.nature.com/srep Competing financial interests: e a Th uthors declare no competing financial interests. How to cite this article: Pilania, G. et al. Machine learning bandgaps of double perovskites. Sci. Rep. 6, 19375; doi: 10.1038/srep19375 (2016). This work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder to reproduce the material. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ Scientific RepoRts | 6:19375 | DOI: 10.1038/srep19375 10

Journal

Scientific ReportsSpringer Journals

Published: Jan 19, 2016

There are no references for this article.