Next Article in Journal
Improving the Adhesiveness of Cemented Glass Components by DBD Plasma Pre-Treatment at Atmospheric Pressure
Previous Article in Journal
The Classification of Inertinite Macerals in Coal Based on the Multifractal Spectrum Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computational Screening of New Perovskite Materials Using Transfer Learning and Deep Learning

1
School of Mechanical Engineering, Guizhou University, Guiyang 550025, China
2
Key laboratory of advanced manufacturing technology, Ministry of education, Guiyang 550025, China
3
Department of Computer Science and Engineering, University of South Carolina, Columbia, SC 29208, USA
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2019, 9(24), 5510; https://doi.org/10.3390/app9245510
Submission received: 8 November 2019 / Revised: 2 December 2019 / Accepted: 12 December 2019 / Published: 14 December 2019
(This article belongs to the Section Materials Science and Engineering)

Abstract

:
As one of the most studied materials, perovskites exhibit a wealth of superior properties that lead to diverse applications. Computational prediction of novel stable perovskite structures has big potential in the discovery of new materials for solar panels, superconductors, thermal electric, and catalytic materials, etc. By addressing one of the key obstacles of machine learning based materials discovery, the lack of sufficient training data, this paper proposes a transfer learning based approach that exploits the high accuracy of the machine learning model trained with physics-informed structural and elemental descriptors. This gradient boosting regressor model (the transfer learning model) allows us to predict the formation energy with sufficient precision of a large number of materials of which only the structural information is available. The enlarged training set is then used to train a convolutional neural network model (the screening model) with the generic Magpie elemental features with high prediction power. Extensive experiments demonstrate the superior performance of our transfer learning model and screening model compared to the baseline models. We then applied the screening model to filter out promising new perovskite materials out of 21,316 hypothetical perovskite structures with a large portion of them confirmed by existing literature.

Graphical Abstract

1. Introduction

The perovskite structure is one of the most common and widely studied structures in materials science. The general chemical formula of the perovskite compound is ABX3, wherein A and B are two different sized cations, and X is an anion bonded to both. Its ideal structure is a cubic structure, and B atoms are at the center of a typical anionic octahedron. This seemingly simple atomic arrangement is deceptive because it hides its diversity of special physical and chemical properties.
As a result of its rich and remarkable properties, perovskites are used in many technical fields. Examples include piezoelectric perovskites, lead zirconate titanate, perovskites with piezoelectric effect for sensors or actuators [1] and high temperature perovskite superconductors such as beryllium copper oxide [2,3]. Some perovskite groups (mainly manganese-based perovskite oxides) exhibit huge magnetoresistance, which can significantly change the electrical resistance in the presence of a magnetic field [4]. In addition, perovskites have also been studied and applied to other fields such as thermoelectric materials [5], catalysts [6,7], light emitting diodes (LEDs) [8,9], lasers [10], and so on. In recent years, since perovskite can be used as an absorbent material for solar cells, it has become a research hotspot. In 2009, Miyasaka et al. [11] made CH3NH3PbBr3 and CH3NH3PbI3 as photosensitizers for dye-sensitized solar cells for the first time, with an efficiency of 3.8%, which laid the foundation for the development of perovskite solar cells. In 2012, Grätzel and Park [12] first used solid Spiro-OMeTAD as a hole transport layer to make a solid perovskite solar cell with an efficiency of 9.7%. In 2014, Hongwei [13] and others from Huazhong University of Science and Technology used a full-printing method to prepare a hole-free transport layer, and a mesoporous structured perovskite solar cell using carbon as the electrode, which achieved an efficiency of 13.4%. In 2016, the Korean Institute of Chemistry (KRICT) and the Ulsan University of Science and Technology (UNIST) jointly developed a perovskite battery with an efficiency of 22.1% [14], making it the most energy-efficient perovskite solar cell. In 2019, Lin et.al. [15] at Nanjing University developed a strategy to reduce tin vacancies in Pb-Sn narrow-band perovskites by a neutralization reaction, thereby improving the performance and stability of a perovskite series tandem solar cell. The large area of perovskite series tandem solar cells achieves 24.8% and 22.1% certification efficiency, respectively. Perovskite solar cell efficiency records are constantly being refreshed, and research results of perovskite solar cells continue to emerge [16]. Since perovskite materials were first used in solar cells in 2009, in just ten years, their energy conversion efficiency has reached 22.1%, far exceeding other thin film solar cells, and has broad commercial prospects.
In recent years, with the accumulation of material database resources, data mining [17] and machine learning (ML) [18] have been used more and more frequently in material research, platform design and analysis, and prediction of material properties [19]. In terms of new material discovery, machine learning algorithms have been used to discover new energy materials [20], soft materials [21], polymer dielectrics [22], etc., and have achieved remarkable results. While machine learning has become a promising tool for scientists, it also has a distinct disadvantage: it usually requires a large amount of training datasets (e.g., 104–106). This is usually not feasible in materials science because the data set (mainly the property characterization data) for most interesting material properties, such as ion-conductivity, has approximately 101–103 samples. Standard machine learning methods that use generic descriptors do not work well with small data unless it contains highly informative physical and structural descriptors. For example, Saad et al. [23] trained a ML model with 44 materials to predict the melting temperature of octagonal compounds; Seko et al. [24] used 101 materials to establish a lattice thermal conductivity model; Ghiringhelli et al. [25] established an energy difference model for sphalerite and rock salt phases using 82 materials; Reed et al. [26] proposed to transfer structural information descriptors to general descriptors, so that billions of unknown lithium-ion conductive components can be screened. All these studies emphasized the importance of descriptors for their ML models.
This paper proposes a transfer learning approach that develops a transfer learning model for annotating a large number of materials samples with structural information but without formation energy information. The enlarged annotated training set then trains a high-performance convolution neural network model with generic Magpie elemental features to predict the formation energy of hypothetical perovskite materials for which only the composition and stoichiometry are available without crystal structure information. The main contributions of this paper are as follows:
(1)
We proposed a transfer learning strategy to convert formation energy related structural features/insights into training data for a perovskite screening model using only elemental Magpie features. This enables us to address the small dataset issue in typical ML based materials discovery.
(2)
We developed a gradient boosting regressor (GBR) ML model trained with structural and elemental features for perovskite formation energy prediction, which outperforms the state-of-the-art artificial neural network (ANN) based model trained with two elemental descriptors. This highly accurate model allows us to annotate the large number of material samples with structural information but no formation energy.
(3)
We built a convolutional neural network model trained with the enlarged large dataset together with generic Magpie elemental descriptors for large-scale screening of hypothetical perovskites.
(4)
Application of our model to a large dataset with 21,316 possible candidates has allowed us to identify interesting stable perovskites available for further experimental or computational density functional theory (DFT) verification.

2. Materials and Methods

All 21,316 compounds are generated by filling the A and B positions (732 × 4 = 21,316) in the ABX3 (X = O, Br, Cl, I) perovskite crystal structure with 73 metals or semi-metals in the periodic table (see Figure 1a) [27]. The ideal ABX3 perovskite crystal structure [28] is composed of an A cation in a 12-coordinate structure located in a cavity composed of octahedrons; and a B cation forms an octahedral coordination with six oxygen ions (see Figure 1b). In addition to the ideal cubic structure, many perovskites undergo local distortions. These distorted perovskites may have a variety of symmetrical structures, including diamond, tetragonal, and orthogonal distortion. In this work, all 21,316 compounds were generated in the ideal cubic structure.
In order to improve the accuracy of the ML model and to screen for perovskites that have never been reported in the literature and cannot be characterized, we propose a transfer learning method [29]. First, we used a descriptor with structural features to train an accurate ML model; then we used the trained ML model to predict the 21,316 perovskites as a label and train the new model through the elemental feature descriptors. Since this data set is much larger, we can train a good prediction model using only generic features. As elemental descriptors only depend on component or stoichiometry information, no structure or other information is needed. Once the accurate general model is trained, it can be used to effectively screen new perovskite materials.

2.1. Materials Dataset Preparation and Features

This study generated 21,316 samples, of which 5329 were from Chris et al. [27] with their stability calculated via high-throughput DFT. We tried to retrieve these 5239 samples from the Material Project (MP) database and obtained 1148 sample data with Crystallographic Information File (CIF) files. Then these samples were divided into two parts (570 and 578, as shown by D1 and D2, respectively), of which the structural features of the 570 samples were calculated by the powerful open source python library Pymatgen [30], and the calculated features and their descriptions are shown in Table 1; the Magpie features of the other 578 samples were calculated by the python library Matminer [31]. The remaining 15,987 samples were generated by replacing the O element with I, Br, and Cl, respectively for all the 5329 ABO3 from Chris et al. Finally, we combined all the data of ABX3 (X = O, I, Cl, Br) leading to a total of 21,316 samples and their Magpie features were calculated. This part of the data set, the D3 as described in Figure 2, was used as the final screening candidates for the convolutional neural network (CNN) screening model.

2.1.1. Structural and Elemental Features

Choosing the right input features is a key step in achieving good predictive performance, thus, the so-called feature engineering. For a materials data set, the features should clearly describe that a single given material can be easily distinguished between different materials. In this study, we selected 31 features, including the elemental and structural features of the atomic composition such as electronegativity, atom radius, band gap, etc. The complete characterization is shown in Table 1.

2.1.2. Magpie Features

The Magpie is a set of extensible attributes, created by Chris et al. [32], that can be used for materials with any number of constituent elements. This set of attributes is broad enough to capture a wide variety of physical/chemical properties that can be used to create accurate models of many material prediction problems. These include stoichiometric characteristics (depending on the proportion of elements only), elemental property statistics (atomic number, atomic radius, melting temperature, etc.), electronic structural properties (valence electrons of s, p, d, and f), and ionic compound characteristics.

2.2. Overview of Our Data-Driven Framework for Computational Screening

The overall framework of our methodology for screening new perovskites is shown in Figure 2. First, a GBR machine learning model (M1) is trained using the hybrid structural and elemental features and the training dataset D1. The M1 model is then used to predict the formation energies of the materials dataset D2 of which all samples come with structural information. D1 and D2 datasets are then combined to train a convolution neural network model M2 using the elemental magpie features which do not need structural information. This M2 model can then be used to do large-scale screening of candidate dataset D3 to identify potential new perovskite materials for further DFT or experimental verification.
In the following sections, we will describe each step of our screening framework.

2.3. GBR Machine Learning Model for Formation Energy Prediction

It has been shown [33] that training a perovskite specific formation energy prediction model using structural information can achieve high accuracy comparable to that of a DFT calculation. Instead of using artificial neural networks and elemental features only, as done before, we propose to use GBR with both structural and elemental features.

Gradient Boosting Regressor

Boosting is a family of algorithms that can promote weak learners to strong learners, and its performance is significantly better than other basic classifiers. Adding new models to the collection in turn is the main idea for improvement. In each particular iteration, a new, basic learner model is trained by the entire integration error that has been learned so far. Its gradient promotion, like other lifting methods, builds the model in stages and uses arbitrary loss functions. It uses the gradient descent method to solve the minimization problem and generates a predictive model in the form of a set of weak predictive models (usually decision trees). Boosting can be used for both regression and classification problems while this study uses it for regression.
The main architecture of GBR includes three elements: the loss function, the weak learner (for prediction), and the addition model. The main idea of this algorithm is to construct a new basic learner, which is most correlated with the negative gradient of the loss function and is related to the whole. Any loss function can be used. In general, the choice of the loss function depends on the user of the algorithm. So far, there have been various loss functions [34]. The mathematical formula for GBR is as follows:
F ( x ) = m = 1 M γ m h m ( x ) ,
where hm(x) is a basic function and is often referred to as a weak learner. GBR uses a fixed-size decision tree as a weak learner. Decision trees have the ability to process mixed-type data and the ability to model complex functions. Like other enhancement algorithms, GBR builds the addition model in a step-by-step manner, with the following formula:
F m ( x ) = F m 1 ( x ) + γ m h m ( x ) ,
At each stage, the decision tree hm(x) selects the minimized loss function L, giving the current model Fm−1 and Fm−1(xi), as follows:
F m ( x ) = F m 1 ( x ) + arg min h i = 1 n L ( y i , F m 1 ( x i ) h m ( x ) ) ,
The initial model F0 is determined for a particular problem, and for least square regression, the average of the target values is typically chosen. Given any divisible loss function L, the algorithm starts with the initial model. GBR solves this minimization problem by gradient descent numerical methods. The gradient descent direction is the negative gradient of the loss function on the current model Fm−1 and can be calculated for any divisible loss function. The formula is:
F m ( x ) = F m 1 ( x ) + γ m i = 1 n F L ( y i , F m 1 ( x i ) ) ,
In the gradient-lifting regression tree, there are multiple hyperparameters. The values of the hyperparameters used in this study are:
  • max_depth = 6; n_estimators = 500; min_samles_split = 0.5;
  • subsample = 0.7; alpha = 0.1; learning_rate = 0.01; loss = ls
To evaluate the performance of GBR, we also evaluated several mainstream machine learning algorithms as the baselines, including random forest regression (RFR), support vector regression (SVR), and least absolute shrinkage and selection operator (Lasso) with the same dataset using the same set of features.
Random forest regression is also a commonly used algorithm in boosting. It is composed of many decision trees trained on each subsample of the dataset and uses their averages to improve prediction accuracy and control overfitting. The sub-sample size is always the same as the original input sample size, but the samples are drawn with replacement if bootstrap = true (default). It is widely used in statistics, data mining, and machine learning. The hyperparameters in random forest (RF) are max_features and n_estimators, where max_features is the number of features to consider when looking for the best segmentation, and n_estimators is the number of trees in the forest. SVR is a commonly used regression algorithm that uses kernel functions to map data from low-dimensional space to high-dimensional space and then uses the support vectors to fit the hyperplane to the final prediction. The SVR has excellent performance in solving prediction problems with high-dimensional features. However, this advantage is reduced when the feature size is much larger than the number of samples. The main hyperparameters in SVR include C, γ, and epsilon, where C is the penalty parameter of the error term and γ is a parameter attached to the rbf function. The last baseline algorithm, Lasso, is a data dimension reduction method that is applicable not only to linear cases but also to nonlinear cases. The Lasso is based on the penalty method to select the variables of the sample data. By compressing the original coefficients, the original small coefficients are directly compressed to 0, so that the variables corresponding to these coefficients are regarded as non-significant variables. These variables are discarded directly. For ordinary linear models, Lasso usually chooses L1 as the penalty term. In this study, all the above machine learning algorithm models and the 10-fold cross-validation method are implemented using the open-source library Scikit-learn [35].

2.4. Structure Information Enabled Transfer Learning and CNN Based Screening ML Model

2.4.1. Transfer Learning

One of the key obstacles of applying machine learning to materials discovery is the limited training data for the most figure of merit (FOM) properties such as the ion-conductivity, thermal conductivity, and formation energy [26]. Here we propose to use a structural information enabled transfer learning method to train a screening model. The basic idea is to train a formation energy prediction model based on structure and elemental features (model M1 in Figure 2) firstly, which usually has high generalization performance due to the informative structural features. This model is then used to predict the formation energies of a large number of samples in the D2 dataset, of which the samples only have structural information but without formation energy. This formation energy annotation step gives us an enlarged dataset (D1 + D2) with a large number of samples with formation energy values. It is then feasible for us to train a deep learning (CNN) based screening model (M2 in Figure 2) based on the enlarged labeled dataset using the Magpie elemental features only. The independence of M1 on the structural feature here is essential as most hypothetical materials only have composition and/or stoichiometry information without the crystal structure information.

2.4.2. Convolutional Neural Network Model

As one of the most successful deep learning models, the convolutional neural network has a special structure compared to traditional neural network models. With multiple data input channels, it can receive multi-dimensional input data, and the complexity of the network model, which greatly reduces the weight sharing structure, reduces the amount of calculation. The neural unit sharing the weights can use the layer-by-layer feature mapping function to perform multi-level understanding of the input data. A general convolutional neural network consists of an input layer, one or more convolutional layers, one or more sampling/pooling layers, a few fully connected layers, and an output layer.
As the core structure in the convolutional neural network, the convolutional layer uses different scale convolution kernels to traverse the input data in the way of weight sharing and extracts different levels of data features for the same data sample through different parameter distributions. The extracted feature maps of different features are stored in different channels of the convolution network in a feature stacking manner to form a high-dimensional data matrix to be used in the next calculation. The calculation formula of the convolution kernel is as follows:
x k l = j = M k x j l 1 w j k l + b k l ,
In Equation (5), x k l represents the kth feature of the lth layer; x j l 1 represents the output of the jth feature of the previous layer; ω j k l represents the l −1th feature of the layer and the kth feature of the layer of the convolution kernel; b k l represents the offset of the kth feature of layer l; M k is the set of all features output after the convolution operation of layer l − 1 and layer l.
The pooling layer [36] is also called the downsampling layer. After acquiring the features through the convolution layer operation, it is usually useful to use the pooling layer to sample the features calculated by the previous convolution layer to reduce the data dimension and the computational overhead due to the limitation of computing resources and time overhead.
The feature map obtained by calculating the upper convolution layer is divided into non-overlapping rectangular regions, and the operation of taking the maximum value for each rectangular region is called maximum pooling, and the operation of averaging is called averaging pooling. Figure 3 shows the matrix obtained by the maximum pooling and average pooling of the convolutional layer output matrices.
The fully connected layer is located at the end of the convolutional neural network and is a common layer for connecting the output features extracted from the convolutional layers to the prediction output to implement tasks such as classification or regression. All neuron nodes in the fully connected layer are connected to the neuron nodes of the previous layer network, and the high-dimensional data obtained from the previous network layers is tiled as an input. Through the activation function which carries out nonlinear transformation, the fully connected layers learn the mapping of extracted abstract features to predict the desired output

2.4.3. The Convolutional Neural Network Training Process

Training of convolutional neural networks refers to training networks with known samples to learn the mapping between the input and output. It is divided into two stages: forward propagation and backward propagation [37]. Forward propagation refers to the input of the sample into the network, and the output value of the network is obtained by weighting the weight, offset, activation function, and full connection layer parameters of the convolution kernel. Backpropagation first calculates the error between the predicted value of the sample obtained from the forward propagation output and the true value of the sample, and then proceeds backward according to the error value to obtain the error information of each layer, and uses the calculated gradients to adjust the network parameters until the network converges or reaches the specified iteration termination condition.
The forward propagation step feeds the training sample into the network and initializes the parameters of each layer of the network. Through the layer-by-layer calculation of the network, the output corresponding to the input sample under the current network parameters is obtained, that is, a forward propagation is completed. In classification, the output vector characterizes the probability distribution of the sample belonging to the corresponding category, which is calculated by the convolutional neural network. In regression, the output is just a real vector which can be compared with the desired values to calculate the regression loss.
The predicted value of the forward propagation output of the convolutional neural network is compared with the true value and their difference is defined as the loss. Usually for regression, mean square error loss function is used. Then, according to the obtained loss function value, the error value, the parameters of each layer of the network are adjusted and updated to minimize the loss function. The parameters that need to be adjusted in the convolutional network are the weights and offsets of the fully connected layer, and the weights and offsets of the convolutional layer.
For the full layer, the weight and offset of the common network can be solved according to the back propagation algorithm of the common network. The formula is as follows:
w i j l = w i j l η w i j l L ,
b i l = b i l η b i l L ,
In Equation (6), w i j l is the weight of the full connection of layer l, b i l is the offset of the full connection of layer l, and η is the network learning rate. According to above formulas, the adjustment process of the network parameter is essentially a process of the loss function’s partial derivative of the weight parameter and the offset, and thus the derivation rule can be obtained:
w i j l L ( w , b ) = a j l δ i l + 1 ,
b i j l L ( w , b ) = δ i l + 1 ,
In Equation (8), a j l is the jth neuron input of the lth layer, and δ i l + 1 in Equation (9) is the ith neuron error of the l + 1th fully connected layer. Through the above two formulas, the partial derivative of the loss function to the weight and the offset can be obtained, thereby completing the update of the parameters of the fully connected layer network.
The weight update formula of the convolutional layer is similar to that of the fully connected layer. The difference is that the partial derivative value makes a 180° rotation operation based on the output of the previous layer. The partial derivative formula is as follows:
w i j l L ( w , b ) = δ i l + 1 r o t 180 ( a j l ) ,
where a j l is the jth neuron input of the lth convolutional layer. δ i l + 1 is the ith neuron error of the l + 1th convolutional layer. For the offset parameters of the convolutional layer, the update method is slightly special. This is because in the convolutional layer, the error δ i l + 1 is a three-dimensional vector, and the offset b i l is a single vector, and the offset update method of the fully connected layer cannot be used. The approach is to sum the respective sub-matrix terms ( δ i l + 1 ) u , v of the error δ i l + 1 to obtain the error vector, which is the gradient of b i l , u and v are the height and width of the gradient of the output image when it is reversed. The partial derivative formula is as follows:
b i j l L ( w , b ) = u , v ( δ i l + 1 ) u , v ,
After the parameters are updated, the training samples will be re-entered into the updated convolutional network model for forward and reverse propagation until the network converges or reaches the specified iteration termination condition, completing the training process.
Figure 4 shows our convolutional neural network model for predicting material formation energy. The CNN input is a 12 × 11 fixed size two-dimensional matrix. The structure of the CNN model consists of three convolutional layers and four fully connected layers. The output of the last convolutional layer is expanded into a one-dimensional vector as input to the subsequent fully connected layer. Both the convolutional layer and the fully connected layer use ReLU as the activation function because it is fast, can help address gradient vanishing problem, and can add sparsity to the network. The output of the network is a continuous value, which is the formation energy of the prediction. Our CNN model uses the Adam optimizer and the MAE (mean absolute error) loss function to train the convolutional neural network. The Adam optimizer combines the advantages of multiple optimizers and demonstrates outstanding performance in many applications. In addition, we used 10-fold cross-validation in the assessment and used root mean square error (RMSE), MAE, and R2 to evaluate the performance of CNN and other machine learning algorithms.

2.5. Verification Whether a Screened ABX3 Material is Perovskite or Non-Perovskite

The ABX3 material after screening by the M2 model is not necessarily a perovskite material. In order to verify whether these ABX3 candidates are stable perovskites, we use a tolerance factor to predict the stability of the perovskite as proposed by Ghirringhell et al. [38]. It can accurately determine whether the selected ABX3 material is perovskite or non-perovskite. It only needs chemical composition to predict the stability of perovskite with τ, which makes it possible to verify perovskite materials with unknown structure. In addition to predicting whether the material is a stable perovskite, τ also provides a monotonic estimate of the probability of the material’s stability in the perovskite structure. Its accuracy and probability, as well as its widespread presence in a single perovskite and double perovskite, provide a new physical insight for the stability of perovskite structure. The formula for τ is as follows:
τ = r X r B n A ( n A r A r B I n ( r A r B ) ) ,
where nA is the oxidation state of A atom, rA and rB are the ionic radii of A and B cations, respectively, rA > rB. A key aspect of τ performance is the degree to which the sum of ionic radii estimates the inter-atomic bond distance for a given structure.

3. Results and Discussions

3.1. Selection of the Best Material Features and Analysis of Feature Importance

Features or descriptors are an important part of machine learning models. In general, choosing different feature descriptors will have a great impact on the prediction results. In order to prove that our calculated material descriptors (annotated by Hybrid_descriptors from now on) can be used to predict the formation energy of perovskites, we compare it with the descriptors proposed by Ong et al. [33] (Ong_Descriptors) and the Magpie features. These three feature sets are used to predict the perovskite formation energy on the same algorithm and dataset, using RMSE, MAE, and R2 as evaluation indicators, all using 10-fold cross-validation. The results are shown in Table 2. Obviously, our feature set is superior over the other two feature descriptors in terms of all three evaluation criteria.
In addition to using a comparative method to validate our proposed feature descriptors, we also analyzed the importance of features using the random forest approach. The random forest consists of a number of decision trees. Every node in the decision trees is a condition on a single feature, designed to split the dataset into two so that similar response values end up in the same set. When training a tree, it can be computed how much each feature decreases the weighted impurity in a tree. For a forest, the impurity decrease from each feature can be averaged and the features are ranked according to this measure. Figure 5 shows the importance scores for all features. It can be seen from the figure that the Pauling electronegativity has a considerable influence on the formation energy of perovskite. It is worth noting that X’s importance score is more than twice higher than the second highest feature, which is consistent with the results of the two literatures [27,39] analyses.

3.2. Performance of the M1 Model with Hybrid Structural and Elemental Features

First, we compared the performances of our GBR and other ML models such as RFR, Lasso, and SVR using the hybrid structural and elemental descriptors as raw features. In order to obtain stable results, each algorithm was evaluated using 10-fold cross-validations ten times. Figure 6 shows the fitting accuracy of all models using the same number of samples. It can be clearly seen that GBR has the best prediction performance, followed by RFR, and the worst is SVR. In terms of RMSE, MAE, and R2 evaluation criteria, the GBR model scores are 0.28, 0.20, and 0.91, respectively. As shown in Table 3, the scores of these three evaluation measures of GBR are the best among all these ML models. In addition to the above three machine learning models, we also tried other machine learning models (such as linear regression models, K-neighbor regression models, etc.) for comparisons. However, their predictions are extremely poor and are therefore not listed here.
The high accuracy of our GBR prediction model can be attributed to the following reasons. One reason is the nonlinear nature of the GBR algorithm; another reason may be the structural similarity of the materials considered in the data set. In a given data set, the crystal structure of all materials is perovskite. Given the same structure, materials with similar chemical compositions may have similar properties, making it more feasible to interpolate properties in predictive models. Discussing structural similarities are beyond the scope of our research, but it has become an important topic in ML research in materials science.

3.3. Performance of M2 Perovskite Screening Model

Before we evaluate the performance of the M2 model, a convolutional neural network model, we first compared the prediction performance before and after data enhancement. First, we used the M1 model to label the structured unlabeled D2 dataset and then calculated the Magpie descriptors for datasets D1 and D2. Finally, D1 and D2 datasets were pooled together to predict the perovskite formation energy with M2, and the D1 dataset was used alone to predict the formation energy of the perovskite. It is worth noting here that the labels of the D2 data set are obtained by migrating the learning model M1, and the label of the D1 data set is calculated by the density functional theory (DFT). The results obtained are shown in Figure 7, with the support of the M1 model, the prediction performance has been significantly improved. This also verifies that the hybrid materials features that we proposed play a role.
To further verify that our chosen convolutional neural network model can be used as the best screening model for screening perovskites, we used the ElemNet model proposed by Ward et al. [40], which is a 17-layer deep neural network model, as comparison. The elemental composition is used to predict the formation energy of the material and then screen the material. In addition to using the ElemNet model for comparison, we also compared it to common machine learning models (RF, GBR, SVR, linear regression, K-neighbor regression, etc.), where RF and GBR performed better. As shown in Table 4, we show the results of our convolutional neural network model, the ElemNet model, and the best two traditional machine learning models, the RF and GBR. The results show that our CNN model is the best in terms of RMSE, MAE, and R2. Thus, our CNN model is used as the ML model for screening hypothetical perovskites. All of the above models are based on the same data set, and all models have been trained and tested using 10-fold cross-validations.

3.4. Screening Results Analysis

Our CNN model is able to make robust, fast predictions, so it can be used to screen 21,316 materials to discover new and stable perovskite materials. After screening by the M2 model, of 21,316 hypothetical ABO3 materials, 4147 had formation energy less than 0. More specifically, 5106 were ABBr3, 5236 were ABCl3, and 4279 were ABI3. The specific formation energy prediction value range distribution is shown in Figure 8.
The candidate materials screened out are not necessarily perovskite materials, even though the formation energy may be less than zero. Here we use the new tolerance coefficient τ as discussed before to further screen the selected candidate materials. If the τ of a candidate material is calculated to be less than 4.18, the candidate material has a high probability of being a stable perovskite material. Since τ requires that the ionic radius of the A cation is greater than the ionic radius of the B cation, a portion of the candidate material is therefore filtered out. After screening the new tolerance coefficient τ, there are 625 ABO3 with τ less than 4.18 along with 52 ABBr3, 55 ABCl3, and 32 ABI3. By reviewing the literature, we found that 98 of the 626 ABO3s were reported in [27], and 98 perovskites were proved to be stable by DFT calculations. In addition, it was reported [41] that the doped lanthanide BaSnO3 can be used as the material of the electron transfer layer for a highly efficient and stable solar cell. This material showed up in our screening results. The specific 98 ABO3s are shown in Table 5. In addition to these 98 compounds, literature [42] proved that one of our predictions, SrMoO3 in 625 ABO3, has paramagnetism. Among the 32 ABI3 screened out by our model, literature [43] verified that CsPbI3 can be surface-coated with surfactants and environmental conditions are stable. It is expected to be used for light collection or LEDs. Among the 52 ABCl3 screened out, literature [44] calculated CsPbCl3 by DFT, and found that the reduction of band gap is due to the limiting effect of carriers, and the increasing number of perovskite layers. Among the 55 ABBr3 screened out, Anni et al. [45] for the first time reported the temperature dependence of the spontaneous emission (ASE) characteristics of CsPbBr3 nanocrystalline thin films. Swarnkar et al. [46] verified the luminescence of colloidal CsPbBr3 perovskite nanocrystals, which surpass traditional quantum dots.
In general, extensive literature inspection shows that our model made reasonable predictions and can be used for discovery of new perovskite materials. The top 200 predicted new perovskites are listed in Table 6. The complete set of predicted perovskites is provided in the supplementary file. The remaining non-reported perovskite materials are also promising.

4. Conclusions

In this paper, we proposed to use deep neural network based transfer learning and a hybrid descriptor set for perovskite formation energy prediction. The hybrid descriptors are composed of structural and elemental features as calculated via the pymatgen library. Using these 31 features, our transfer learning algorithm can be used to address the small data issue typical in machine learning based material discovery. It works by first training an annotation model using structured perovskite data sets and then using it to predict the formation energy of unannotated perovskite materials with structures. The experimental results show that the proposed hybrid feature descriptors perform better than the Ong_Descriptors and Magpie descriptors in predicting the perovskite formation energy. Moreover, compared to the commonly used machine learning model, the gradient-enhanced regression that we used outperformed random forest regression, Lasso, and support vector regression models.
Based on the transfer learning method, we established a convolutional neural network perovskite screening model by first labeling unannotated perovskite materials with high-precision structure feature based model and then built a CNN model and trained it with the Magpie descriptors to get a generic screening model, the M2 model, which does not require structural information. Compared to the ElemNet model and several machine learning models, the experiments show that our CNN model is the best in formation energy prediction of perovskites given only composition information. The ABX3 materials with formation energy greater than 0 were screened out from 21,316 candidates through the CNN model, and the new tolerance factor τ was used to verify whether a screened material is a stable perovskite material. Extensive literature inspection showed that many predicted perovskite materials have been reported in the literature, and the rest is subject to further experimentation or DFT calculation verification.

Supplementary Materials

The following are available online at https://www.mdpi.com/2076-3417/9/24/5510/s1, supplementary file: supplement_file.csv 764 predicted perovskites.

Author Contributions

Conceptualization, J.H. and X.L.; methodology, J.H., X.L., and Y.D.; software, X.L. and Y.D.; validation, X.L., C.N., and J.H.; investigation, X.L., Y.D., Z.C., R.D., Y.S., and J.H.; resources J.H. and S.L.; writing—original draft preparation, X.L. and J.H.; writing—review and editing, X.L., J.H., and S.L.; supervision, J.H. and S.L.; project administration J.H.; funding acquisition J.H. and S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by The National Natural Science Foundation of China under Grant No. 51741101; J.H. and Y.S. were partially supported by the National Science Foundation EPSCoR Program under NSF Award # OIA-1655740, 1940099, and 1905775. S.L. was partially supported by the National Important Project under grant No. 2018AAA0101803 and by Guizhou Province Science and Technology Project under grant No. [2015]4011. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect those of the National Science Foundation.

Acknowledgments

We gratefully acknowledge the support of NVIDIA Corporation with the donation of the Titan X Pascal GPU used for this research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Aksel, E.; Jones, J.L. Advances in lead-free piezoelectric materials for sensors and actuators. Sensors 2010, 10, 1935–1954. [Google Scholar] [CrossRef] [PubMed]
  2. Vinila, V.; Jacob, R.; Mony, A.; Nair, H.G.; Issac, S.; Rajan, S.; Nair, A.S.; Satheesh, D.; Isac, J. Ceramic Nanocrystalline Superconductor Gadolinium Barium Copper Oxide (GdBaCuO) at Different Treating Temperatures. J. Cryst. Process Technol. 2014, 4, 168–176. [Google Scholar] [CrossRef] [Green Version]
  3. Beno, M.; Soderholm, L.; Capone, D.; Hinks, D.; Jorgensen, J.; Grace, J.; Schuller, I.K.; Segre, C.; Zhang, K. Structure of the single-phase high-temperature superconductor YBa2Cu3O7—δ. Appl. Phys. Lett. 1987, 51, 57–59. [Google Scholar] [CrossRef] [Green Version]
  4. Laffez, P.; Tendeloo, G.V.; Seshadri, R.; Hervieu, M.; Martin, C.; Maignan, A.; Raveau, B. Microstructural and physical properties of layered manganites oxides related to the magnetoresistive perovskites. J. Appl. Phys. 1996, 80, 5850–5856. [Google Scholar] [CrossRef]
  5. Maignan, A.; Hébert, S.; Pi, L.; Pelloquin, D.; Martin, C.; Michel, C.; Hervieu, M.; Raveau, B. Perovskite manganites and layered cobaltites: Potential materials for thermoelectric applications. Cryst. Eng. 2002, 5, 365–382. [Google Scholar] [CrossRef]
  6. Song, K.-S.; Kang, S.-K.; Kim, S.D. Preparation and characterization of Ag/MnO x/perovskite catalysts for CO oxidation. Catal. Lett. 1997, 49, 65–68. [Google Scholar] [CrossRef]
  7. Suntivich, J.; May, K.J.; Gasteiger, H.A.; Goodenough, J.B.; Shao-Horn, Y. A perovskite oxide optimized for oxygen evolution catalysis from molecular orbital principles. Science 2011, 334, 1383–1385. [Google Scholar] [CrossRef]
  8. Yuan, M.; Quan, L.N.; Comin, R.; Walters, G.; Sabatini, R.; Voznyy, O.; Hoogland, S.; Zhao, Y.; Beauregard, E.M.; Kanjanaboos, P. Perovskite energy funnels for efficient light-emitting diodes. Nat. Nanotechnol. 2016, 11, 872–877. [Google Scholar] [CrossRef]
  9. Cho, H.; Jeong, S.-H.; Park, M.-H.; Kim, Y.-H.; Wolf, C.; Lee, C.-L.; Heo, J.H.; Sadhanala, A.; Myoung, N.; Yoo, S. Overcoming the electroluminescence efficiency limitations of perovskite light-emitting diodes. Science 2015, 350, 1222–1225. [Google Scholar] [CrossRef]
  10. Veldhuis, S.A.; Boix, P.P.; Yantara, N.; Li, M.; Sum, T.C.; Mathews, N.; Mhaisalkar, S.G. Perovskite materials for light-emitting diodes and lasers. Adv. Mater. 2016, 28, 6804–6834. [Google Scholar] [CrossRef]
  11. Kojima, A.; Teshima, K.; Shirai, Y.; Miyasaka, T. Organometal halide perovskites as visible-light sensitizers for photovoltaic cells. J. Am. Chem. Soc. 2009, 131, 6050–6051. [Google Scholar] [CrossRef] [PubMed]
  12. Kim, H.-S.; Lee, C.-R.; Im, J.-H.; Lee, K.-B.; Moehl, T.; Marchioro, A.; Moon, S.-J.; Humphry-Baker, R.; Yum, J.-H.; Moser, J.E. Lead iodide perovskite sensitized all-solid-state submicron thin film mesoscopic solar cell with efficiency exceeding 9%. Sci. Rep. 2012, 2, 6022–6025. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Yang, Y.; Ri, K.; Mei, A.; Liu, L.; Hu, M.; Liu, T.; Li, X.; Han, H. The size effect of TiO2 nanoparticles on a printable mesoscopic perovskite solar cell. J. Mater. Chem. A 2015, 3, 9103–9107. [Google Scholar] [CrossRef]
  14. Yang, W.S.; Park, B.-W.; Jung, E.H.; Jeon, N.J.; Kim, Y.C.; Lee, D.U.; Shin, S.S.; Seo, J.; Kim, E.K.; Noh, J.H. Iodide management in formamidinium-lead-halide—Based perovskite layers for efficient solar cells. Science 2017, 356, 1376–1379. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Lin, R.; Xiao, K.; Qin, Z.; Han, Q.; Zhang, C.; Wei, M.; Saidaminov, M.I.; Gao, Y.; Xu, J.; Xiao, M. Monolithic all-perovskite tandem solar cells with 24.8% efficiency exploiting comproportionation to suppress Sn (ii) oxidation in precursor ink. Nat. Energy 2019, 4, 864–873. [Google Scholar] [CrossRef]
  16. Shi, Z.; Jayatissa, A.H. Perovskites-based solar cells: A review of recent progress, materials and processing methods. Materials 2018, 11, 729. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Ceder, G.; Morgan, D.; Fischer, C.; Tibbetts, K.; Curtarolo, S. Data-mining-driven quantum mechanics for the prediction of structure. MRS Bull. 2006, 31, 981–985. [Google Scholar] [CrossRef] [Green Version]
  18. Michie, D.; Spiegelhalter, D.J.; Taylor, C. Machine learning. Neural Stat. Classif. 1994, 13, 19–22. [Google Scholar]
  19. Rajan, K. Materials informatics: The materials “gene” and big data. Annu. Rev. Mater. Res. 2015, 45, 153–169. [Google Scholar] [CrossRef] [Green Version]
  20. De Luna, P.; Wei, J.; Bengio, Y.; Aspuru-Guzik, A.; Sargent, E. Use machine learning to find energy materials. Nature 2017, 552, 23–27. [Google Scholar] [CrossRef]
  21. Ferguson, A.L. Machine learning and data science in soft materials engineering. J. Phys. Condens. Matter 2017, 30, 043002. [Google Scholar] [CrossRef] [PubMed]
  22. Mannodi-Kanakkithodi, A.; Huan, T.D.; Ramprasad, R. Mining materials design rules from data: The example of polymer dielectrics. Chem. Mater. 2017, 29, 9001–9010. [Google Scholar] [CrossRef]
  23. Saad, Y.; Gao, D.; Ngo, T.; Bobbitt, S.; Chelikowsky, J.R.; Andreoni, W. Data mining for materials: Computational experiments with A B compounds. Phys. Rev. B 2012, 85. [Google Scholar] [CrossRef] [Green Version]
  24. Seko, A.; Togo, A.; Hayashi, H.; Tsuda, K.; Chaput, L.; Tanaka, I. Prediction of low-thermal-conductivity compounds with first-principles anharmonic lattice-dynamics calculations and Bayesian optimization. Phys. Rev. Lett. 2015, 115. [Google Scholar] [CrossRef]
  25. Ghiringhelli, L.M.; Vybiral, J.; Levchenko, S.V.; Draxl, C.; Scheffler, M. Big data of materials science: Critical role of the descriptor. Phys. Rev. Lett. 2015, 114. [Google Scholar] [CrossRef] [Green Version]
  26. Cubuk, E.D.; Sendek, A.D.; Reed, E.J. Screening billions of candidates for solid lithium-ion conductors: A transfer learning approach for small data. J. Chem. Phys. 2019, 150, 214701. [Google Scholar] [CrossRef] [Green Version]
  27. Emery, A.A.; Wolverton, C. High-throughput dft calculations of formation energy, stability and oxygen vacancy formation energy of abo 3 perovskites. Sci. Data 2017, 4. [Google Scholar] [CrossRef] [Green Version]
  28. Schmidt, J.; Shi, J.; Borlido, P.; Chen, L.; Botti, S.; Marques, M.A. Predicting the thermodynamic stability of solids combining density functional theory and machine learning. Chem. Mater. 2017, 29, 5090–5103. [Google Scholar] [CrossRef]
  29. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2009, 22, 1345–1359. [Google Scholar] [CrossRef]
  30. Ong, S.P.; Richards, W.D.; Jain, A.; Hautier, G.; Kocher, M.; Cholia, S.; Gunter, D.; Chevrier, V.L.; Persson, K.A.; Ceder, G. Python Materials Genomics (pymatgen): A robust, open-source python library for materials analysis. Comput. Mater. Sci. 2013, 68, 314–319. [Google Scholar] [CrossRef] [Green Version]
  31. Ward, L.; Dunn, A.; Faghaninia, A.; Zimmermann, N.E.; Bajaj, S.; Wang, Q.; Montoya, J.; Chen, J.; Bystrom, K.; Dylla, M. Matminer: An open source toolkit for materials data mining. Comput. Mater. Sci. 2018, 152, 60–69. [Google Scholar] [CrossRef] [Green Version]
  32. Ward, L.; Agrawal, A.; Choudhary, A.; Wolverton, C. A general-purpose machine learning framework for predicting properties of inorganic materials. npj Comput. Mater. 2016, 2. [Google Scholar] [CrossRef] [Green Version]
  33. Ye, W.; Chen, C.; Wang, Z.; Chu, I.-H.; Ong, S.P. Deep neural networks for accurate predictions of crystal stability. Nat. Commun. 2018, 9, 3800. [Google Scholar] [CrossRef] [PubMed]
  34. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  35. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  36. Zeiler, M.D.; Fergus, R. Stochastic pooling for regularization of deep convolutional neural networks. arXiv 2013, arXiv:1301.3557. [Google Scholar]
  37. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  38. Bartel, C.J.; Sutton, C.; Goldsmith, B.R.; Ouyang, R.; Musgrave, C.B.; Ghiringhelli, L.M.; Scheffler, M. New tolerance factor to predict the stability of perovskite oxides and halides. Sci. Adv. 2019, 5. [Google Scholar] [CrossRef] [Green Version]
  39. Im, J.; Lee, S.; Ko, T.-W.; Kim, H.W.; Hyon, Y.; Chang, H. Identifying Pb-free perovskites for solar cells by machine learning. npj Comput. Mater. 2019, 5. [Google Scholar] [CrossRef] [Green Version]
  40. Jha, D.; Ward, L.; Paul, A.; Liao, W.-K.; Choudhary, A.; Wolverton, C.; Agrawal, A. Elemnet: Deep learning the chemistry of materials from only elemental composition. Sci. Rep. 2018, 8, 1–13. [Google Scholar] [CrossRef]
  41. Shin, S.S.; Yeom, E.J.; Yang, W.S.; Hur, S.; Kim, M.G.; Im, J.; Seo, J.; Noh, J.H.; Seok, S.I. Colloidally prepared La-doped BaSnO3 electrodes for efficient, photostable perovskite solar cells. Science 2017, 356, 167–171. [Google Scholar] [CrossRef]
  42. Mehmood, S.; Ali, Z.; Khan, I.; Khan, F.; Ahmad, I. First-Principles Study of Perovskite Molybdates AMoO 3 (A= Ca, Sr, Ba). J. Electron. Mater. 2019, 48, 1730–1739. [Google Scholar]
  43. Swarnkar, A.; Marshall, A.R.; Sanehira, E.M.; Chernomordik, B.D.; Moore, D.T.; Christians, J.A.; Chakrabarti, T.; Luther, J.M. Quantum dot–induced phase stabilization of α-CsPbI3 perovskite for high-efficiency photovoltaics. Science 2016, 354, 92–95. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Chang, Y.; Park, C.H.; Matsuishi, K. First-principles study of the Structural and the electronic properties of the lead-Halide-based inorganic-organic perovskites (CH~ 3NH~ 3) PbX~ 3 and CsPbX~ 3 (X= Cl, Br, I). J.-Korean Phys. Soc. 2004, 44, 889–893. [Google Scholar]
  45. Balena, A.; Perulli, A.; Fernandez, M.; De Giorgi, M.L.; Nedelcu, G.; Kovalenko, M.V.; Anni, M. Temperature Dependence of the Amplified Spontaneous Emission from CsPbBr3 Nanocrystal Thin Films. J. Phys. Chem. C 2018, 122, 5813–5819. [Google Scholar] [CrossRef]
  46. Swarnkar, A.; Chulliyil, R.; Ravi, V.K.; Irfanullah, M.; Chowdhury, A.; Nag, A. Colloidal CsPbBr3 perovskite nanocrystals: Luminescence beyond traditional quantum dots. Angew. Chem. Int. Ed. 2015, 54, 15424–15428. [Google Scholar] [CrossRef]
Figure 1. (a) is a list of elements considered by the A and B sites. The color coding of the element is a function of the amount of stabilized perovskite and the corresponding element of the A and B positions; (b) is a perovskite structure with 12-fold coordinated A cation (green A atom) and the octahedral coordinated B cation (blue B atom). The X atoms in the three-dimensional (3D) Wyckoff positions are in red.
Figure 1. (a) is a list of elements considered by the A and B sites. The color coding of the element is a function of the amount of stabilized perovskite and the corresponding element of the A and B positions; (b) is a perovskite structure with 12-fold coordinated A cation (green A atom) and the octahedral coordinated B cation (blue B atom). The X atoms in the three-dimensional (3D) Wyckoff positions are in red.
Applsci 09 05510 g001
Figure 2. Framework for the computational screening of perovskite materials. Abbreviations: RFR, random forest regression; SVR, support vector regression; GBR, gradient boosting regressor; Crystallographic Information File (CIF).
Figure 2. Framework for the computational screening of perovskite materials. Abbreviations: RFR, random forest regression; SVR, support vector regression; GBR, gradient boosting regressor; Crystallographic Information File (CIF).
Applsci 09 05510 g002
Figure 3. Pool processing matrix.
Figure 3. Pool processing matrix.
Applsci 09 05510 g003
Figure 4. Convolutional neural network architecture for material formation energy prediction.
Figure 4. Convolutional neural network architecture for material formation energy prediction.
Applsci 09 05510 g004
Figure 5. Feature importance from random forest (RF) for the formation energy of perovskite.
Figure 5. Feature importance from random forest (RF) for the formation energy of perovskite.
Applsci 09 05510 g005
Figure 6. Prediction performances of different models using hybrid features: (a) GBR; (b) least absolute shrinkage and selection operator (Lasso); (c) RFR; and (d) SVR.
Figure 6. Prediction performances of different models using hybrid features: (a) GBR; (b) least absolute shrinkage and selection operator (Lasso); (c) RFR; and (d) SVR.
Applsci 09 05510 g006
Figure 7. Prediction performance comparison of transfer learning model and the baseline model, both using Magpie descriptors. Left: transfer learning model trained with 578 samples labeled by M1 model +570 samples with density functional theory (DFT) formation energies. Right: baseline model trained only with 570 samples with DFT formation energies.
Figure 7. Prediction performance comparison of transfer learning model and the baseline model, both using Magpie descriptors. Left: transfer learning model trained with 578 samples labeled by M1 model +570 samples with density functional theory (DFT) formation energies. Right: baseline model trained only with 570 samples with DFT formation energies.
Applsci 09 05510 g007
Figure 8. Formation energy distribution of hypothetical perovskite materials predicted by M2 model.
Figure 8. Formation energy distribution of hypothetical perovskite materials predicted by M2 model.
Applsci 09 05510 g008aApplsci 09 05510 g008b
Table 1. Thirty-one materials features, their descriptions, and descriptions of attributes for D1 dataset CSV file.
Table 1. Thirty-one materials features, their descriptions, and descriptions of attributes for D1 dataset CSV file.
NameTypeUnitDescription
Chemical formulastringNoneChemical composition of the compound. The first and second elements correspond to the A- and B-site, respectively. The third element is oxygen
Hull distancenumbereV/atomHull distance as calculated by the equation of the distortion with the lowest energy. A compound is considered stable if it is within 0.025 eV per atom of the convex hull
BandgapnumbereVPBE band gap obtained from the relaxed structure
X_AnumbereVThe electronegativity of the A atom in the compound
X_BnumbereVThe electronegativity of the B atom in the compound
X_OnumbereVThe electronegativity of the O atom in the compound
A_atomic_massnumberamuAtomic mass of atom A
B_atomic_massnumberamuAtomic mass of atom B
O_atomic_massnumberamuAtomic mass of atom O
A_average_ionic_radiusnumberangThe average is taken over all oxidation states of the A element for which data is present
B_average_ionic_radiusnumberangThe average is taken over all oxidation states of the B element for which data is present
O_average_ionic_radiusnumberangThe average is taken over all oxidation states of the O element for which data is present
minDistAnumberNoneAtomic distance between the A cation and the nearest oxygen atom
minDistBnumberNoneAtomic distance between the B cation and the nearest oxygen atom
A_molar_volumenumbermolmolar volume of A element
B_molar_volumenumbermolmolar volume of B element
O_molar_volumenumbermolmolar volume of O element
A_electrical_resistivitynumberohmmelectrical resistivity of A element
B_electrical_resistivitynumberohmmelectrical resistivity of B element
O_electrical_resistivitynumberohmmelectrical resistivity of O element
A_atomic_radiusnumberangAtomic radius of element A
B_atomic_radiusnumberangAtomic radius of element B
O_atomic_radiusnumberangAtomic radius of element O
LUMOnumbereVThe lowest unoccupied molecular orbital
HOMOnumbereVThe highest occupied molecular orbital
volumenumberÅ3Volumes of crystal structures.
anumberÅLattice parameter a of the relaxed structure
bnumberÅLattice parameter b of the relaxed structure
cnumberÅLattice parameter c of the relaxed structure
alphanumber°α angle of the relaxed structure. α = 90 for the cubic, tetragonal, and orthorhombic distortion
betanumber°β angle of the relaxed structure. β = 90 for the cubic, tetragonal, and orthorhombic distortion
gammanumber°γ angle of the relaxed structure. γ = 90 for the cubic, tetragonal, and orthorhombic distortion
° degree, the unit of angle.
Table 2. Comparison of three feature sets for formation energy prediction.
Table 2. Comparison of three feature sets for formation energy prediction.
DescriptorsRMSEMAER2
Ong_Descriptors0.150.290.78
Magpie0.110.250.83
Hybrid_descriptors0.080.200.88
RMSE: root mean square error; MAE, mean absolute error.
Table 3. RMSE (eV/atom), MAE(eV/atom), and R2 values of 10-fold cross-validation results of all prediction models using the hybrid descriptor.
Table 3. RMSE (eV/atom), MAE(eV/atom), and R2 values of 10-fold cross-validation results of all prediction models using the hybrid descriptor.
Regression ModelRMSEMAER2
SVR0.830.670.20
RFR0.400.290.80
Lasso0.430.330.75
GBR0.280.200.91
Table 4. RMSE (eV/atom), MAE (eV/atom), and R2 values of cross-validation results of all prediction models using the Magpie descriptor. CNN, convolutional neural network.
Table 4. RMSE (eV/atom), MAE (eV/atom), and R2 values of cross-validation results of all prediction models using the Magpie descriptor. CNN, convolutional neural network.
ModelRMSEMAER2
ElemNet0.490.640.52
RF0.360.270.83
GBR0.350.250.84
CNN0.340.240.85
Table 5. Ninety-eight ABO3 perovskites calculated by DFT in the literature and our calculated τ value (stability less than 0.025, indicating that ABO3 perovskite is stable).
Table 5. Ninety-eight ABO3 perovskites calculated by DFT in the literature and our calculated τ value (stability less than 0.025, indicating that ABO3 perovskite is stable).
FormulaτStabilityFormulaτStabilityFormulaτStability
NiPtO3−58.1089−0.729RbPuO33.4805−0.103CeRhO31.3071−0.061
KPaO34.0629−0.567EuMoO33.2674−0.101SmVO32.9774−0.057
RbPaO33.6251−0.543KPuO33.78−0.1CsNpO33.3106−0.056
LiPaO3−5.7739−0.311SrTcO33.8734−0.097SmAlO31.9819−0.054
CsPaO33.3524−0.298BaZrO33.7699−0.096EuNbO34.09−0.053
BaHfO33.7284−0.247EuVO33.3069−0.093BaSnO33.6543−0.052
KNbO33.5401−0.221AcTiO33.8113−0.092NpTiO33.6701−0.051
EuGeO33.1904−0.212CeAlO3−0.5164−0.086KTcO33.555−0.048
EuTcO32.8001−0.15EuAsO32.0406−0.081LiOsO3−2.1229−0.044
RbNpO33.5353−0.15CeMnO3−0.9994−0.079CeGaO31.8339−0.043
EuOsO3−65.0247−0.143AcCuO33.2679−0.077EuIrO33.1159−0.043
KNpO33.8902−0.142AcNiO32.3005−0.077GdAlO33.1334−0.041
NpAlO3−20.7485−0.129LaVO33.64−0.077EuCoO33.1653−0.038
EuRhO32.8422−0.128PrVO33.2155−0.077CeNiO31.144−0.037
AcPdO33.7105−0.122EuAlO32.1447−0.075YbSiO31.9755−0.032
AcMnO31.718−0.12AcGaO32.4961−0.073BaTiO33.7351−0.03
AcFeO33.8271−0.116LaAlO32.3073−0.072DyAlO32.6055−0.03
AcVO32.6908−0.116EuPtO33.7879−0.071LaGaO33.3386−0.028
EuRuO32.0471−0.112NdVO32.555−0.071PuGaO3−10.427−0.028
AcAlO31.8388−0.11EuReO32.472−0.067YAlO33.5597−0.023
ErAlO33.699−0.022SrRhO33.8864−0.006KReO33.5840.008
LaCoO33.4776−0.022BaPdO33.7136−0.005YbReO33.64510.009
NaOsO3−11.1333−0.021KWO33.5403−0.004CsUO33.30370.01
PuNiO3−13.1563−0.02BaFeO33.73850EuGaO33.04390.013
LaNiO33.0358−0.019EuNiO32.77950SrCoO33.98920.014
TmAlO32.926−0.017LaMnO32.10320CeTmO3−300.8720.015
SrRuO33.6809−0.013EuSiO31.68490.001DyGaO33.88820.017
UAlO3−11.2029−0.011NdMnO31.66220.001BaPtO33.57780.019
YbAlO33.0372−0.011SmMnO31.83540.001DyCoO34.06170.02
SmCoO32.8557−0.009TbMnO30.54640.002YMnO33.12680.023
SmNiO32.5241−0.009EuMnO31.96940.003DyMnO32.34840.024
LuAlO34.1571−0.008SmGaO32.75140.006SrIrO33.97320.025
PuVO3−7.538−0.008NaReO34.10880.007
Table 6. Top 200 predicted perovskites with predicted τ values (tolerance factor).
Table 6. Top 200 predicted perovskites with predicted τ values (tolerance factor).
FormulaτFormulaτFormulaτFormulaτ
YbTlBr3−20.162CaTlCl3−14.381CaTlI3−14.0108PrBO3−2.0396
CaTlBr3−14.3041NaHgCl3−13.5492TlSnI32.7082LiNdO3−2.0303
CrCoBr31.55MgTlCl3−6.3625TlFeI32.9924LiTeO3−1.945
TlGeBr31.6125TlNiCl31.2386CsCrI33.2934TaBeO3−1.754
TlFeBr32.5339TlCoCl31.4647CsInI33.2934ThBeO3−1.7537
CsFeBr32.8551TlGeCl31.4794CsMgI33.3089CeBeO3−1.7537
CsScBr32.8556TlVCl31.5469CsTiI33.3142CrRhO3−1.747
CsPdBr32.8561TlCuCl31.9923CsSnI33.3312PuPtO3−1.6802
CsPtBr32.8713CsPdCl32.735CsGeI33.404ZrBO3−1.6005
CsInBr32.8777CsSnCl32.7368RbTiI33.4157PrBeO3−1.5988
CsGeBr32.8964CsCuCl32.739RbSnI33.4195HfBO3−1.5526
CsNiBr32.9272CsCoCl32.7645CsYbI33.4316TbSiO3−1.4989
RbCuBr32.9466CsCrCl32.7686CsTmI33.4519SnBO3−1.4453
RbSnBr32.9486CsInCl32.7686RbVI33.4526CuBO3−1.3853
RbVBr32.9499BaNiCl32.8079RbCrI33.4599HgRhO3−1.2912
RbGeBr32.9525RbGeCl32.8194RbInI33.4599GeBO3−1.0071
RbCoBr32.9532RbCuCl32.8216RbCrI33.4599PaCoO3−1.0055
RbPdBr32.9543RbSnCl32.8251RbGeI33.4601LiLaO3−2.8329
RbTiBr32.9568RbPdCl32.8331CsPbI33.5094HgIrO3−0.9862
CsAuBr33.0427RbFeCl32.8369CsDyI33.5292HgGeO3−0.9036
RbInBr33.0441RbMgCl32.8421CsCaI33.5514CeAsO3−0.8
KGeBr33.0488RbScCl32.8635KSnI33.5609TbBeO3−0.7902
CsYbBr33.0711KCoCl32.9149KTiI33.5761MnSiO3−0.7002
CsAgBr33.0767KGeCl32.9156CsMnI33.6832CoSiO3−2.4934
CsCdBr33.0839KVCl32.9194KInI33.7193BiMoO3−0.2924
ZrSiO30.0224EuMnO31.9694GdBO32.2335TmMnO32.6111
ThReO30.3455ThSiO3−2.1251GdBeO32.2465PmRuO32.6249
NpTaO3−4.4473NpSnO3−2.2195HoBO32.2479BeAgO3−4.6185
CuSiO30.8738NdReO31.986YBO32.2485TlSbO32.6921
BiRhO3−2.1807GdSiO32.0075EuBO32.2486SmGaO32.7514
LiSmO3−2.3252TlGaO32.0255ErBO32.2548InSiO32.7544
TlSiO31.5291LiEuO3−2.5704SmBO32.2647LiCdO3−4.1549
NdBeO31.5459BeAuO3−4.675LuBO32.2789BePbO3−4.5119
AcSiO31.5949PmBeO32.0662AcBO32.285AgAsO32.815
ZrBeO31.6089TlWO32.0764VSiO3-2.9169AgRuO32.8273
SmSiO31.6357ThWO32.0875LiPmO3-3.731LiGdO3−4.5398
SmBeO31.6514TlCoO32.0895LuSiO32.3438SmCoO32.8557
LiYbO3−4.301TlGeO32.1027TlBO32.3453SmGeO32.8773
BeHgO3−4.2571BeOsO3−4.2343DyMnO32.3484GdAsO32.9255
BeCdO3−4.6077TmBeO32.1393CrGeO32.3535BeInO3−5.3548
EuBeO31.7345HoSiO32.1444ThMoO32.423TlFeO32.9768
CrWO31.7353LiCaO3−3.2345HoBeO32.4577SmVO32.9774
LaBeO31.8183BiPtO32.145LiTmO3−4.0403LiPrO3−5.0002
CrTcO3−2.2119YSiO32.1489NdGeO32.4752YbAlO33.0372
LiCeO3−4.404LaAsO32.1864LiDyO3−3.369EuGaO33.0439
BeBiO3−4.8111YbBeO32.1968SmNiO32.5241GaBO33.0837
TiBeO31.8655DyBO32.2289ErBeO32.535NdTaO33.0888
TlNiO31.8873TmBO32.2293FeAsO3−3.8096HgRuO3−2.1943
TlTcO31.898AgBO32.2307TlCuO32.5776HoMnO33.1159
HfBeO31.9189YbBO32.2312LiAcO3−2.1229ThBO3−2.0482
Note: the ABO3 in this table does not contain the 98 ABO3 in Table 5. Please check supplementary file for the total 765 predictions.

Share and Cite

MDPI and ACS Style

Li, X.; Dan, Y.; Dong, R.; Cao, Z.; Niu, C.; Song, Y.; Li, S.; Hu, J. Computational Screening of New Perovskite Materials Using Transfer Learning and Deep Learning. Appl. Sci. 2019, 9, 5510. https://doi.org/10.3390/app9245510

AMA Style

Li X, Dan Y, Dong R, Cao Z, Niu C, Song Y, Li S, Hu J. Computational Screening of New Perovskite Materials Using Transfer Learning and Deep Learning. Applied Sciences. 2019; 9(24):5510. https://doi.org/10.3390/app9245510

Chicago/Turabian Style

Li, Xiang, Yabo Dan, Rongzhi Dong, Zhuo Cao, Chengcheng Niu, Yuqi Song, Shaobo Li, and Jianjun Hu. 2019. "Computational Screening of New Perovskite Materials Using Transfer Learning and Deep Learning" Applied Sciences 9, no. 24: 5510. https://doi.org/10.3390/app9245510

APA Style

Li, X., Dan, Y., Dong, R., Cao, Z., Niu, C., Song, Y., Li, S., & Hu, J. (2019). Computational Screening of New Perovskite Materials Using Transfer Learning and Deep Learning. Applied Sciences, 9(24), 5510. https://doi.org/10.3390/app9245510

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop