Abstract
Periodic material or crystal property prediction using machine learning has grown popular in recent years as it provides a computationally efficient replacement for classical simulation methods. A crucial first step for any of these algorithms is the representation used for a periodic crystal. While similar objects like molecules and proteins have a finite number of atoms and their representation can be built based upon a finite point cloud interpretation, periodic crystals are unbounded in size, making their representation more challenging. In the present work, we adapt the Pointwise Distance Distribution (PDD), a continuous and generically complete isometry invariant for periodic point sets, as a representation for our learning algorithm. The PDD distinguished all (more than 660 thousand) periodic crystals in the Cambridge Structural Database as purely periodic sets of points without atomic types. We develop a transformer model with a modified self-attention mechanism that combines PDD with compositional information via a spatial encoding method. This model is tested on the crystals of the Materials Project and Jarvis-DFT databases and shown to produce accuracy on par with state-of-the-art methods while being several times faster in both training and prediction time.
Similar content being viewed by others
Introduction
A solid crystalline material is made up of a periodically repeated unit cell containing a motif of atoms (ions or molecules). Crystals can distinguish themselves by atomic types (chemical elements and possibly charges of ions) and by the geometry of atomic centers. Both of these aspects can determine the various properties of a crystal. Knowledge of these properties is pertinent for determining whether a crystal can be experimentally synthesized or is useful for a particular application.
Determination of property values can be done using ab initio calculations with techniques like density functional theory (DFT)1. These techniques are often computationally expensive2. Further, they require extensive domain knowledge to be applied correctly, making them inaccessible. Recently, machine learning has become very popular as a substitute and has experienced success in decreasing computational costs while producing accurate predictions.
Any learning algorithm requires an input representation that adequately describes the object of interest. Objects similar to crystals, like molecules, are often treated as finite point clouds. This makes their representation more easily constructible than a representation for crystals, which are not bounded in size.
While a crystal can be described in several ways, descriptors that are easily human-interpretable, such as unit cell parameters or atomic coordinates are not useful for machine learning algorithms. Atomic coordinates do not retain invariance under rigid motion. Unit cell based descriptors are also ambiguous as there are infinitely many valid unit cells for a single structure. Such ambiguities can result in different model outputs for the same structure. Techniques such as data augmentation3 and parameter sharing4 can mitigate these effects but still do not guarantee the aforementioned consistency.
The structure-property relationship5 dictates that changes in the structure of a material result in changes in its properties. Distinction between crystals then allows for distinction between their respective property values. A machine learning algorithm (for a regression task) is a map from a crystal representation to value. If a representation cannot distinguish periodic crystals then two different crystals can incorrectly be perceived to be the same and so will the output property values. Similarly, if the same crystal can be represented in different ways, consistent mapping cannot be guaranteed.
The fundamental model of a crystal is a periodic set of points at all atomic centers (even without atomic types), see details in Definition 2.1. Since crystal structures are determined in a rigid form, their strongest practical equivalence is rigid motion, which is a composition of translations and rotations in \(\mathbb {R}^n\). We consider a slightly weaker isometry, which is a composition of rigid motion and mirror reflections. Two periodic point sets S and Q are isometric if they are related by an isometry \(f:\mathbb {R}^n\rightarrow \mathbb {R}^n\), so \(f(S)=Q\). This work exclusively applies these concepts to dimension \(n=3\).
An isometry invariant I is a property (descriptor or function) that is preserved under any isometry. The values of I should be simpler than the initial periodic set, for example, a scalar, vector, or matrix. Not all invariants can be considered equally useful, however. Space groups, for example, reflect the symmetry of a given material but tiny perturbations of atomic coordinates can change the material’s space group. Hence the important question is to quantify how different crystals are. Figure 1 illustrates the relationship between geometric descriptors based on their properties. A practically useful invariant should satisfy the following conditions introduced by6
Problem 1
Find a function I on all periodic point sets in \(\mathbb {R}^n\) subject to the following conditions:
-
1a.
Invariance: If two periodic point sets, S and Q are isometric, then \(I(S) = I(Q)\).
-
1b.
Generic Completeness: If \(I(S) = I(Q)\), then the two periodic sets are isometric.
-
1c.
Computability: the invariant I, the metric d, and the reconstruction of S can be computed in polynomial time with respect to the size of the motif of any periodic point set S.
-
1d.
Reconstructability: any periodic set S can be fully reconstructed from its invariant I(S).
-
1e.
Metric: There exists a distance function d on the codomain I that satisfies the following: 1) \(d(I(S), I(Q)) = 0\) iff \(I(S) = I(Q)\). 2) \(d(I(S), I(Q)) = d(I(Q), I(S))\). 3) For any three periodic sets S, Q, and T, \(d(I(S), I(Q)) + d(I(Q), I(T)) \ge d(I(S), I(T))\)
-
1f.
Lipschitz continuity: If a periodic set Q is obtained by shifting points within periodic set S by at most \(\epsilon\), then the distance between the two periodic sets can be bound according to some distance function d such that \(d(I(S), I(Q)) \le C\epsilon\) for some fixed constant C.
In the present work, an isometry invariant called the Pointwise Distance Distribution (PDD), defined formally in Definition 2.2, which has properties 1a and 1c (see Theorem 5.1 of6), and satisfies 1b and 1d for any periodic set in the general position6 along with sufficiently large k, and inclusion of the lattice (see theorem 4.4 of6). Conditions 1e and 1f (see Theorem 4.3 of6) are satisfied under the Earth Mover’s Distance (EMD) between PDDs.
The contribution of this work is a Transformer model7 which utilizes the PDD to make predictions on the properties of materials in a highly efficient manner compared to state-of-the-art models. In doing this, the gap between unambiguous crystal descriptors and machine learning models is bridged. Use of such a representation produces results on par or better than graph-based models, despite the additional structuring of data that comes with edges and edge embeddings. This model is faster in both prediction and training speed compared to two state-of-the-art models. To prove the method’s robustness, the model is applied to the crystals of the Materials Project8 and Jarvis-DFT9. Further experimentation on material classification, hyper-parameter sensitivity testing, and prediction of crystal properties without compositional information is included in the Supplemental Materials.
Related work
Early works in crystal property prediction used more classical statistical methods like kernel regression10 before eventually moving towards deep learning11. More recent works have shifted to Graph Neural Networks (GNN)12,13,14,15,16,17,18,19,20,21 due to their ability to make use of structured data. Several of these focus on predicting the properties of the crystals contained within the Materials Project8 using a multigraph representation where vertices represent atoms and edges are embedded with the pairwise distances to an atom’s nearest neighbors. Some state-of-the-art models use line graphs to incorporate more geometric information like angles and dihedrals13,22. The derived line graphs can contain significantly more vertices and edges, incurring a higher computational cost. Other method23 take a physics principled approach and substitute the interatomic distances for interatomic potentials and capture a crystal’s periodicity using the infinite sum of these potentials.
While effective in modeling crystal structures, graphs are discontinuous under perturbations24. Small movements in the atomic positioning can cause significant changes to the graph’s topology. Some graphs are not unit cell invariant. Due to an infinite number of possible unit cells, the graph is then reliant on the data or the cell reduction technique used.
SE(3)-equivariant models such as Tensor Flow Networks25, SE(3)-Transformers26, and SE(3)-GNNs27 impose constraints on the set of learnable functions of the network such that the output is equivariant with respect to the input points. While effective on finite point clouds, they offer no promise of completeness with periodic point sets which is necessary for distinguishing between structures. These architectures can also be beneficial when predicting properties that are equivariant with respect to rigid motion, but the crystal properties examined here are invariant to such symmetries, and thus invariance of the model output through either the input or model architecture is required.
In addition to the properties mentioned earlier, the invariant needs to be able to be adapted for a learning algorithm. Further, it needs a way to incorporate compositional information as invariants typically only consider structure. Some invariants have been adapted for use in machine learning algorithms such as symmetry functions28,29 and Voronoi cells30. Both of these, however, still lack continuity. The Partial Radial Distribution Function is invariant and continuous but is not complete for homometric crystals. Smooth Overlapped Atomic Positions31 has been incorporated into models for property prediction, but is an invariant for atomic environments, not entire structures. Coulomb matrices use electrostatic interactions between atoms instead of Euclidean distances and are invariant and complete for molecules but have not been proven to retain these properties for the periodic case32. Average Minimum Distance (AMD)33 is invariant and continuous and has been used to predict lattice energies via Gaussian Process Regression34, but is incomplete and does not currently have a way to incorporate compositional information. The PDD has been used to derive a graph representation35, but this graph does not retain continuity.
Methods
A periodic crystal can be represented as a periodic point set36 with points located at the atomic centers of the structure. They do not differentiate between the types of atoms and instead treat every point as unlabeled. A periodic point set (periodic set) can defined like so:
Definition 2.1
(Periodic Point Set) For a set of n basis vectors \({\varvec{v}}_1 \ldots {\varvec{v}}_n \in \mathbb {R}^n\), the lattice L is formed by the integer linear combinations of these basis vectors \(\{ \sum _{i=1}^n c_i {\varvec{v}}_i \vert c_i \in \mathbb {Z} \}\). The unit cell is the parallelepiped \(U = \{ \sum _{i=1}^n t_i {\varvec{v}}_i | t_i \in [0, 1) \}\). For a unit cell U, the motif M is a finite subset of U. Then, a periodic point set S of lattice L and motif M is defined by \(\{ \varvec{\lambda } + {\varvec{p}}: \varvec{\lambda } \in L , {\varvec{p}}\in M \}\).
The PDD of a periodic set is the \(m \times (k+1)\) matrix where m is the number of atoms in the motif M and k is a positive integer indicating the number of nearest neighbors to use. Each row corresponds to a point in the motif and the entries within the row consist of the Euclidean distance to each of this point’s k-nearest neighbors within the entire periodic set S. The first entry of the row is assigned to be a weight equal to \(\frac{1}{m}\) (the distances follow). Once the matrix is formed, rows that are the same are collapsed into a single row and their respective weights are added. Due to very small differences between rows caused by floating point arithmetic or atomic perturbations, it is common to use a tolerance, henceforth called the collapse tolerance, that allows rows with small non-zero differences (e.g. with respect to \(L_\infty\) distance) to be treated as the same. By collapsing rows in the PDD, the resulting matrix representation is always the same for a given crystal, regardless of the unit cell. Formally,
Definition 2.2
(Pointwise Distance Distribution) For a periodic set \(S = L + M\) with a set of motif points \(M = \{{\varvec{p}}_1, \ldots , {\varvec{p}}_m \}\) within a unit cell U of lattice L, the uncollapsed PDD matrix for a parameter \(k \in \mathbb {N}^+\) is a \(m \times (k+1)\) matrix where the \(i^{th}\) row consists of the row weight \(w_i = \frac{1}{m}\) followed by the euclidean distances \(d_1 \ldots d_k\) from the point \({\varvec{p}}_i\) to its k-nearest neighbors such that \(d_1 \le d_2 \ldots \le d_k\). If a group of rows is found to be identical (or close enough using a valid distance measure within some tolerance) then the matrix rows are collapsed and the weights of the involved rows are summed. The resulting matrix will then have less than m rows.
This matrix is referred to as \(\textsc {PDD}(S;k)\) for a periodic set S and positive integer k.
Periodic set transformer
In our model, rather than being considered a matrix of values, the PDD will be considered a set of grouped atoms. A single group of atoms corresponds to the k-nearest neighbor distances in a given row within the PDD matrix. Each member of the set will carry the weight provided by the row in the PDD. Any set A can trivially be turned into a weighted set by weighing each element by \(\frac{1}{|A|}\). When the PDD is not collapsed, then there can be more than a single occurrence of any given element, making the uncollapsed PDD a multiset. Now, let A be a multiset of the form \(A = \{ a_{i}^{(j)}: i \in [1, \dots , n], j\in [1,\dots , n_i] \}\) where \(n_i\) is the multiplicity of element \(a_i\) and \(a_{i}^{(j)}\) is the \(j^{th}\) occurrence of element \(a_i\). This multiset can be turned into a weighted set by assigning each element \(a_i\) with the weight \(\frac{n_i}{n}\). We can recover the influence of multiplicity by the use of weights in our model.
When a periodic crystal has its unit cell modified, the proportion of each atom is expanded or reduced. The use of weights captures this behavior in the form of a concentration or frequency.
We use an attention mechanism to find the interactions between members of the set. The rows of the PDD contain the pairwise distance information, but they do not indicate which atoms these distances correspond to. Application of the attention mechanism can help the model learn these interactions.
Let \({\varvec{R}}\in \mathbb {R}^{m \times k}\) be the PDD matrix containing m rows without the associated weight column. Let \({\varvec{w}}\in \mathbb {R}^{m \times 1}\) be the column vector containing the weights from the PDD matrix. The initial embedding is \({\varvec{X}}^{(0)} = {\varvec{R}}{\varvec{W}}_d\) where \({\varvec{W}}_d \in \mathbb {R}^{k \times d}\) is the initial trainable weight matrix. The embedding is updated according to:
where \({\varvec{Q}}= {\varvec{X}}^{(0)} {\varvec{W}}_Q\), \({\varvec{K}}= {\varvec{X}}_i^{(0)} {\varvec{W}}_K\) and \({\varvec{V}}= {\varvec{X}}^{(0)} {\varvec{W}}_{V}\) and d is the embedding dimension of the weight matrices for the query, key, and value \({\varvec{W}}_Q, {\varvec{W}}_K,\) and \({\varvec{W}}_V\) respectively as described in7. The function \(\sigma\) is the softmax function with the PDD weights integrated into it; \(\sigma\) is applied to each row \({\varvec{z}}\) of the input matrix, and i and j are used to index entries in \({\varvec{z}}\) and \({\varvec{w}}\). The \(i^{th}\) entry of the output vector is defined by:
The result is passed through a single-layer perceptron SLP. The layer normalization order described by37 is used for increased stability during training. Equation (2) describes the case for a single attention head. When multiple attention heads are used, the PDD weights are applied to each individually and the result is concatenated before being passed to the SLP like so:
where h is the number of attention heads, \(\oplus\) is the concatenation operator and \({\varvec{Q}}_i, {\varvec{K}}_i\) and \({\varvec{V}}_i\) are the query, key and value for the \(i^{th}\) head. This process is repeated l times; this determines the depth of the model. The embeddings are finally pooled into a single vector by reincorporating the PDD weights into a weighted sum of the row vectors \({\varvec{x}}_i\) of the final embedding \({\varvec{X}}^{(l)}\).
This final embedding can be passed to a perceptron layer to predict the property value.
This version of self-attention can be applied to a weighted set or distribution. The weights are applied in such a way that the output of the PST is invariant to an arbitrary splitting of rows within the PDD. We provide a formal proof of this in Supplementary Material. An overview of the Periodic Set Transformer (PST) architecture with PDD encoding (described in the next section) can be seen in Fig. 2.
PDD encoding
While structure is a powerful indicator of a crystal’s properties, there may be datasets in which it is not the primary differentiator of a set of crystals. In such cases, the composition of the atoms contained within the material has a heavy influence. The previously described transformer does a good job of utilizing the structural information within the PDD but does not provide an obvious way to include atomic composition.
Transformers for natural language processing tasks use positional encoding to allow the model to distinguish the position of words within a given sentence38. A recent transformer model, Uni-Mol39, which performed property prediction for molecules (among other tasks), used 3D spatial encoding first proposed by40 to give the model an understanding of each atom’s position in space, relative to one another. This encoding is done at the pair level, using the Euclidean distance between atoms and a pair-type aware Gaussian kernel41. A transformer model for finite 3D points clouds is provided by42 via vector attention. The case for crystals is more difficult because they are not bounded in size and can exhibit many symmetries. Fortunately, by using the rows of the PDD we can distinguish each atom with structural information. We refer to this as PDD encoding.
When rows are grouped together, they are done so by having the same k-nearest neighbor distances. Though rare, it is possible for rows corresponding to different atom types to be collapsed. If this occurs, the selection of either atom type will result in information loss. To prevent this, we add the condition that the groups must be formed on the basis of having the same k-nearest neighbor distances and the same atomic species. In this case, the periodic point set has points that are labeled according to atomic type.
For a periodic set S, let PDD(S; k) be the resulting PDD matrix with parameter k. Let \({\varvec{R}}\) be PDD(S; k) without the initial weight column and \({\varvec{T}}\) be the matrix whose rows correspond to the vector of atomic properties used to describe the type of atom associated with each row of PDD(S; k). The initial set of embeddings for the attention mechanism is defined as \({\varvec{X}}^{(0)} = {\varvec{R}}{\varvec{W}}_s + {\varvec{T}}{\varvec{W}}_c\) where \({\varvec{W}}_s\) and \({\varvec{W}}_c\) are initial embedding weights. By starting with a linear embedding, the PDD row can be transformed to match the dimension of composition embedding. The parameter k used can then be changed as needed to include distance information from further neighbors.
Results and discussion
Prediction of materials project properties
The model will be applied to the data within the Materials Project. To make fair comparisons to other models we report the performance according to Matbench43, which contains data for various crystal properties. The error rates are reported using five-fold cross-validation with standardized training and testing sets for each fold. Further, tuning is done according to the models’ authors and thus our model can be compared to others more fairly.
The crystals in the Materials Project are highly diverse in composition. For all predictions, we include the composition of the crystal with PDD encoding. To incorporate this compositional information the mat2vec atomic embeddings supplied by44 are used. The embeddings have empirically been found to produce better performance than the one-hot encoded method used by CGCNN45. They also have the added convenience of not missing any atomic property information for certain elements.
In Table 1 we report the average mean-absolute-error (MAE) across the five test sets. We include the reported accuracies of other models to allow for comparison. The selection of models aims to present a high diversity in approaches while also coming from relatively recent publications. CrabNet45 is the only other Transformer model listed on Matbench. This model, in terms of architecture, is the most similar to the PST. Additionally, the atomic embeddings used to describe each chemical element are the same as those used in our model. The majority of models used in crystal property prediction use GNNs. Matbench features several of these, but the model that provides the best results on several properties is coGN. coGN is a GNN that includes angular and dihedral information through the use of line graphs. As such, the amount of information used is significantly more than the PST, which uses the distribution of pairwise distances. Crystal Twins (CT)46 is a model based on the convolutional layer developed by CGCNN12 (also used by several other models16,18,47) that uses self-supervised learning to create embeddings based on maximizing the similarity between augmented instances of a crystal.
The PST and CrabNet share similarities in their construction. Both use mat2vec atomic embeddings and utilize a Transformer architecture with self-attention. While CrabNet uses fractional encoding to embed the multiplicity of each element type, we opt for the PDD-weighted attention mechanism and pooling described by Eqs. (2) and (4). The PST also uses PDD encoding to add structural information. This could also be done for CrabNet, but the combination of both fractional and PDD encoding is not guaranteed to aid in performance and simple summation of the encodings can cause ambiguities in the final embeddings, reducing performance. Across all properties the PST outperforms CrabNet, further indicating the usefulness of PDD encoding.
The performance disparity between the PST and coGN on formation and band gap energy can be difficult to discern. GNNs allow embeddings at the vertex and edge level. These embeddings can not only carry different information but allow for simultaneous updates to each of these embeddings, adding richness to the learned representation. CoGN takes this further and updates the original edges with message-passing from the derived line graph, allowing the inclusion of angular information. The edges of the line graph are further updated by its line graph, incorporating dihedral angles. These updates allow the model to learn a better representation of a crystal in latent space, which can be necessary for these larger datasets. This points to a current limitation for the PST. By using PDD encoding, we effectively limit opportunities for such updates to a single embedding representing both the atom’s properties and its structural behavior.
In Table 2 the time taken to train and test the PST and coGN are listed. Our model was trained for only 250 epochs while coGN is trained for 800 epochs. While this accounts for a large portion of the disparity, the time taken to train each model per epoch is also faster for our model in all properties except the refractive index. A batch size of 32 is used for datasets containing less than 5, 000 samples. For exfoliation energy and phonon peak, PST takes \(68.1\%\) and \(92.6\%\) of the time of coGN per epoch. The larger batch size of coGN allows it to have greater GPU utilization and thus, better training efficiency. For properties with greater than 5, 000 samples, the batch size for both models is the same. For these properties, the training time per epoch is between \(65-70\%\) of coGN.
The prediction time disparity is more pronounced. For band gap and formation energy, the PST makes predictions approximately five times faster than coGN. Using nested line graphs introduces significant computational cost, but coGN is able to shrink the size of the graph by using the atoms in the asymmetric unit. The unit cell based approach was initially proposed by CGCNN12 and used in several follow-up works13,16,18,46, but the unit cell is inherently ambiguous and unnecessarily large in terms of the number of atoms needed to fully describe a crystal’s symmetry. The PDD at a collapse tolerance of exactly zero will have a number of rows less than or equal to the number of atoms in the asymmetric unit. At higher tolerances, this number will be reduced until reaching the number of unique chemical elements in the crystal.
Ablation study
In Table 3 we list the results for each “component” within the Periodic Set Transformer. In the row indicating PDD as the component, we train and test the model using only the structural information within the PDD. In the “Composition” component we pass the atomic encoding for the elements in the crystal and their concentration in the form of the PDD weights to the model without the PDD encoding. The model is run using \(k = 15\) and a collapse tolerance of exactly zero.
By separating out each component of the model, we can interpret the importance of each to a particular property. Properties that experience a more significant decrease in performance when the PDD encoding is not used, can be ascribed to be more dependent on structural information. In all cases, the combination of both the composition and PDD encoding results in significantly lower error rates. We can conclude that this encoding method is effective in combining the structural and compositional information of a crystal structure.
In Table 4 the effect of including the PDD weight in the attention mechanism described in Eq. (2) and in the pooling layer described by Eq. (4) is listed. A collapse tolerance of zero is used to remove any regularization effect (further described in the Supplemental Material).
The exclusion of weights from both the attention mechanism and pooling decreases accuracy significantly. Doing this removes all indications of multiplicity making discernment of crystals more difficult. The inclusion of the weights in the pooling layer is more impactful than when applied in the attention mechanism. The use of the weights in the pooling layer alone allows the model to perform better when the number of samples in the dataset is low. Datasets with fewer samples likely have less diversity amongst their crystals, making the need for recognizing the multiplicity of atoms less necessary.
Prediction of Jarvis-DFT properties
The Jarvis-DFT dataset9 is a commonly used set of materials with VASP48 calculated properties. The list of properties computed for the materials within the dataset is more extensive than that of the Materials Project. Its inclusion provides further evidence of the robustness of the model on an even wider variety of crystal properties.
The prediction MAE produced by PST and Matformer for 12 different properties from the dataset are included in Table 5. For Matformer, we retrain the model to ensure the training and testing sets are the same. We use the default parameters for the model defined by the authors’ codebase. We make one alteration to the training procedure; the number of epochs trained is reduced to 250. The number of epochs is the same as for our model.
The PST outperforms Matformer in nine of the twelve properties tested. In particular, properties for which data is sparse yield results that favor the PST significantly (i.e. exfoliation energy, \(e_{ij}\) and \(d_{ij}\)). Jarvis-DFT has two band gap values that are computed for its crystals, one which uses the optimized Becke88 functional (OPT)49 and the other uses the Tran-Blaha modified Becke Johnson potential (MBJ)50. The latter is more accurate (when compared to experimentally observed values) but also more computationally expensive. For this reason, there are significantly fewer computed values in the database. Interestingly, the PST produces a smaller error for the more accurate band gap values compared to Matformer, but a larger error for the less accurate OPT calculated values. A possible reason for this is the smaller sample size for which the PST has shown to be more effective. The disparity in performance for formation and total energy can be attributed to Matformer’s architecture which uses a GNN that updated both node and edge embeddings. This additional level of expression is helpful particularly when the size of the data grows larger, though it does come with added computational cost.
Matformer has been shown to produce even better results than the previous state-of-the-art model ALIGNN13 while taking roughly a third of the time to do both training and prediction. In Table 6, the training and prediction time for each of the properties in the Jarvis-DFT dataset is reported for the PST and Matformer. For the training time, the validation and pre-processing times are not included. The prediction time listed is the number of seconds taken to make predictions on the test set.
In the closest training time comparison, the PST is still more than six times faster than Matformer. The training times for all properties fall between six and twelve times faster for the PST compared to Matformer. The performance increase can be attributed to several factors. Primarily, Matformer relies on a line graph (similar to coGN22) in order to update edge embeddings. While this increases the information used and leads to richer learned embeddings, the size of line graphs is considerably larger than the graph they are derived from. This, in turn, incurs a higher computational cost.
The difference in prediction times is more significant. Exfoliation energy is predicted over fifty times faster using the PST than with Matformer. This is the closest the two models perform to each other. Notably, exfoliation energy also has the fewest samples. For the bulk of the other properties, the speedup ranges between eighty and ninety times faster for the PST.
Conclusion
The PDD is a generically complete and continuous invariant under isometry and permutations of points, hence independent of a unit cell. By using weights and creating a distribution, the PDD is able to represent an infinitely spanning object by its finite forms of behavior. Further, by collapsing rows in the PDD, the resulting representation can also be much smaller in comparison to the number of atoms within the unit cell, even when the cell is reduced.
The model is applied to the crystals of the Materials Project and Jarvis-DFT on a variety of material properties. Despite using less information in the model than more commonly employed graph-based models, the PST is able to produce results on par or even exceeding that of models like coGN and Matformer while taking significantly less time to train and make predictions.
Data availability
The data from the Materials Project is automatically downloaded through the code in the Github Repository. The Jarvis-DFT data can be downloaded through the Jarvis-Tools python package using the dft_3d_2021 database. Examples are included in the documentation here: https://pages.nist.gov/jarvis/databases/. The dataset for the crystals used in the lattice energy experiments is available at https://eprints.soton.ac.uk/404749/.
Code availability
The code for the experiments is located at: https://github.com/jonathanBalasingham/Periodic-set-transformer. The code contains what is necessary to re-run the experiments done in “Prediction of Jarvis-DFT properties” and “Prediction of materials project properties” sections. It also contains the source code necessary to recreate the Figures included in the Supplemental Material. Details of how these plots are created are included in the Supplementary Material. The individual predictions for the Materials Project data are contained in JSON format within the repository. Proofs for the properties of the PDD mentioned in51, Problem 1.1 are included in the original paper6. More details on the actual implementation of the model, data pre-processing, and training are contained in Supplementary Material.
References
Sholl, D. S. & Steckel, J. A. Density Functional Theory: A Practical Introduction (Wiley, 2022).
Cohen, A. J., Mori-Sánchez, P. & Yang, W. Challenges for density functional theory. Chem. Rev. 112, 289–320 (2012).
Quiroga, F., Ronchetti, F., Lanzarini, L. & Bariviera, A. F. Revisiting data augmentation for rotational invariance in convolutional neural networks. In Modelling and Simulation in Management Sciences: Proceedings of the International Conference on Modelling and Simulation in Management Sciences (MS-18), 127–141 (Springer, 2020).
Ravanbakhsh, S., Schneider, J. & Póczos, B. Equivariance through parameter-sharing. In Precup, D. & Teh, Y. W. (eds.) Proceedings of the 34th International Conference on Machine Learning, vol. 70 of Proceedings of Machine Learning Research, 2892–2901 (PMLR, 2017).
Le, T., Epa, V. C., Burden, F. R. & Winkler, D. A. Quantitative structure-property relationship modeling of diverse materials properties. Chem. Rev. 112, 2889–2919. https://doi.org/10.1021/cr200066h (2012).
Widdowson, D. & Kurlin, V. Resolving the data ambiguity for periodic crystals. In Advances in Neural Information Processing Systems (Proceedings of NeurIPS 2022) 35, 24625–24638 (2022).
Vaswani, A. et al. Attention is all you need. Adv. Neural Inf. Process. Syst.30 (2017).
Jain, A. et al. The Materials Project: A materials genome approach to accelerating materials innovation. Appl. Phys. Lett. Mater. 1, 011002. https://doi.org/10.1063/1.4812323 (2013).
Choudhary, K. et al. The joint automated repository for various integrated simulations (jarvis) for data-driven materials design. NPJ Comput. Mater. 6, 173 (2020).
Calfa, B. A. & Kitchin, J. R. Property prediction of crystalline solids from composition and crystal structure. AIChE J. 62, 2605–2613. https://doi.org/10.1002/aic.15251 (2016).
Ye, W., Chen, C., Wang, Z., Chu, I.-H. & Ong, S. P. Deep neural networks for accurate predictions of crystal stability. Nat. Commun. 9, 3800–3800. https://doi.org/10.1038/s41467-018-06322-x (2018).
Xie, T. & Grossman, J. C. Crystal graph convolutional neural networks for an accurate and interpretable prediction of material properties. Phys. Rev. Lett. 120, 145301. https://doi.org/10.1103/PhysRevLett.120.145301 (2018).
Choudhary, K. & DeCost, B. Atomistic line graph neural network for improved materials property predictions. NPJ Comput. Mater.https://doi.org/10.1038/s41524-021-00650-1 (2021).
Yan, K., Liu, Y., Lin, Y. & Ji, S. Periodic graph transformers for crystal material property prediction. In Oh, A. H., Agarwal, A., Belgrave, D. & Cho, K. (eds.) Advances in Neural Information Processing Systems (2022).
Omee, S. S. et al. Scalable deeper graph neural networks for high-performance materials property prediction. Patterns 100491 (2022).
Park, C. W. & Wolverton, C. Developing an improved crystal graph convolutional neural network framework for accelerated materials discovery. Phys. Rev. Mater. 4, 063801 (2020).
Chen, C., Ye, W., Zuo, Y., Zheng, C. & Ong, S. P. Graph networks as a universal machine learning framework for molecules and crystals. Chem. Mater. 31, 3564–3572. https://doi.org/10.1021/acs.chemmater.9b01294 (2019).
Das, K. et al. CrysXPP: An explainable property predictor for crystalline materials. NPJ Comput. Mater. 8, 43. https://doi.org/10.1038/s41524-022-00716-8 (2022).
Cheng, J., Zhang, C. & Dong, L. A geometric-information-enhanced crystal graph network for predicting properties of materials. Commun. Mater. 2, 1–11 (2021).
Sanyal, S. et al. Mt-cgcnn: Integrating crystal graph convolutional neural network with multitask learning for material property prediction. arXiv preprint arXiv:1811.05660 (2018).
Schütt, K. T., Sauceda, H. E., Kindermans, P.-J., Tkatchenko, A. & Müller, K.-R. Schnet-a deep learning architecture for molecules and materials. J. Chem. Phys. 148, 241722 (2018).
Ruff, R., Reiser, P., Stühmer, J. & Friederich, P. Connectivity optimized nested graph networks for crystal structures (2023). arXiv:2302.14102.
Lin, Y. et al. Efficient approximations of complete interatomic potentials for crystal property prediction. In Krause, A. et al. (eds.) Proceedings of the 40th International Conference on Machine Learning, vol. 202 of Proceedings of Machine Learning Research, 21260–21287 (PMLR, 2023).
Edelsbrunner, H., Heiss, T., Kurlin, V., Smith, P. & Wintraecken, M. The density fingerprint of a periodic point set. In 37th International Symposium on Computational Geometry (SoCG 2021), 189, 395–408 (2021).
Thomas, N. et al. Tensor field networks: Rotation- and translation-equivariant neural networks for 3d point clouds. CoRRabs/1802.08219 (2018). arXiv:1802.08219.
Fuchs, F., Worrall, D., Fischer, V. & Welling, M. Se (3)-transformers: 3d roto-translation equivariant attention networks. Adv. Neural. Inf. Process. Syst. 33, 1970–1981 (2020).
Du, W. et al. Se (3) equivariant graph neural networks with complete local frames. In International Conference on Machine Learning, 5583–5608 (PMLR, 2022).
Behler, J. Atom-centered symmetry functions for constructing high-dimensional neural network potentials. J. Chem. Phys. 134, 074106 (2011).
Egorova, O., Hafizi, R., Woods, D. C. & Day, G. M. Multifidelity statistical machine learning for molecular crystal structure prediction. J. Phys. Chem. A 124, 8065–8078 (2020).
Ward, L. et al. Including crystal structure attributes in machine learning models of formation energies via voronoi tessellations. Phys. Rev. B 96, 024104 (2017).
Bartók, A. P., Kondor, R. & Csányi, G. On representing chemical environments. Phys. Rev. B 87, 184115 (2013).
Faber, F., Lindmaa, A., Von Lilienfeld, O. A. & Armiento, R. Crystal structure representations for machine learning models of formation energies. Int. J. Quantum Chem. 115, 1094–1101 (2015).
Widdowson, D., Mosca, M., Pulido, A., Cooper, A. & Kurlin, V. Average minimum distances of periodic point sets - fundamental invariants for mapping all periodic crystals. MATCH Commun. Math. Comput. Chem. 87, 529–559 (2022).
Ropers, J., Mosca, M. M., Anosova, O., Kurlin, V. & Cooper, A. I. Fast predictions of lattice energies by continuous isometry invariants of crystal structures. In Pozanenko, A., Stupnikov, S., Thalheim, B., Mendez, E. & Kiselyova, N. (eds.) Data Analytics and Management in Data Intensive Domains, 178–192 (Springer International Publishing, Cham, 2022).
Balasingham, J., Zamaraev, V. & Kurlin, V. Material property prediction using graphs based on generically complete isometry invariants. Integr. Mater. Manuf. Innov. https://doi.org/10.1007/s40192-024-00351-9 (2024).
Smith, P. & Kurlin, V. A practical algorithm for degree-k voronoi domains of three-dimensional periodic point sets. In Lecture Notes in Computer Science (Proceedings of ISVC), 13599, 377–391 (2022).
Xiong, R. et al. On layer normalization in the transformer architecture. In International Conference on Machine Learning, 10524–10533 (PMLR, 2020).
Dufter, P., Schmitt, M. & Schütze, H. Position information in transformers: An overview. Comput. Linguist. 48, 733–763 (2022).
Zhou, G. et al. Uni-mol: A universal 3d molecular representation learning framework. In The Eleventh International Conference on Learning Representations (2023).
Ying, C. et al. Do transformers really perform badly for graph representation?. Adv. Neural. Inf. Process. Syst. 34, 28877–28888 (2021).
Shuaibi, M. et al. Rotation invariant graph neural networks using spin convolutions. arXiv preprint arXiv:2106.09575 (2021).
Zhao, H., Jiang, L., Jia, J., Torr, P. H. & Koltun, V. Point transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 16259–16268 (2021).
Dunn, A., Wang, Q., Ganose, A., Dopp, D. & Jain, A. Benchmarking materials property prediction methods: The matbench test set and automatminer reference algorithm. NPJ Comput. Mater. 6, 138 (2020).
Tshitoyan, V. et al. Unsupervised word embeddings capture latent knowledge from materials science literature. Nature 571, 95–98 (2019).
Wang, A.Y.-T., Kauwe, S. K., Murdock, R. J. & Sparks, T. D. Compositionally restricted attention-based network for materials property predictions. NPJ Comput. Mater. 7, 77 (2021).
Magar, R., Wang, Y. & Barati Farimani, A. Crystal twins: Self-supervised learning for crystalline material property prediction. NPJ Comput. Mater. 8, 231. https://doi.org/10.1038/s41524-022-00921-5 (2022).
Cao, Z., Magar, R., Wang, Y., Farimani, B. & Moformer, A. Self-supervised transformer model for metal-organic framework property prediction. J. Am. Chem. Soc. 145, 2958–2967. https://doi.org/10.1021/jacs.2c11420 (2023).
Kresse, G. & Furthmüller, J. Efficiency of ab-initio total energy calculations for metals and semiconductors using a plane-wave basis set. Comput. Mater. Sci. 6, 15–50 (1996).
Klimeš, J., Bowler, D. R. & Michaelides, A. Chemical accuracy for the van der waals density functional. J. Phys.: Condens. Matter 22, 022201 (2009).
Tran, F. & Blaha, P. Accurate band gaps of semiconductors and insulators with a semilocal exchange-correlation potential. Phys. Rev. Lett. 102, 226401 (2009).
Widdowson, D. & Kurlin, V. Pointwise distance distributions of periodic sets. CoRRabs/2108.04798 (2021). arXiv:2108.04798.
Acknowledgements
We would like to thank both of the anonymous reviewers whose feedback helped improve the manuscript. This research was supported by the Royal Academy of Engineering Industry Fellowship (IF2122/186) at the Cambridge Crystallographic Data Centre, the New Horizons EPSRC grant (EP/X018474/1), and the Royal Society APEX fellowship (APX/R1/231152). The funder played no role in the study design, data collection, analysis and interpretation of data, or the writing of this manuscript.
Author information
Authors and Affiliations
Contributions
J.B. proposed the original model, V.K. proposed the spatial encoding method used. J.B. developed and tested the codebase used for the experiments. J.B., V.Z. and V.K. contributed in writing the first draft of the article, and subsequent revisions were completed by all three authors.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Balasingham, J., Zamaraev, V. & Kurlin, V. Accelerating material property prediction using generically complete isometry invariants. Sci Rep 14, 10132 (2024). https://doi.org/10.1038/s41598-024-59938-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-024-59938-z
This article is cited by
-
Generic families of finite metric spaces with identical or trivial 1-dimensional persistence
Journal of Applied and Computational Topology (2024)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.