Crystal structure prediction (CSP) is a foundational computational technique for determining the atomic arrangements of crystalline materials, especially under high-pressure conditions. While CSP plays a critical role in materials science, traditional approaches often encounter significant challenges related to computational efficiency and scalability, particularly when applied to complex systems. Recent advances in machine learning (ML) have shown tremendous promise in addressing these limitations, enabling the rapid and accurate prediction of crystal structures across a wide range of chemical compositions and external conditions. This review provides a concise overview of recent progress in ML-assisted CSP methodologies, with a particular focus on machine learning potentials and generative models. By critically analyzing these advances, we highlight the transformative impact of ML in accelerating materials discovery, enhancing computational efficiency, and broadening the applicability of CSP. Additionally, we discuss emerging opportunities and challenges in this rapidly evolving field.

Pressure, as a fundamental thermodynamic variable, exerts a profound influence on interatomic distances, electronic configurations, and bonding patterns in materials. These effects can induce phase transitions and unlock a wide range of intriguing properties that are often unattainable under ambient conditions.1 The investigation of materials under compression not only advances our understanding of fundamental condensed matter physics, but also provides essential insights into the behavior of planetary interiors and supports the development of functional materials, including superhard, catalytic, optoelectronic, thermoelectric, and superconducting materials.1–3 A notable example is diamond, which, synthesized under high-pressure and high-temperature conditions, remains stable at ambient pressure and exhibits exceptional hardness.4 Another significant case is the theoretical prediction and subsequent experimental discovery of the high-temperature superconducting hydride LaH10, which exhibits a record-breaking critical temperature of 260 K under high-pressure conditions.5–8 

High pressure in experiments can be generated using static methods, such as diamond anvil cells (DACs) and large-volume presses, or dynamic approaches, such as shock-wave facilities. However, these techniques are subject to limitations, including the small sample volumes of DACs and the transient nature of dynamic compression, both of which complicate data collection and interpretation. In this context, crystal structure prediction (CSP) methods play a pivotal role in complementing experimental studies. They not only assist in interpreting experimental measurements and resolving complex crystal structures, but also enable efficient predictions of high-pressure phases with intriguing properties, thereby providing valuable guidance for experimental exploration.

CSP seeks to determine the most stable atomic arrangement within a crystal for a given chemical composition and set of external conditions, such as pressure. From a thermodynamic perspective, this problem can be formulated as a global optimization task on a high-dimensional energy surface, where the goal is to identify the configuration corresponding to the lowest thermodynamic potential. In recent decades, significant progress has been achieved in addressing the computational challenges associated with CSP through the development of advanced methodologies and software packages. Prominent examples include AIRSS,9 basin hopping,10 CALYPSO,11,12 minima hopping,13 USPEX,14 MAGUS,15,16 and XtalOpt17 (see Ref. 18 for a more comprehensive list). These methods combine first-principles structural relaxation—most commonly performed using density functional theory (DFT)—with advanced sampling techniques for exploring the potential energy surface (PES) efficiently. They employ diverse strategies, such as random structure generation, molecular dynamics, evolutionary algorithms, and swarm intelligence. These advances have enabled significant theory-driven breakthroughs,18 including the solution of the complex crystal structures of alkali metals19–21 and the discovery of superhydride superconductors with remarkable critical temperatures.5–8,22,23

Despite significant advances, CSP remains fundamentally constrained by its NP-hard nature, since the computational complexity increases exponentially with system size. This inherent limitation highlights the need to address two critical challenges: efficient energy evaluation and effective PES sampling. Continued innovation in these areas is crucial for advancing CSP methodologies and enabling the prediction of increasingly complex crystal structures.

In recent years, machine learning (ML), as a rapidly evolving data-driven approach, has demonstrated exceptional capabilities in regression, pattern recognition, and generative tasks across diverse data types, including natural language24 and images.25 These techniques have increasingly become integral to computational physics, chemistry, and materials science, offering a promising route to address the aforementioned challenges in CSP.26–30 On the one hand, advances in supervised learning have led to the development of machine learning potentials (MLPs), which enable efficient and accurate evaluation of energies and forces, significantly accelerating atomic-scale simulations. On the other hand, the emergence of generative models (GMs) has introduced powerful tools for efficiently sampling the vast crystal configuration space, providing novel approaches for PES exploration and optimization.

In this review, we provide a brief overview of recent advances in CSP methodologies facilitated by ML techniques, with particular emphasis on MLPs and GMs. A schematic representation of the typical methods and their interrelations is provided in Fig. 1. We examine their applications to the discovery of high-pressure materials and critically analyze the associated challenges and opportunities within this rapidly evolving domain.

FIG. 1.

Schematic representation of crystal structure prediction, machine learning potentials, and crystal generative models, and their interrelations.

FIG. 1.

Schematic representation of crystal structure prediction, machine learning potentials, and crystal generative models, and their interrelations.

Close modal

MLPs are among the most promising approaches for accelerating quantum-mechanical-based calculations, enabling high-throughput, large-scale, and long-timescale atomic simulations with significantly reduced computational cost.31 In most MLP frameworks, the total energy of a system is expressed as the sum of the energies of its constituent atoms, where each atomic energy is determined by its local chemical environment.32 This localized representation allows MLPs to efficiently model the complex interactions within a system while maintaining computational efficiency. By leveraging the superior learning capacity of neural networks to approximate high-dimensional complex functions, MLPs are trained on data generated from high-fidelity quantum mechanical calculations, such as DFT. As a result, MLPs are designed to reproduce the accuracy of these quantum mechanical methods while achieving computational speeds that are orders of magnitude faster.

The first MLP applied to crystalline systems was the Behler–Parrinello neural network (BPNN).32 In this approach, the local atomic environment is represented using hand-crafted descriptors known as atom-centered symmetry functions (ACSFs).32 ACSFs capture two- and three-body interactions by projecting the atomic environment into a set of basis functions, effectively encoding the local geometry and chemical surroundings of each atom. Using ACSFs as input, a feedforward fully connected neural network is employed to regress the atomic energies, which are then summed to obtain the total energy of the system. Subsequently, many MLPs based on carefully devised descriptors—referred to as descriptor-based models—have been developed, including Gaussian approximation potentials (GAPs),33 moment tensor potentials (MTPs),34 and spectral neighbor analysis potentials (SNAPs),35 among others.36,37 These models offer robust performances for a wide range of applications within the scope of their designed capabilities. Among descriptor-based approaches, the Deep Potential (DP) model38 introduced a significant innovation by employing deep neural networks to learn more flexible and expressive representations of atomic environments during the training process.

Meanwhile, graph neural networks (GNNs)39 have emerged as powerful deep learning techniques for modeling interatomic potentials, in what are referred to as graph-based models. In this framework, a crystal structure is represented as a graph, with atoms treated as nodes and the interactions between them as edges. These graph-based models typically incorporate two-body (distances),40–44 three-body (angles),45–48 or even higher-order49,50 information. In the manner of message passing neural networks (MPNNs),51 messages between atoms are updated through multiple layers of the networks, and finally the regression target, typically the energy, is read out. A key factor contributing to the success of MLPs is their ability to exploit the inherent symmetries of crystals, including permutation, rotation, and translation invariance. While invariance under translation and rotation can be easily achieved by feeding invariant features (e.g., relative distance), some graph-based MLPs utilizing equivariant GNNs have been proposed. These equivariant models typically use specially designed architectures42,52 or Wigner-D matrices, spherical harmonics, and tensor product methods44,49,50,53–55 to preserve equivariance under rotation and invariance under translation. The inclusion of high-order tensors and equivariance in models can indeed improve data efficiency and generalization and result in a better learning curve.44,50 Furthermore, integrating transformer models with graph representations of structure also provides a promising route to the modeling of interatomic potentials. Equiformer/EquiformerV256,57 uses graph neural networks leveraging the strength of transformer architectures, while incorporating an equivariant graph attention mechanism to model interatomic potentials. Similar approaches that combine attention mechanisms with GNNs are also common in the literature.58–61 

Initially, MLPs were typically developed and trained from scratch for each specific system under investigation. This approach inherently restricted their generalizability and significantly limited their applicability to broader materials discovery and exploration. To overcome these limitations, Takamoto et al.62 introduced a universal MLP, PreFerred Potential (PFP), capable of handling combinations of 45 elements. PFP has demonstrated promising performance in diverse applications, including modeling lithium diffusion in LiFeSO4F and phase transitions in Cu–Au alloys. Similarly, M3GNet,47 a model based on GNNs incorporating three-body interactions, was developed and trained on the Materials Project63 database, which spans 89 elements in the Periodic Table. M3GNet successfully identified 1578 stable structures after screening over 31 × 106 candidates, highlighting its effectiveness in large-scale materials discovery. Furthermore, CHGNet,48 another universal MLP based on graph neural networks and trained on the Materials Project Trajectory (MPtrj) dataset,64 extended the predictive capabilities of MLPs by accurately predicting energies, forces, stresses, and magnetic moments. In parallel, DPA-265 introduced a pretraining–fine-tuning approach to develop a universal MLP, in which a universal descriptor was first trained on a diverse dataset calculated at different levels of theory and then fine-tuned for specific chemical systems of interest. MatterSim66 further advanced the development of universal MLPs by training on a highly diverse dataset that incorporated not only a wide range of chemical species, but also varying pressure and temperature conditions, thereby broadening its applicability. Other notable efforts include MACE,50,67 GNoME-NequIP,68 EquiformerV2-OMat24,69 Orb,70 SevenNet,71 and grACE.72 Collectively, these advances represent significant strides toward the realization of universal MLPs, offering transformative potential for comprehensive materials discovery, design, and exploration.

Given the optimal trade-off between accuracy and computational efficiency provided by MLPs, leveraging a well-trained MLP specifically tailored to a target system provides a practical and effective method for accelerating structure prediction by replacing DFT in energy evaluations. For instance, Deringer et al.73 combined a GAP, trained on liquid and amorphous structures, with the AIRSS method, successfully discovering several new carbon allotropes. Similarly, Wang et al.74 integrated the DP model with the CALYPSO method to conduct structure searches on binary Al–Mg alloy systems, uncovering six previously unreported metastable structures. By combining an MLP with the stochastic surface walking global optimization method,75,76 Huang et al.77 proposed several competitive candidates after exploring the PES of β-B, and revealed the key principles governing the filling of interstitial sites. These examples78 highlight the growing potential of MLPs in accelerating and enhancing CSP across a wide range of material systems.

MLPs have demonstrated notable accuracy and generalization, but often suffer from significant prediction uncertainty in regions of the PES beyond the training data domain. This limitation is particularly critical for CSP tasks, which require extensive PES exploration to identify low-energy structures.78 Integrating MLPs with CSP in an active learning (AL) framework offers a promising solution. Specifically, an AL strategy iteratively performs configuration extrapolation detection, DFT labeling, training dataset refinement, and MLP updates, thereby improving the reliability of MLP predictions and accelerating the CSP process.

In this context, Deringer et al.79,80 combined a GAP with AIRSS9 in an AL framework. Starting with DFT calculations of randomly generated structures to train an initial GAP, they iteratively refined the model by using the GAP to relax new structures and incorporating selected relaxed structures into the training dataset. Enhancements, such as selecting structures on the basis of physical constraints, have enabled successful CSP applications to systems such as phosphorus, carbon, silicon, and titanium.80,81 Tong et al.82 identified a novel core–shell-type boron B84 cluster using an active learning scheme combining an ACSF-based GAP with CALYPSO.11,12 Similarly, Yang et al.83 iteratively constructed an ACSF-based GAP with CALYPSO, discovering a new boron phase, c-B24, composed of B6 octahedra, B2 pairs, and B12 icosahedra, which exhibits a superconducting critical temperature of 13.8 K at ambient pressure. Podryabinkin and Shapeev84 developed an AL framework for MTPs using the D-optimality criterion for efficient configuration selection, and Podryabinkin et al.85 subsequently integrated this with USPEX14 to accelerate CSP. This method successfully predicted crystal structures of carbon, high-pressure sodium phases, and boron allotropes, including a previously unknown 54-atom boron structure,85 with minimal computational cost.86 Cheng et al.87 used a GNN to model the relationship between crystal structures and formation enthalpies, integrated three optimization algorithms to enhance the efficiency of searching for the structure with the lowest formation enthalpy, and updated the GNN prediction model using AL. By identifying crystal structures of 29 compounds, they showcased the potential of this approach to accelerate materials discovery.

The effectiveness of AL approaches is particularly evident in the discovery of novel high-pressure phases. Pickard88 introduced the ephemeral data-driven potentials (EDDP) framework, a methodology designed to dynamically construct and refine interatomic potentials using minimal, on-the-fly generated data. A previously unreported high-pressure silane (SiH4) phase with 12 formula units in the primitive cell was discovered through the application of EDDP and was determined to be thermodynamically stable at ∼300 GPa. Moreover, the EDDP framework facilitated the revisitation and identification of a series of stable scandium hydride phases at 300 GPa,89 thus demonstrating its efficacy in exploring high-pressure phase stability and uncovering novel structures. Sun and co-workers used MAGUS15,16 in combination with AL to predict a series of new compounds under planetary interior conditions, primarily high pressure and high temperature. They successfully predicted a novel ground state of silica at pressures between 645 and 890 GPa, featuring six-, eight-, and nine-coordinated silicon atoms.90 Additionally, they identified several stable magnesium oxide–water compounds within the pressure range of 270–600 GPa, which could potentially serve as a water reservoir on the early Earth.91 By exploring the phase diagram of the lithium–aluminum system from 0 to 150 GPa, they revealed a pressure-stabilized compound exhibiting both superconducting and superionic behaviors.92 These findings underscore the power of MLPs in high-pressure CSP.

Wang et al.93 employed an AL scheme to construct a DP model suitable for lithium at high pressure. Using this model, they explored over 600 000 crystal structures at 50 and 60 GPa, identifying four candidate structures for previously unknown crystal phases located at the minima of the melting curve. Among these, two structures, aP160 and oP192, containing more than 150 atoms in the unit cell, with P1 and Pcc2 space groups respectively, were discovered.93 These findings underscore the predictive power of global structure search methods when combined with accurate ML potentials for uncovering complex crystal structures. Furthermore, Wang et al.94 combined CALYPSO with DP in an AL manner. They revisited the complex ternary superhydride compound Li–La–H, a material with potential for high-temperature superconductivity, at 300 GPa. After exploring 375 000 DP-based and 1475 DFT-based local minima structures, a new phase, Cmcm-Li2La2H23,94 was revealed. This phase is thermodynamically stable at 300 GPa and exhibits a critical temperature of ∼130 K. This approach reduced the computational cost by two orders of magnitude compared with traditional DFT-based CSP methods.

Additionally, MLPs have demonstrated considerable efficiency, making them suitable for large-scale simulations. The focus of the present paper is on CSP, and therefore a more detailed discussion of MLPs is beyond its scope. However, comprehensive reviews on this topic can be found in the literature.27,31,95,96

Unsupervised learning GMs have demonstrated a significant capacity to approximate the underlying probability distributions of complex datasets, enabling the generation of novel data instances through sampling from these learned distributions. Traditionally, CSP methods have relied primarily on techniques such as Monte Carlo simulations, molecular dynamics, and global optimization algorithms to explore the PES and generate new structures. However, the recent emergence of comprehensive materials databases, encompassing both structural and property information, has facilitated the integration of unsupervised learning algorithms into materials design and CSP. This integration represents a paradigm shift in CSP methodologies, offering a complementary and often more efficient approach to exploring the vast configuration space. Prominent examples of GMs in ML include generative adversarial networks (GANs), variational aautoencoders (VAEs), diffusion models, flow models, and autoregressive models.

The applications of GMs to crystal structures began with descriptor-based rough/compressed representation of structures,97–111 mostly utilizing GANs and VAEs. For instance, CrystalGAN97 employed a 2D raw coordinate-based representation and trained a GAN on a dataset of binary hydrides to generate ternary hydrides. Hoffmann et al.99 used a voxel representation of structures, akin to image generation techniques, employing a VAE/U-Net framework to generate voxelized structure images. Compression of crystal structure representations inevitably results in a loss of information, and therefore additional efforts are required for these methods to reconstruct structures from their representations. Voxel representation approaches99–102,106,109 usually employ object detection techniques to ascertain atomic positions, whereas fingerprint representation methods98,104,112,113 often rely on composition substitution or random generation to identify structures that match the desired fingerprint. Furthermore, neglecting the fundamental permutation, rotation, and translation invariance that are inherent in crystal structures could shorten a model’s expressive capacity. In addition, the restriction of training data to specific atomic species or lattice types constrains the ability of these models to accommodate diverse and realistic crystalline datasets.

The advent of invariant/equivalent GNNs marked a significant breakthrough in this field, particularly with the introduction of CDVAE,114 which utilized crystal graph representations within a VAE/diffusion framework. This has led to a series of methodologies capable of learning from comprehensive elemental datasets and efficiently generating high-quality crystalline structures.115–125 For example, DiffCSP118 employs a periodic-E(3)-equivariant denoising model to generate structures under a fractional coordinate system, and FlowMM121 adopts a generalized Riemannian flow matching technique for crystal structure. The integration of symmetry-aware structure generation into graph-based models is also a crucial development.119,126,127 CHGFlowNet126 achieves this by using a sequential mechanism to generate symmetry-nonequivalent atoms one-by-one, and DiffCSP++119 enforces symmetry constraints during the diffusion process to ensure that the generated structures retain the symmetry of the input configuration. Furthermore, MatterGen,120 Con-CDVAE,128 Cond-CDVAE,122 and CrystalFlow123 integrate conditional generation techniques into these graph-based models to generate structures under targeted external conditions or material properties, such as space group label, magnetic density, bandgap, bulk modulus, formation energy, and pressure, significantly advancing the fields of CSP and inverse material design.

In parallel, inspired by advances in natural language processing and large language models (LLMs), another class of LLM-based models utilizing character-type representation has rapidly emerged. These methods employ sequential tokenization of crystal structures and utilize an autoregressive mechanism—typically recurrent neural networks (RNNs) or transformer models—to achieve dialogue-oriented generation schemes.129–140 For example, CrystalLLM130 utilizes a LLM to generate the texts of crystallographic information files (CIFs), and CrystalFormer135 utilizes a transformer model to generate structures by a sequence of symmetric Wyckoff sites. Such approaches are well suited for facilitating the integration of multimodal learning strategies, encompassing literature review and material synthesis path retrieval, thus demonstrating a promising direction for new material design.

As a data-driven approach, the development of GMs capable of producing crystal structures under specified high-pressure conditions necessitates a comprehensive high-pressure crystal structure dataset. In this context, a CALYPSO high-pressure dataset containing more than 657 000 structures and mainly covering pressures between 0 and 300 GPa has been curated, leading to the development of two models, Cond-CDVAE122 and CrystalFlow.123 Cond-CDVAE employs a diffusion process to produce promising structures, while CrystalFlow utilizes a flow model architecture. Both models adopt a conditional generation approach that leverages pressure as a conditioning variable. It has been shown that with Cond-CDVAE, significantly less effort is required to identity the stable high-pressure phases of lithium, boron, and silica compared with traditional CSP methods. Meanwhile, CrystalFlow exhibits enhanced capabilities in generating local stable structures under high-pressure conditions. Despite the shared mathematical formulation of diffusion and flow models,141,142 diffusion models benefit from the accumulated results of their application in other fields, particularly computer vision,143 and thus at present offer a more mature approach compared with flow models. However, research on flow models is rapidly advancing.144 Overall, these advances represent a substantial step forward in the data-driven exploration of crystal structures under extreme conditions, paving the way for future research in materials science.

The exploration of materials under high-pressure conditions is pivotal for understanding fundamental condensed matter physics, planetary interiors, and the development of functional materials with exceptional properties. In this review, we have highlighted the important role of CSP methodologies, particularly those enhanced by ML techniques, in addressing the challenges faced by traditional computational techniques. By leveraging advances in MLPs and GMs, researchers have made significant strides in accelerating the discovery of complex crystal structures and uncovering novel high-pressure phases.

Looking ahead, several key opportunities and challenges define the future trajectory of this rapidly evolving field. First, the development of universal and transferable MLPs that can accurately model diverse chemical systems across varying pressure and temperature conditions remains a critical goal. Recent progress in AL frameworks and pretraining–fine-tuning strategies has demonstrated the potential to enhance MLP generalizability, but further efforts are needed to expand these models to encompass more elements, complex bonding environments, and extreme conditions.

Second, generative models offer a promising avenue for efficient exploration of the vast configuration space of crystal structures. However, to fully realize their potential, future work must address challenges such as ensuring physical validity, incorporating symmetry constraints, and integrating property-driven condition generation. The incorporation of advanced techniques, such as equivariant neural networks and graph-based representations, has already shown promise in generating high-quality structures, but continued innovation is required to improve data efficiency, scalability, and the ability to target specific material properties.

Finally, as the field progresses, interdisciplinary collaboration will play a crucial role in addressing the multifaceted challenges associated with high-pressure materials discovery. Advances in computational physics and ML techniques must be synergistically combined to push the boundaries of what is achievable. Moreover, the establishment of open-access datasets, standardized benchmarks, and community-driven initiatives will be essential for fostering innovation and accelerating progress.

The work is supported by the National Key Research and Development Program of China (Grant No. 2022YFA1402304), the National Natural Science Foundation of China (Grant Nos. 12034009, 12374005, 52288102, 52090024, and T2225013), the Fundamental Research Funds for the Central Universities, and the Program for JLU Science and Technology Innovative Research Team.

The authors have no conflicts to disclose.

Zhenyu Wang: Investigation (equal); Writing – original draft (equal); Writing – review & editing (equal). Xiaoshan Luo: Investigation (equal); Writing – original draft (equal); Writing – review & editing (equal). Qingchang Wang: Investigation (equal); Writing – review & editing (equal). Heng Ge: Investigation (equal); Writing – review & editing (equal). Pengyue Gao: Investigation (equal); Writing – review & editing (equal). Wei Zhang: Conceptualization (equal); Supervision (equal); Writing – review & editing (equal). Jian Lv: Conceptualization (equal); Supervision (equal); Writing – original draft (equal); Writing – review & editing (equal). Yanchao Wang: Conceptualization (equal); Supervision (equal); Writing – review & editing (equal).

The data that support the findings of this study are available from the corresponding authors upon reasonable request.

1.
L.
Zhang
,
Y.
Wang
,
J.
Lv
, and
Y.
Ma
, “
Materials discovery at high pressures
,”
Nat. Rev. Mater.
2
,
17005
(
2017
).
2.
M.
Xu
,
Y.
Li
, and
Y.
Ma
, “
Materials by design at high pressures
,”
Chem. Sci.
13
,
329
(
2022
).
3.
H.-K.
Mao
,
B.
Chen
,
J.
Chen
,
K.
Li
,
J.-F.
Lin
et al, “
Recent advances in high-pressure science and technology
,”
Matter Radiat. Extremes
1
,
59
(
2016
).
4.
Q.
Huang
,
D.
Yu
,
B.
Xu
,
W.
Hu
,
Y.
Ma
et al, “
Nanotwinned diamond with unprecedented hardness and stability
,”
Nature
510
,
250
(
2014
).
5.
H.
Liu
,
I. I.
Naumov
,
R.
Hoffmann
,
N. W.
Ashcroft
, and
R. J.
Hemley
, “
Potential high-Tc superconducting lanthanum and yttrium hydrides at high pressure
,”
Proc. Natl. Acad. Sci. U. S. A.
114
,
6990
(
2017
).
6.
F.
Peng
,
Y.
Sun
,
C. J.
Pickard
,
R. J.
Needs
,
Q.
Wu
et al, “
Hydrogen clathrate structures in rare Earth hydrides at high pressures: Possible route to room-temperature superconductivity
,”
Phys. Rev. Lett.
119
,
107001
(
2017
).
7.
A. P.
Drozdov
,
P. P.
Kong
,
V. S.
Minkov
,
S. P.
Besedin
,
M. A.
Kuzovnikov
et al, “
Superconductivity at 250 K in lanthanum hydride under high pressures
,”
Nature
569
,
528
(
2019
).
8.
M.
Somayazulu
,
M.
Ahart
,
A. K.
Mishra
,
Z. M.
Geballe
,
M.
Baldini
et al, “
Evidence for superconductivity above 260 K in lanthanum superhydride at megabar pressures
,”
Phys. Rev. Lett.
122
,
027001
(
2019
).
9.
C. J.
Pickard
and
R. J.
Needs
, “
Ab initio random structure searching
,”
J. Phys. Condens. Matter
23
,
053201
(
2011
).
10.
M. T.
Oakley
,
R. L.
Johnston
, and
D. J.
Wales
, “
Symmetrisation schemes for global optimisation of atomic clusters
,”
Phys. Chem. Chem. Phys.
15
,
3965
(
2013
).
11.
Y.
Wang
,
J.
Lv
,
L.
Zhu
, and
Y.
Ma
, “
Calypso: A method for crystal structure prediction
,”
Comput. Phys. Commun.
183
,
2063
(
2012
).
12.
X.
Shao
,
J.
Lv
,
P.
Liu
,
S.
Shao
,
P.
Gao
et al, “
A symmetry-orientated divide-and-conquer method for crystal structure prediction
,”
J. Chem. Phys.
156
,
014105
(
2022
).
13.
S.
Goedecker
, “
Minima hopping: An efficient search method for the global minimum of the potential energy surface of complex molecular systems
,”
J. Chem. Phys.
120
,
9911
(
2004
).
14.
C. W.
Glass
,
A. R.
Oganov
, and
N.
Hansen
, “
USPEX—Evolutionary crystal structure prediction
,”
Comput. Phys. Commun.
175
,
713
(
2006
).
15.
K.
Xia
,
H.
Gao
,
C.
Liu
,
J.
Yuan
,
J.
Sun
et al, “
A novel superhard tungsten nitride predicted by machine-learning accelerated crystal structure search
,”
Sci. Bull.
63
,
817
(
2018
).
16.
J.
Wang
,
H.
Gao
,
Y.
Han
,
C.
Ding
,
S.
Pan
et al, “
MAGUS: Machine learning and graph theory assisted universal structure searcher
,”
Natl. Sci. Rev.
10
,
nwad128
(
2023
).
17.
D. C.
Lonie
and
E.
Zurek
, “
XtalOpt: An open-source evolutionary algorithm for crystal structure prediction
,”
Comput. Phys. Commun.
182
,
372
(
2011
).
18.
Y.
Wang
,
J.
Lv
,
P.
Gao
, and
Y.
Ma
, “
Crystal structure prediction via efficient sampling of the potential energy surface
,”
Acc. Chem. Res.
55
,
2068
(
2022
).
19.
J.
Lv
,
Y.
Wang
,
L.
Zhu
, and
Y.
Ma
, “
Predicted novel high-pressure phases of lithium
,”
Phys. Rev. Lett.
106
,
015503
(
2011
).
20.
M.
Marqués
,
M. I.
McMahon
,
E.
Gregoryanz
,
M.
Hanfland
,
C. L.
Guillaume
et al, “
Crystal structures of dense lithium: A metal-semiconductor-metal transition
,”
Phys. Rev. Lett.
106
,
095502
(
2011
).
21.
Y.
Ma
,
M.
Eremets
,
A. R.
Oganov
,
Y.
Xie
,
I.
Trojan
et al, “
Transparent dense sodium
,”
Nature
458
,
182
(
2009
).
22.
H.
Wang
,
J. S.
Tse
,
K.
Tanaka
,
T.
Iitaka
, and
Y.
Ma
, “
Superconductive sodalite-like clathrate calcium hydride at high pressures
,”
Proc. Natl. Acad. Sci. U. S. A.
109
,
6463
(
2012
).
23.
L.
Ma
,
K.
Wang
,
Y.
Xie
,
X.
Yang
,
Y.
Wang
et al, “
High-temperature superconducting phase in clathrate calcium hydride CaH6 up to 215 K at a pressure of 172 GPa
,”
Phys. Rev. Lett.
128
,
167001
(
2022
).
24.
T. B.
Brown
,
B.
Mann
,
N.
Ryder
,
M.
Subbiah
,
J.
Kaplan
et al, “
Language models are few-shot learners
,” arXiv:2005.14165 (
2020
).
25.
R.
Rombach
,
A.
Blattmann
,
D.
Lorenz
,
P.
Esser
, and
B.
Ommer
, “
High-resolution image synthesis with latent diffusion models
,” in
2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022)
(
IEEE
,
2022
), pp.
10674
10685
.
26.
J.
Schmidt
,
M. R. G.
Marques
,
S.
Botti
, and
M. A. L.
Marques
, “
Recent advances and applications of machine learning in solid-state materials science
,”
npj Comput. Mater.
5
,
83
(
2019
).
27.
G.
Wang
,
C.
Wang
,
X.
Zhang
,
Z.
Li
,
J.
Zhou
et al, “
Machine learning interatomic potential: Bridge the gap between small-scale models and realistic device-scale simulations
,”
iScience
27
,
109673
(
2024
).
28.
L.
Chuan-Nan
,
L.
Han-Pu
,
Z.
Bai-Qing
,
W.
Su-Huai
, and
Z.
Xie
, “
Machine learning assisted crystal structure prediction made simple
,”
J. Mater. Inf.
4
,
15
(
2024
).
29.
Z.
Wang
,
H.
Hua
,
W.
Lin
,
M.
Yang
, and
K. C.
Tan
, “
Crystalline material discovery in the era of artificial intelligence
,” arXiv:2408.08044 (
2024
).
30.
X.-Q.
Han
,
X.-D.
Wang
,
M.-Y.
Xu
,
Z.
Feng
,
B.-W.
Yao
et al, “
AI-driven inverse design of materials: Past, present and future
,” arXiv:2411.09429 (
2024
).
31.
J.
Behler
, “
Four generations of high-dimensional neural network potentials
,”
Chem. Rev.
121
,
10037
(
2021
).
32.
J.
Behler
and
M.
Parrinello
, “
Generalized neural-network representation of high-dimensional potential-energy surfaces
,”
Phys. Rev. Lett.
98
,
146401
(
2007
).
33.
A. P.
Bartók
,
M. C.
Payne
,
R.
Kondor
, and
G.
Csányi
, “
Gaussian approximation potentials: The accuracy of quantum mechanics, without the electrons
,”
Phys. Rev. Lett.
104
,
136403
(
2010
).
34.
A. V.
Shapeev
, “
Moment tensor potentials: A class of systematically improvable interatomic potentials
,”
Multiscale Model. Simul.
14
,
1153
(
2016
).
35.
A. P.
Thompson
,
L. P.
Swiler
,
C. R.
Trott
,
S. M.
Foiles
, and
G. J.
Tucker
, “
Spectral neighbor analysis method for automated generation of quantum-accurate interatomic potentials
,”
J. Comput. Phys.
285
,
316
(
2015
).
36.
R.
Drautz
, “
Atomic cluster expansion for accurate and transferable interatomic potentials
,”
Phys. Rev. B
99
,
014104
(
2019
).
37.
Y.
Zhang
,
C.
Hu
, and
B.
Jiang
, “
Embedded atom neural network potentials: Efficient and accurate machine learning with a physically inspired representation
,”
J. Phys. Chem. Lett.
10
,
4962
(
2019
).
38.
L.
Zhang
,
J.
Han
,
H.
Wang
,
W. A.
Saidi
,
R.
Car
et al, “
End-to-end symmetry preserving inter-atomic potential energy model for finite and extended systems
,” arXiv:1805.09003 (
2018
).
39.
F.
Scarselli
,
M.
Gori
,
A. C.
Tsoi
,
M.
Hagenbuchner
, and
G.
Monfardini
, “
The graph neural network model
,”
IEEE Trans. Neural Netw.
20
,
61
(
2009
).
40.
K. T.
Schütt
,
F.
Arbabzadah
,
S.
Chmiela
,
K. R.
Müller
, and
A.
Tkatchenko
, “
Quantum-chemical insights from deep tensor neural networks
,”
Nat. Commun.
8
,
13890
(
2017
).
41.
T.
Xie
and
J. C.
Grossman
, “
Crystal graph convolutional neural networks for an accurate and interpretable prediction of material properties
,”
Phys. Rev. Lett.
120
,
145301
(
2018
).
42.
K. T.
Schütt
,
P.-J.
Kindermans
,
H. E.
Sauceda
,
S.
Chmiela
,
A.
Tkatchenko
et al, “
SchNet: A continuous-filter convolutional neural network for modeling quantum interactions
,” arXiv:1706.08566 (
2017
).
43.
O. T.
Unke
and
M.
Meuwly
, “
PhysNet: A neural network for predicting energies, forces, dipole moments, and partial charges
,”
J. Chem. Theor. Comput.
15
,
3678
(
2019
).
44.
S.
Batzner
,
A.
Musaelian
,
L.
Sun
,
M.
Geiger
,
J. P.
Mailoa
et al, “
E(3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials
,”
Nat. Commun.
13
,
2453
(
2022
).
45.
J.
Gasteiger
,
J.
Groß
, and
S.
Günnemann
, “
Directional message passing for molecular graphs
,” arXiv:2003.03123 (
2020
).
46.
J.
Gasteiger
,
S.
Giri
,
J. T.
Margraf
, and
S.
Günnemann
, “
Fast and uncertainty-aware directional message passing for non-equilibrium molecules
,” arXiv:2011.14115 (
2020
).
47.
C.
Chen
and
S. P.
Ong
, “
A universal graph deep learning interatomic potential for the periodic table
,”
Nat. Comput. Sci.
2
,
718
(
2022
).
48.
B.
Deng
,
P.
Zhong
,
K.
Jun
,
J.
Riebesell
,
K.
Han
et al, “
CHGNet as a pretrained universal neural network potential for charge-informed atomistic modelling
,”
Nat. Mach. Intell.
5
,
1031
(
2023
).
49.
J.
Gasteiger
,
F.
Becker
, and
S.
Günnemann
, “
GemNet: Universal directional graph neural networks for molecules
,” arXiv:2106.08903 (
2021
).
50.
I.
Batatia
,
D. P.
Kovács
,
G. N. C.
Simm
,
C.
Ortner
, and
G.
Csányi
, “
MACE: Higher order equivariant message passing neural networks for fast and accurate force fields
,” arXiv:2206.07697 (
2022
).
51.
J.
Gilmer
,
S. S.
Schoenholz
,
P. F.
Riley
,
O.
Vinyals
, and
G. E.
Dahl
, “
Neural message passing for quantum chemistry
,” arXiv:1704.01212 (
2017
).
52.
V. G.
Satorras
,
E.
Hoogeboom
, and
M.
Welling
, “
E(n) equivariant graph neural networks
,” arXiv:2102.09844 (
2021
).
53.
M.
Geiger
and
T.
Smidt
, “
e3nn: Euclidean neural networks
,” arXiv:2207.09453 (
2022
).
54.
N.
Thomas
,
T.
Smidt
,
S.
Kearnes
,
L.
Yang
,
L.
Li
et al, “
Tensor field networks: Rotation- and translation-equivariant neural networks for 3D point clouds
,” arXiv:1802.08219 (
2018
).
55.
J.
Wang
,
Y.
Wang
,
H.
Zhang
,
Z.
Yang
,
Z.
Liang
et al, “
E(n)-Equivariant cartesian tensor message passing interatomic potential
,”
Nat. Commun.
15
,
7607
(
2024
).
56.
Y.-L.
Liao
and
T.
Smidt
, “
Equiformer: Equivariant graph attention transformer for 3D atomistic graphs
,” arXiv:2206.11990 (
2022
).
57.
Y.-L.
Liao
,
B.
Wood
,
A.
Das
, and
T.
Smidt
, “
EquiformerV2: Improved equivariant transformer for scaling to higher-degree representations
,” arXiv:2306.12059 (
2023
).
58.
F. B.
Fuchs
,
D. E.
Worrall
,
V.
Fischer
, and
M.
Welling
, “
SE(3)-Transformers: 3D roto-translation equivariant attention networks
,” arXiv:2006.10503 (
2020
).
59.
M.
Hutchinson
,
C. L.
Lan
,
S.
Zaidi
,
E.
Dupont
,
Y. W.
Teh
et al, “
LieTransformer: Equivariant self-attention for lie groups
,” arXiv:2012.10885 (
2020
).
60.
C.
Ying
,
T.
Cai
,
S.
Luo
,
S.
Zheng
,
G.
Ke
et al, “
Do transformers really perform bad for graph representation?
,” arXiv:2106.05234 (
2021
).
61.
P.
Thölke
and
G. D.
Fabritiis
, “
TorchMD-NET: Equivariant transformers for neural network based molecular potentials
,” arXiv:2202.02541 (
2022
).
62.
S.
Takamoto
,
C.
Shinagawa
,
D.
Motoki
,
K.
Nakago
,
W.
Li
et al, “
Towards universal neural network potential for material discovery applicable to arbitrary combination of 45 elements
,”
Nat. Commun.
13
,
2991
(
2022
).
63.
A.
Jain
,
S. P.
Ong
,
G.
Hautier
,
W.
Chen
,
W. D.
Richards
et al, “
Commentary: The materials Project: A materials genome approach to accelerating materials innovation
,”
APL Mater.
1
,
011002
(
2013
).
64.
B.
Deng
(
2023
). “
Materials project trajectory (MPtrj) dataset,” Figshare
. https://doi.org/10.6084/m9.figshare.23713842.v2
65.
D.
Zhang
,
X.
Liu
,
X.
Zhang
,
C.
Zhang
,
C.
Cai
et al, “
DPA-2: A large atomic model as a multi-task learner
,”
npj Comput. Mater.
10
,
293
(
2024
).
66.
H.
Yang
,
C.
Hu
,
Y.
Zhou
,
X.
Liu
,
Y.
Shi
et al, “
MatterSim: A deep learning atomistic model across elements, temperatures and pressures
,” arXiv:2405.04967 (
2024
).
67.
I.
Batatia
,
P.
Benner
,
Y.
Chiang
,
A. M.
Elena
,
D. P.
Kovács
et al, “
A foundation model for atomistic materials chemistry
,” arXiv:2401.00096 (
2023
).
68.
A.
Merchant
,
S.
Batzner
,
S. S.
Schoenholz
,
M.
Aykol
,
G.
Cheon
et al, “
Scaling deep learning for materials discovery
,”
Nature
624
,
80
(
2023
).
69.
L.
Barroso-Luque
,
M.
Shuaibi
,
X.
Fu
,
B. M.
Wood
,
M.
Dzamba
et al, “
Open materials 2024 (OMat24) inorganic materials dataset and models
,” arXiv:2410.12771 (
2024
).
70.
M.
Neumann
,
J.
Gin
,
B.
Rhodes
,
S.
Bennett
,
Z.
Li
et al, “
Orb: A fast, scalable neural network potential
,” arXiv:2410.22570 (
2024
).
71.
Y.
Park
,
J.
Kim
,
S.
Hwang
, and
S.
Han
, “
Scalable parallel algorithm for graph neural network interatomic potentials in molecular dynamics simulations
,”
J. Chem. Theor. Comput.
20
,
4857
(
2024
).
72.
A.
Bochkarev
,
Y.
Lysogorskiy
, and
R.
Drautz
, “
Graph atomic cluster expansion for semilocal interactions beyond equivariant message passing
,”
Phys. Rev. X
14
,
021036
(
2024
).
73.
V. L.
Deringer
,
G.
Csányi
, and
D. M.
Proserpio
, “
Extracting crystal chemistry from amorphous carbon structures
,”
ChemPhysChem
18
,
873
(
2017
).
74.
H.
Wang
,
Y.
Zhang
,
L.
Zhang
, and
H.
Wang
, “
Crystal structure prediction of binary alloys via deep potential
,”
Front. Chem.
8
,
589795
(
2020
).
75.
C.
Shang
and
Z.-P.
Liu
, “
Stochastic surface walking method for structure prediction and pathway searching
,”
J. Chem. Theor. Comput.
9
,
1838
(
2013
).
76.
C.
Shang
,
X.-J.
Zhang
, and
Z.-P.
Liu
, “
Stochastic surface walking method for crystal structure and phase transition pathway prediction
,”
Phys. Chem. Chem. Phys.
16
,
17845
(
2014
).
77.
S.-D.
Huang
,
C.
Shang
,
P.-L.
Kang
, and
Z.-P.
Liu
, “
Atomic structure of boron resolved using machine learning and global sampling
,”
Chem. Sci.
9
,
8644
(
2018
).
78.
Q.
Tong
,
P.
Gao
,
H.
Liu
,
Y.
Xie
,
J.
Lv
et al, “
Combining machine learning potential and structure prediction for accelerated materials design and discovery
,”
J. Phys. Chem. Lett.
11
,
8710
(
2020
).
79.
V. L.
Deringer
,
C. J.
Pickard
, and
G.
Csányi
, “
Data-driven learning of total and local energies in elemental boron
,”
Phys. Rev. Lett.
120
,
156001
(
2018
).
80.
V. L.
Deringer
,
D. M.
Proserpio
,
G.
Csányi
, and
C. J.
Pickard
, “
Data-driven learning and prediction of inorganic crystal structures
,”
Faraday Discuss.
211
,
45
(
2018
).
81.
N.
Bernstein
,
G.
Csányi
, and
V. L.
Deringer
, “
De novo exploration and self-guided learning of potential-energy surfaces
,”
npj Comput. Mater.
5
,
99
(
2019
).
82.
Q.
Tong
,
L.
Xue
,
J.
Lv
,
Y.
Wang
, and
Y.
Ma
, “
Accelerating CALYPSO structure prediction by data-driven learning of a potential energy surface
,”
Faraday Discuss.
211
,
31
(
2018
).
83.
Q.
Yang
,
J.
Lv
,
Q.
Tong
,
X.
Du
,
Y.
Wang
et al, “
Hard and superconducting cubic boron phase via swarm-intelligence structural prediction driven by a machine-learning potential
,”
Phys. Rev. B
103
,
024505
(
2021
).
84.
E. V.
Podryabinkin
and
A. V.
Shapeev
, “
Active learning of linearly parametrized interatomic potentials
,”
Comput. Mater. Sci.
140
,
171
(
2017
).
85.
E. V.
Podryabinkin
,
E. V.
Tikhonov
,
A. V.
Shapeev
, and
A. R.
Oganov
, “
Accelerating crystal structure prediction by machine-learning interatomic potentials with active learning
,”
Phys. Rev. B
99
,
064114
(
2019
).
86.
K.
Gubaev
,
E. V.
Podryabinkin
,
G. L. W.
Hart
, and
A. V.
Shapeev
, “
Accelerating high-throughput searches for new alloys with active learning of interatomic potentials
,”
Comput. Mater. Sci.
156
,
148
(
2019
).
87.
G.
Cheng
,
X.-G.
Gong
, and
W.-J.
Yin
, “
Crystal structure prediction by combining graph network and optimization algorithm
,”
Nat. Commun.
13
,
1492
(
2022
).
88.
C. J.
Pickard
, “
Ephemeral data derived potentials for random structure search
,”
Phys. Rev. B
106
,
014102
(
2022
).
89.
P. T.
Salzbrenner
,
S. H.
Joo
,
L. J.
Conway
,
P. I. C.
Cooke
,
B.
Zhu
et al, “
Developments and further applications of ephemeral data derived potentials
,”
J. Chem. Phys.
159
,
144801
(
2023
).
90.
C.
Liu
,
J.
Shi
,
H.
Gao
,
J.
Wang
,
Y.
Han
et al, “
Mixed coordination silica at megabar pressure
,”
Phys. Rev. Lett.
126
,
035701
(
2021
).
91.
S.
Pan
,
T.
Huang
,
A.
Vazan
,
Z.
Liang
,
C.
Liu
et al, “
Magnesium oxide-water compounds at megabar pressure and implications on planetary interiors
,”
Nat. Commun.
14
,
1165
(
2023
).
92.
X.
Wang
,
Y.
Wang
,
J.
Wang
,
S.
Pan
,
Q.
Lu
et al, “
Pressure stabilized lithium-aluminum compounds with both superconducting and superionic behaviors
,”
Phys. Rev. Lett.
129
,
246403
(
2022
).
93.
X.
Wang
,
Z.
Wang
,
P.
Gao
,
C.
Zhang
,
J.
Lv
et al, “
Data-driven prediction of complex crystal structures of dense lithium
,”
Nat. Commun.
14
,
2924
(
2023
).
94.
Z.
Wang
,
X.
Wang
,
X.
Luo
,
P.
Gao
,
Y.
Sun
et al, “
Concurrent learning scheme for crystal structure prediction
,”
Phys. Rev. B
109
,
094117
(
2024
).
95.
P.
Friederich
,
F.
Häse
,
J.
Proppe
, and
A.
Aspuru-Guzik
, “
Machine-learned potentials for next-generation matter simulations
,”
Nat. Mater.
20
,
750
(
2021
).
96.
V. L.
Deringer
,
A. P.
Bartók
,
N.
Bernstein
,
D. M.
Wilkins
,
M.
Ceriotti
et al, “
Gaussian process regression for materials and molecules
,”
Chem. Rev.
121
,
10073
(
2021
).
97.
A.
Nouira
,
N.
Sokolovska
, and
J.-C.
Crivello
, “
CrystalGAN: Learning to discover crystallographic structures with generative adversarial networks
,” in
Proceedings of the AAAI 2019 Spring Symposium on
Combining Machine Learning with Knowledge Engineering (AAAI
,
2018
).
98.
K.
Ryan
,
J.
Lengyel
, and
M.
Shatruk
, “
Crystal structure prediction via deep learning
,”
J. Am. Chem. Soc.
140
,
10158
(
2018
).
99.
J.
Hoffmann
,
L.
Maestrati
,
Y.
Sawada
,
J.
Tang
,
J. M.
Sellier
et al, “
Data-driven approach to encoding and decoding 3-D crystal structures
,” arXiv:1909.00949 (
2019
).
100.
J.
Noh
,
J.
Kim
,
H. S.
Stein
,
B.
Sanchez-Lengeling
,
J. M.
Gregoire
et al, “
Inverse design of solid-state materials via a continuous representation
,”
Matter
1
,
1370
(
2019
).
101.
C. J.
Court
,
B.
Yildirim
,
A.
Jain
, and
J. M.
Cole
, “
3-D inorganic crystal structure generation and property prediction via representation learning
,”
J. Chem. Inf. Model.
60
,
4518
(
2020
).
102.
B.
Kim
,
S.
Lee
, and
J.
Kim
, “
Inverse design of porous materials using artificial neural networks
,”
Sci. Adv.
6
,
eaax9324
(
2020
).
103.
S.
Kim
,
J.
Noh
,
G. H.
Gu
,
A.
Aspuru-Guzik
, and
Y.
Jung
, “
Generative adversarial networks for crystal structure prediction
,”
ACS Cent. Sci.
6
,
1412
(
2020
).
104.
I.-H.
Lee
and
K. J.
Chang
, “
Crystal structure prediction in a continuous representative space
,”
Comput. Mater. Sci.
194
,
110436
(
2021
).
105.
Y.
Zhao
,
M.
Al-Fahdi
,
M.
Hu
,
E. M. D.
Siriwardane
,
Y.
Song
et al, “
High-throughput discovery of novel cubic crystal materials using deep generative neural networks
,”
Adv. Sci.
8
,
2100566
(
2021
).
106.
T.
Long
,
N. M.
Fortunato
,
I.
Opahle
,
Y.
Zhang
,
I.
Samathrakis
et al, “
Constrained crystals deep convolutional generative adversarial network for the inverse design of crystal structures
,”
npj Comput. Mater.
7
,
66
(
2021
).
107.
Z.
Ren
,
S. I. P.
Tian
,
J.
Noh
,
F.
Oviedo
,
G.
Xing
et al, “
An invertible crystallographic representation for general inverse design of inorganic crystals with targeted properties
,”
Matter
5
,
314
(
2022
).
108.
Y.
Zhao
,
E. M. D.
Siriwardane
,
Z.
Wu
,
N.
Fu
,
M.
Al-Fahdi
et al, “
Physics guided deep learning for generative design of crystal materials with symmetry constraints
,”
npj Comput. Mater.
9
,
38
(
2023
).
109.
J.
Park
,
A. P. S.
Gill
,
S. M.
Moosavi
, and
J.
Kim
, “
Inverse design of porous materials: A diffusion model approach
,”
J. Mater. Chem. A
12
,
6507
(
2024
).
110.
M.
Alverson
,
S. G.
Baird
,
R.
Murdock
,
E. S.-H.
Ho
et al, “
Generative adversarial networks and diffusion models in material discovery
,”
Digit. Discov.
3
,
62
(
2024
).
111.
O.
Boyar
,
Y.
Gu
,
Y.
Tanaka
,
S.
Tonogai
,
T.
Itakura
et al, “
Crystal-LSBO: Automated design of de novo crystals with latent space bayesian optimization
,” arXiv:2405.17881 (
2024
).
112.
A.
Sinha
,
S.
Jia
, and
V.
Fung
, “
Representation-space diffusion models for generating periodic materials
,” arXiv:2408.07213 (
2024
).
113.
R.
Zhu
,
W.
Nong
,
S.
Yamazaki
, and
K.
Hippalgaonkar
, “
WyCryst: Wyckoff inorganic crystal generator framework
,”
Matter
7
,
3469
(
2024
).
114.
T.
Xie
,
X.
Fu
,
O.-E.
Ganea
,
R.
Barzilay
, and
T.
Jaakkola
, “
Crystal diffusion variational autoencoder for periodic material generation,” in International Conference on Learning Representations (ICLR) (OpenReview.net,
2022).
115.
S.
Wang
,
X.
Guo
, and
L.
Zhao
, “
Deep generative model for periodic graphs
,” arXiv:2201.11932 (
2022
).
116.
M.
Yang
,
K.
Cho
,
A.
Merchant
,
P.
Abbeel
,
D.
Schuurmans
et al, “
Scalable diffusion for materials generation
,” in
Advances in Neural Information Processing Systems (NeurIPS)
(
MIT Press
,
2023
).
117.
A.
Klipfel
,
Y.
Frégier
,
A.
Sayede
, and
Z.
Bouraoui
, “
Unified model for crystalline material generation
,” arXiv:2306.04510 (
2023
).
118.
R.
Jiao
,
W.
Huang
,
P.
Lin
,
J.
Han
,
P.
Chen
et al, “
Crystal structure prediction by joint equivariant diffusion
,” in
Advances in Neural Information Processing Systems (NeurIPS)
(
MIT Press
,
2023
).
119.
R.
Jiao
,
W.
Huang
,
Y.
Liu
,
D.
Zhao
, and
Y.
Liu
, “
Space group constrained crystal generation,” in International Conference on Learning Representations (
ICLR)
(OpenReview.net, 2024)
.
120.
C.
Zeni
,
R.
Pinsler
,
D.
Zügner
,
A.
Fowler
,
M.
Horton
et al, “
A generative model for inorganic materials design
,”
Nature
(published online)
(
2025
).
121.
B. K.
Miller
,
R. T. Q.
Chen
,
A.
Sriram
, and
B. M.
Wood
, “
FlowMM: Generating materials with riemannian flow matching
,” arXiv:2406.04713 (
2024
).
122.
X.
Luo
,
Z.
Wang
,
P.
Gao
,
J.
Lv
,
Y.
Wang
et al, “
Deep learning generative model for crystal structure prediction
,”
npj Comput. Mater.
10
,
254
(
2024
).
123.
X.
Luo
,
Z.
Wang
,
J.
Lv
,
L.
Wang
,
Y.
Wang
et al, “
CrystalFlow: A flow-based generative model for crystalline materials
,” arXiv:2412.11693 (
2024
).
124.
X.-Q.
Han
,
Z.
Ouyang
,
P.-J.
Guo
,
H.
Sun
,
Z.-F.
Gao
et al, “
InvDesFlow: An AI search engine to explore possible high-temperature superconductors
,” arXiv:2409.08065 (
2024
).
125.
P.
Lin
,
P.
Chen
,
R.
Jiao
,
Q.
Mo
,
J.
Cen
et al, “
Equivariant diffusion for crystal structure prediction,” in International Conference on Machine Learning (
ICML)
(ACM, 2024)
.
126.
T. M.
Nguyen
,
S.
Tawfik
,
r.
Tran
,
S.
Gupta
,
S.
Rana
et al, “
Hierarchical GFlownet for crystal structure generation
,” in
Advances in Neural Information Processing Systems (NeurIPS)
(
MIT Press
,
2023
).
127.
D.
Levy
,
S. S.
Panigrahi
,
S.-O.
Kaba
,
Q.
Zhu
,
M.
Galkin
et al, “
SymmCD: Symmetry-Preserving crystal generation with diffusion models
,” in
AI for Accelerated Materials Disc
overy -
NeurIPS (MIT Press
,
2024
).
128.
C.-Y.
Ye
,
H.-M.
Weng
, and
Q.-S.
Wu
, “
Con-CDVAE: A method for the conditional generation of crystal structures
,”
Comput. Mater. Today
1
,
100003
(
2024
).
129.
H.
Xiao
,
R.
Li
,
X.
Shi
,
Y.
Chen
,
L.
Zhu
et al, “
An invertible, invariant crystal representation for inverse design of solid-state materials using generative deep learning
,”
Nat. Commun.
14
,
7027
(
2023
).
130.
L. M.
Antunes
,
K. T.
Butler
, and
R.
Grau-Crespo
, “
Crystal structure generation with autoregressive large language modeling
,”
Nat. Commun.
15
,
10570
(
2024
).
131.
N.
Sitapure
and
J. S.-I.
Kwon
, “
CrystalGPT: Enhancing system-to-system transferability in crystallization prediction and control using time-series-transformers
,” arXiv:2306.03099 (
2023
).
132.
Y.
Chen
,
X.
Wang
,
X.
Deng
,
Y.
Liu
,
X.
Chen
et al, “
MatterGPT: A generative transformer for multi-property inverse design of solid-state materials
,” arXiv:2408.07608 (
2024
).
133.
B.
Wang
,
Z.
Xu
,
Z.
Han
,
Q.
Nie
,
H.
Xiao
et al, “
SLICES-PLUS: A crystal representation leveraging spatial symmetry
,” arXiv:2410.22828 (
2024
).
134.
N.
Kazeev
,
W.
Nong
,
I.
Romanov
,
R.
Zhu
,
A.
Ustyuzhanin
et al, “
Wyckoff transformer: Generation of symmetric crystals
,” arXiv:2503.02407 (
2025
).
135.
Z.
Cao
,
X.
Luo
,
J.
Lv
, and
L.
Wang
, “
Space group informed transformer for crystalline materials generation
,” arXiv:2403.15734 (
2024
).
136.
N.
Gruver
,
A.
Sriram
,
A.
Madotto
,
A. G.
Wilson
,
C. L.
Zitnick
et al, “
Fine-tuned language models generate stable inorganic materials as text,” in International Conference on Learning Representations (
ICLR)
(OpenReview.net, 2024)
.
137.
A.
Sriram
,
B. K.
Miller
,
R. T. Q.
Chen
, and
B. M.
Wood
, “
FlowLLM: Flow matching for material generation with large language models as base distributions
,” in
Advances in Neural Information Processing Systems (NeurIPS)
(
MIT Press
,
2024
).
138.
K.
Choudhary
, “
AtomGPT: Atomistic generative pretrained transformer for forward and inverse materials design
,”
J. Phys. Chem. Lett.
15
,
6909
(
2024
).
139.
T.
Mohanty
,
M.
Mehta
,
H. M.
Sayeed
,
V.
Srikumar
, and
T. D.
Sparks
, “
CrysText: A generative AI approach for text-conditioned crystal structure generation using LLM
,” ChemRxiv:10.26434/chemrxiv-2024-gjhpq (
2024
).
140.
K.
Yan
,
X.
Li
,
H.
Ling
,
K.
Ashen
,
C.
Edwards
et al, “
Invariant tokenization of crystalline materials for language model enabled generation
,” in
Advances in Neural Information Processing Systems (NeurIPS)
(
MIT Press
,
2024
).
141.
D.
Maoutsa
,
S.
Reich
, and
M.
Opper
, “
Interacting particle solutions of Fokker–Planck equations through gradient–log–density estimation
,”
Entropy
22
,
802
(
2020
).
142.
Y.
Song
,
J.
Sohl-Dickstein
,
D. P.
Kingma
,
A.
Kumar
,
S.
Ermon
et al, “
Score-based generative modeling through stochastic differential equations
,” arXiv:2011.13456 (
2021
).
143.
L.
Yang
,
Z.
Zhang
,
S.
Hong
,
R.
Xu
,
Y.
Zhao
et al, “
Diffusion models: A comprehensive survey of methods and applications
,” arXiv:2209.00796 (
2022
).
144.
S.
Zhai
,
R.
Zhang
,
P.
Nakkiran
,
D.
Berthelot
,
J.
Gu
et al, “
Normalizing flows are capable generative models
,” arxiv:2412.06329 (
2024
).