This paper delves into the transformative role of Machine Learning (ML) and Artificial Intelligence (AI) in materials science, spotlighting their capability to expedite the discovery and development of newer, more efficient, and stronger compounds. It underscores the shift from traditional, resource-intensive approaches toward data-driven methodologies that leverage large datasets to predict properties, identify new materials, and optimize synthesis conditions with a satisfactory level of accuracy. Highlighting various techniques, including supervised, unsupervised, and reinforcement learning, alongside deep learning potential, the chapter presents case studies and applications ranging from predicting stress points in stochastic fields to optimizing thermal protection systems for spacecraft re-entry. It also explores the challenges and future directions, emphasizing the need for integrating experimental validations and developing tailored algorithms to overcome data and computational constraints. The narrative showcases ML and AI’s promise in revolutionizing material discovery, paving the way for innovative solutions in science and engineering.

In the field of materials science, researchers typically rely on experiments and simulations to develop and produce new materials. This approach requires significant resources and often involves a certain degree of luck and skills. Recent discoveries in data-driven knowledge fields have given researchers the opportunity to use what was defined in Material Gnosis in Refs. 1 and 2 as the fourth paradigm of science based upon data driven methodologies. The purpose of using big data is to improve the efficiency and accuracy of their work, as well as produce more sustainable and durable results with less computational power and fewer testing facilities.

In May 2016, Nature published3 a paper that explained how failed experiments could help the discovery of new materials. The accuracy of these predictions outperformed that of expert chemists, leading researchers to believe that Machine Learning (ML) could significantly alter the traditional approach to material discovery and open new possibilities for inventing materials. Since then, the use of Machine Learning (ML) and Artificial Intelligence (AI) in materials science has continued to gain momentum. Researchers are using these technologies to quickly predict candidate materials based on desired performance requirements, which can then be developed and tested using computer simulations. This approach can significantly speed up the development process and improve efficiency (Figs. 1-4).

FIG. 1.

Four paradigms of science.3 

FIG. 1.

Four paradigms of science.3 

Close modal
FIG. 2.

Principal component analysis.4,5

FIG. 2.

Principal component analysis.4,5

Close modal
FIG. 3.

High stress clustering identifications using PCA (principal component analysis).6,7

FIG. 3.

High stress clustering identifications using PCA (principal component analysis).6,7

Close modal
FIG. 4.

Surrogate model to study stress distribution.8,9

FIG. 4.

Surrogate model to study stress distribution.8,9

Close modal

In recent years, machine learning and artificial intelligence have become increasingly popular in materials science. One reason for this is the significant potential that these technologies hold for improving the efficiency and accuracy of material discovery and development. Traditionally, materials scientists have relied on experiments and simulations to identify and produce new materials. However, this approach can be time-consuming and resource-intensive, requiring significant amounts of trial and error to achieve the desired results. Machine learning and AI offer an alternative approach that can speed up the discovery and development process.10 

For example, researchers can use machine learning algorithms to analyze large amounts of experimental data, identify patterns and correlations, and predict new results based on this information. These predictions can then be tested using computer simulations, which can save significant time and resources compared to traditional experimental methods. Another advantage of machine learning and AI is that they can help overcome some of the limitations of traditional material discovery methods. For example, machine learning algorithms can quickly analyze large datasets and identify new materials with desired properties that may have been overlooked using traditional approaches. This can help open up new avenues for material discovery and development. Overall, the use of machine learning and AI in materials science has the potential to revolutionize the field, leading to breakthroughs in many different areas. As researchers continue to explore the possibilities of these technologies, it is likely that we will see new and exciting advances in material discovery and development in the years to come.11 

The discovery of new materials with desirable properties is a critical step in advancing various fields of science and engineering, such as electronics, energy, and medicine. Traditional methods of material discovery are time-consuming and expensive, as they involve synthesizing and characterizing numerous samples. Machine learning has emerged as a powerful tool in material discovery, as it can accelerate the process by predicting material properties, identifying new materials, and optimizing synthesis conditions. In this section, an overview of the different techniques that can be used in material discovery will be provided. Advantages, limitations, and applications in predicting material properties, identifying new materials, and optimizing synthesis conditions.

The three main types of machine learning techniques are as follows: supervised learning, unsupervised learning, and reinforced learning.

  • Supervised learning is a technique in which a machine learning model is trained on a labeled dataset to predict the output of new, unseen data. In material discovery, supervised learning can be used to predict material properties based on the chemical composition, crystal structure, and processing conditions. The labeled dataset may include data from experiments, simulations, or a combination of both. Examples of supervised learning algorithms include regression, support vector machines, decision trees, and neural networks.4 

  • Unsupervised learning is a technique in which a machine learning model is trained on an unlabeled dataset to identify patterns and relationships in the data. In material discovery, unsupervised learning can be used to cluster similar materials based on their properties or to identify outliers that exhibit unique properties. Examples of unsupervised learning algorithms include clustering, principal component analysis (PCA), and autoencoders.1 

  • Reinforcement learning is a technique in which a machine learning model learns to make decisions based on feedback from the environment. In material discovery, reinforcement learning can be used to optimize synthesis conditions by selecting the best set of parameters to maximize a desired property. Examples of reinforcement learning algorithms include Q-learning and policy gradients.10 

1. Predicting material properties

Machine learning can be used to predict material properties such as electronic, magnetic, and mechanical properties based on the chemical composition, crystal structure, and processing conditions. This can be achieved by training a model on a labeled dataset that includes experimental or simulated data. For example, machine learning models have been used to predict the bandgap of materials, which is a critical property for applications in electronics and optoelectronics. Models based upon these methodologies have also been used to predict the mechanical properties of materials, such as the Young’s modulus and Poisson’s ratio, which are important for designing materials with specific mechanical properties.5 

2. Identify new materials

Machine learning can be used to identify new materials with desirable properties by screening a large database of materials. This can be achieved by training a machine learning model on a labeled dataset that includes data on the properties of known materials. The ML model can then be used to predict the properties of new materials based on their chemical composition and crystal structure. For example, ML models have been used to predict the properties of new battery materials, which can potentially improve the energy density and safety of batteries.2 

3. Optimizing conditions

Machine learning can be used to optimize synthesis conditions by selecting the best set of parameters to maximize a desired property. This can be achieved by training an ML model on a labeled dataset that includes data on the properties of materials synthesized under different conditions. The ML model can then be used to predict the properties of materials synthesized under new conditions. For example, ML models have been used to optimize the synthesis of catalysts by selecting the best set of parameters to maximize their activity and selectivity.12 

Deep learning (DL) is a subset of machine learning that has the potential to revolutionize the field of material discovery by accelerating the process of identifying new materials with desirable properties. In this section, we will explore the potential of such methodology in material discovery, including an overview of the different techniques, their advantages and limitations, and their applications in predicting material properties, identifying new materials, and optimizing synthesis conditions. Deep learning techniques are neural network architectures that consist of multiple layers of interconnected nodes that can learn and make predictions based on input data. The most commonly used DL architectures in material discovery are convolutional neural networks (CNNs) and recurrent neural networks (RNNs).6,13

CNNs are used for image-based data and are designed to automatically learn and extract features from image data. In material discovery, CNNs can be used to predict material properties based on images of the material’s microstructure, such as the crystal structure and morphology. RNNs are used for sequential data and are designed to process data with temporal dependencies, such as time-series data. In material discovery, RNNs can be used to predict the properties of materials based on the synthesis conditions, which are often represented as time-series data.

1. Advantages and limitations of deep learning

Deep learning has several advantages over traditional machine learning techniques, such as the ability to automatically learn features from raw data without the need for feature engineering. DL also has the potential to achieve higher accuracy and efficiency in predicting material properties compared to traditional machine learning techniques. However, DL also has several limitations, such as the requirement for large amounts of training data and the need for significant computational resources. In addition, DL models can be difficult to interpret, which can limit their usefulness in providing insights into the underlying relationships between material properties and their predictors.7,13

Deep learning has several applications in material discovery, including predicting material properties, identifying new materials, and optimizing synthesis conditions. For example, deep learning can be used to predict the mechanical properties of materials based on images of their microstructure, which can potentially accelerate the design of new materials with specific mechanical properties. Deep learning can also be used to identify new materials with desirable properties by predicting the properties of hypothetical materials that have not yet been synthesized. In addition, deep learning can be used to optimize synthesis conditions by predicting the properties of materials synthesized under different conditions and identifying the optimal conditions for achieving specific material properties.6,7 Despite its potential, deep learning also faces several challenges in material discovery. One challenge is the interpretability of DL models, which can limit their usefulness in providing insights into the underlying relationships between material properties and their predictors. Another challenge is the requirement for large amounts of training data, which may not be available for certain materials or properties. In addition, DL models require significant computational resources, which can limit their scalability and accessibility. Using this new methodology has the potential to revolutionize the field of material discovery by accelerating the process of identifying new materials with desirable properties. The use of deep learning in predicting material properties, identifying new materials, and optimizing synthesis conditions has several advantages and limitations. However, challenges such as model interpretability and the requirement for large amounts of training data must be addressed to fully realize the potential of deep learning in material discovery.

Stochastic fields are random fields that exhibit random variations, such as the wind load on a bridge or the pressure distribution in a pipeline. Predicting high stress points in such fields is critical for ensuring the safety and reliability of structures and components. Traditional methods for predicting high stress points in stochastic fields involve computationally expensive simulations and the use of probabilistic models.14 In this section, an algorithm is proposed that can efficiently predict high stress points in a stochastic field by combining two known statistical methods and creating a simplified neural network. Considering a continuous field with a finite number of random variables that follow a Gaussian distribution of probability and assuming the principle of continuous entropy is applied, the function of stress σ(x) will be described by the general probability density function,
(1)
where
  • mx is the mean function, often assumed to be zero or another known function based upon prior knowledge.

  • k(x, x′) is the covariance function (kernel), which defines the covariance between any two points in the field, reflecting how stress values correlate with each other across the field.

The next step is to update the function using observed data, which could be the data used by previous stress distribution plots. Let X be the matrix of stress distribution plots and y be the corresponding observed stress value, the posterior distribution σx function is given by
(2)
with mean and variance given by
(3)
(4)
where
  • k is the mean vector of covariances between the new point x* and all points in X.

  • K is the covariance matrix for points in X.

  • I is the identity matrix.

  • varn2 is the variance of the observation noise.

High stress points are predicted by leveraging on the principle of maximum entropy, which involves finding points where the posterior variance is maximized, hence the high stress points σxhigh will be given by
(5)
Eq. (5) generates a series of high stress points based on probability distributions and observed data. Such methodology could be enhanced by performing some data preprocessing using Principal Component Analysis (PCA), which is a statistical technique used to reduce the dimensionality of a dataset while retaining most of the variation in the dataset. It identifies the directions (principal components) that maximize the variance in the data. Prior to apply Gaussian regression, a hypothesized dataset that contains data from known experiments, for example, could be reduced using PCA. If the stochastic field data are high-dimensional (e.g., derived from multiple sensors or simulation parameters), PCA can reduce the number of dimensions to the most significant ones. This step simplifies the model by focusing on the principal components that account for most of the variance in the stress data. PCA can also transform the original variables into a new set of uncorrelated features (principal components). These features can then be used as inputs to the GPR model, potentially improving the efficiency and accuracy of the stress prediction.

The resulting datasets or stress distribution plots can be further reduced by training a simplified neural network that can increase the accuracy of stress prediction levels and reduce computational costs.4 The supervised technique of clustering could help further enhance the results by linking together the results of multiple probabilistic simulations. The picture below shows how the k-mean clustering application led to the identification of key stress points in an I-beam that underwent cyclical loading.15 

Such neural networks, combined with clustering, can be used to predict high stress points in stochastic fields. Although, at this point in time, there are not enough results to verify the accuracy of the models, using ML has certainly led to less computational power usage and more accessibility of results.5 If such algorithms need to be implemented within a data or ML operations pipeline, they would need evaluation using synthetic or dummy data as well as real world data. Such AI driven mythologies have potential applications in engineering and mechanics, where predicting high stress points is crucial for maintaining the structural integrity and safety of various systems. For example, in a case study where there was a need to estimate high stress points within a wind turbine case, this Gaussian based stochastic algorithm was trained using stress data from a historical series of failure data about wind turbines and, once applied within the FEA based model, gave a fairly accurate prediction of the locations of the high stress points using less computational power. Results were validated using experimental data, and the algorithm was shown to be highly accurate in predicting high-stress points in wind turbine blades.16 In another case study, the need was to estimate high-stress points in an underground pipeline. The data regarding previous pipeline inspections were used as training data in a convoluted neural network pipeline similar to the one described above, and the prediction of high-stress points was compared with real data that were acquired in subsequent inspections, showing a good degree of accuracy.17 

Machine learning can be applied in parallel and isolation, but also in sequential logic, meaning that the output or dataset generated by one algorithm could be further included within a different pipeline or even used in conjunction with traditional computational methods to generate a new way of computing simulation called surrogate learning or surrogate modeling.18 In the evolving landscape of material science, the integration of ML algorithms has marked a significant shift toward more innovative and efficient approaches to studying and optimizing materials for extreme conditions. One key application of the surrogate model has been carried out by Zuccarini et al. and involved the analysis of stress distribution in Ultra-High Temperature Ceramics (UHTCs), critical for aerospace applications. The model was built upon a combination, also known as surrogacy, of traditional computational fluid dynamic and deep learning to estimate the stress distribution in a Zirconium Diboride and Silica Carbine (ZrB2–SiC) based Thermal Protection System (TPS) under re-entry conditions.9 

UHTCs such as ZrB2/SiC are indispensable for safeguarding spacecraft against the intense thermal and mechanical stresses encountered during hypersonic re-entry into Earth’s atmosphere, where temperatures can soar to 5000 K and pressures can reach up to 205 MPa. The study embarked on a comprehensive methodology that combined finite element modeling and computational fluid dynamics to simulate the innovative use of the Keras library for DL, underscoring both feasibility as well as a significant reduction of computational time and resource demands, thereby presenting a scalable and efficient tool for predicting complex material behaviors under extreme conditions.8 Findings from the study revealed a maximum stress threshold indicative of the ablative behavior of the TPS material, with specific attention drawn to the thermal stress concentration at boundary layer separation points. These insights are vital for the design and optimization of TPS materials, ensuring spacecraft can withstand the rigors of re-entry. This example underscores the transformative potential of surrogate modeling in material science, particularly in the study of UHTCs for aerospace applications. By enabling more accurate and efficient predictions of material behavior under extreme conditions, such algorithms are not only optimizing the design and testing of thermal protection systems but also paving the way for advancements in the broader field of materials science. Applications in such contexts illustrate a pivotal step toward leveraging computational power to enhance our understanding and innovation in material technologies, marking a significant milestone in the ongoing synergy between materials science and artificial intelligence.7,19

Machine learning and deep learning have the potential to revolutionize the field of material discovery by accelerating the process of identifying new materials with desirable properties. In this section, we will discuss the outlook of machine learning and DL in material discovery, including potential avenues for future research, such as the integration of experimental and computational approaches and the development of algorithms specifically tailored for material discovery.1 While machine learning and deep learning can accelerate the screening of materials, experimental validation is still necessary to confirm the predicted properties of new materials. Therefore, an integrated approach that combines such methodologies with experimental methods, integrated within a stable data driven pipeline, could potentially speed up the process of material discovery while also providing experimental validation of the predicted properties.

For example, machine learning and deep learning could be used to guide the synthesis of new materials, and the properties of the synthesized materials can be characterized using experimental techniques such as x-ray diffraction, scanning electron microscopy, and spectroscopy. Another potential avenue for future research is the development of new machine learning and DL algorithms specifically tailored for material discovery. Traditional machine learning algorithms may not be optimized for the unique challenges associated with material discovery, such as the high dimensionality of the data and the complex relationships between the chemical composition, crystal structure, and properties of materials.20–26 

Therefore, tailored algorithms that are specifically designed for material discovery can potentially improve the accuracy and efficiency of the predictions. For example, Convolutional Neural Networks (CNNs) and recurrent neural networks (RNNs) can be adapted for material discovery by incorporating domain-specific knowledge into the architecture. In conclusion, the integration of experimental and computational approaches and the development of new ML and DL algorithms specifically tailored for material discovery are potential avenues for future research. However, challenges such as data availability and quality must also be addressed to fully realize the potential of these methodologies in the future.4 

The authors have no conflicts to disclose.

Carmine Zuccarini: Writing – original draft (equal). Karthikeyan Ramachandran: Writing – original draft (equal). Doni Daniel Jayaseelan: Writing – review & editing (lead).

The data that support the findings of this study are available from the corresponding author upon reasonable request.

1.
T.
Mueller
,
A.
Gilad Kusne
, and
R.
Ramprasad
, “
Machine learning in materials - recent progress and emerging applications
,” in
Reviews in Computational Chemistery
(
John Wiley & Sons, Inc
,
Baltimore MD
,
2016
), p.
291
.
2.
M. H.
Mobarak
,
M. A.
Mimona
,
M. A.
Islam
,
N.
Hossain
,
F. T.
Zohura
,
I.
Imtiaz
, and
M. I.
Hossain Rimon
, “
Scope of machine learning in materials research—A Review
,”
Appl. Surf. Sci. Adv.
18
,
100523
(
2023
).
3.
R.
LeSar
, “
Materials informatics: An emerging technology for materials development
,”
Stat. Anal. Data Min.: ASA Data Sci. J.
1
,
372
(
2009
).
5.
E.
Ford
et al, “
Machine learning-based accelerated property prediction of two-phase materials using microstructural descriptors and finite element analysis
,”
Comput. Mater. Sci.
191
,
110328
(
2021
).
6.
Stephanie
, Statistics how to (
2018
) Available at: https://www.statisticshowto.com/levenberg-marquardt-algorithm/
7.
K.
Choudhary
et al, “
Recent advances and applications of deep learning methods in materials science
,”
Nature
59
,
8
(
2022
).
8.
V.
Casseau
,
T. J.
Scanlon
,
B.
John
,
D. R.
Emerson
, and
R. E.
Brown
, “
Hypersonic simulations using open-source CFD and DSMC
,”
AIP Conf. Proc.
1786
,
050006
(
2016
) Strathclyde, UK.
9.
C.
Zuccarini
et al, “
Stress distribution analysis in zirconium diboride and silica carbide (ZrB2-SiC) based TPS under hypersonic flight conditions using machine learning driven approach
,” in
Ultra-High Temperature Ceramics: Materials For Extreme Environment Applications
,
2021
.
10.
L.
Alzubaidi
et al, “
Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions
,”
J. Big Data
8
,
53
(
2021
).
11.
P.
Raccuglia
et al, “
Machine-learning-assisted materials discovery using failed experiments
,”
Nature
533
,
73
76
(
2016
).
12.
Z.
Wang
et al, “
The role of machine learning in carbon neutrality: Catalyst property prediction, design, and synthesis for carbon dioxide reduction
,”
eScience
3
(
4
),
100136
(
2023
).
13.
J.
Behler
, “
Neural network potential-energy surfaces in chemistry: A tool for large-scale simulations
,”
Phys. Chem. Chem. Phys.
17930
,
13
(
2011
).
14.
V. Q.
Tran
, “
Using artificial intelligence approach for investigating and predicting yield stress of cemented paste backfill
,”
Sustainability
15
,
2892
(
2023
).
15.
M.
Granzner
,
A.
Strauss
,
M.
Reiterer
,
M.
Cao
, and
D.
Novák
, “
Data-driven condition assessment and life cycle analysis methods for dynamically and fatigue-loaded railway infrastructure components
,”
Infrastructures
8
(
11
),
162
(
2023
).
16.
J.
Xiao
et al, “
Prediction of wind turbine blade icing fault based on selective deep ensemble model
,”
Knowl.-Based Syst.
242
,
108290
(
2022
).
17.
M.
Zhang
et al, “
A data-driven based method for pipeline additional stress prediction subject to landslide geohazards
,”
Sustainability
14
(
19
),
11999
(
2022
).
18.
F.
Qin
et al, “
Deep-learning-based surrogate model for fast and accurate simulation in pipeline transport
,”
Front. Energy Res.
10
,
50
88
(
2022
).
19.
K.
Hansen
et al, “
Machine learning predictions of molecular properties: Accurate many-body potentials and nonlocality in chemical space
,”
J. Phys. Chem. Lett.
6
,
2326
(
2015
).
20.
E.
Burns
and
J.
Burke
, Tech target, (
2021
) Available at: https://searchenterpriseai.techtarget.com/definition/neural-network
21.
A. J.
Litta
,
S. M.
Idicula
, and
U. C.
Mohanty
, “
Artificial neural network model in prediction of meteorological parameters during premonsoon thunderstorms
,”
Int. J. Atmos. Sci.
13
,
161
184
(
2013
).
22.
S.
Samarasinghe
,
Neural Networks for Applied Sciences and Engineering
(
Auerbach Publications, Taylor and Francis group
,
New York
,
2006
).
23.
K.
Abhishek
,
M.
Singh
,
S.
Ghosh
, and
A.
Anand
, “
Weather forecasting model using artificial neural network
,”
Proc. Technol.
4
,
311
318
(
2012
).
24.
K.
Rajan
, “
Materials informatics
,”
Mater. Today
8
(
10
),
38
(
2005
).
25.
C.
White
,
M. K.
Borg
,
T. J.
Scanlon
,
S. M.
Longshaw
,
B.
John
,
D. R.
Emerson
, and
J. M.
Reese
, “
dsmcFoam+: An openFOAM based direct simulation Monte Carlo solver
,”
Comput. Phys. Commun.
224
(
1
),
22
43
(
2018
).
26.
A. P.
Bartók
,
M. C.
Payne
,
R.
Kondor
, and
G.
Csányi
, “
Gaussian approximation potentials: The accuracy of quantum mechanics
,”
Phys. Rev. Lett.
136403
,
104
(
2010
).