Uploaded by XY xx

AI in Materials Design, Discovery & Manufacturing: A Review

advertisement
Computational Materials Science 235 (2024) 112793
Contents lists available at ScienceDirect
Computational Materials Science
journal homepage: www.elsevier.com/locate/commatsci
Full length article
AI methods in materials design, discovery and manufacturing: A review
Ioannis Papadimitriou ∗, Ilias Gialampoukidis, Stefanos Vrochidis, Ioannis Kompatsiaris
Information Technology Institute, Centre for Technology Hellas, 6th km Charilaou Thermis, Thessaloniki, 57001, Greece
ARTICLE
INFO
ABSTRACT
Keywords:
Artificial intelligence
Materials design
Materials discovery
In the advent of the digital revolution, Artificial Intelligence (AI) has emerged as a pivotal tool in various
domains, including materials design and discovery. This paper provides a comprehensive review of the AI
methodologies integrated within this context, encompassing materials informatics, density functional theory,
molecular dynamics, and finite elements analysis. We further delve into the transformative role of AI within
process engineering, manufacturing, and industry 4.0, with a focus on manufacturing process optimization
techniques. Highlighting the importance of active learning, self-correcting processing, and digital twins in
the era of smart manufacturing, this review underscores the impact of big data and data quality. The paper
provides an insight into the challenges and future prospects, pointing towards the tremendous potential AI
holds for revolutionizing the field of materials science.
1. Introduction
to identify optimal material compositions and structures for specific
applications. For example, AI has been used to design materials for
energy storage, such as batteries [8] and supercapacitors [9], metallic
alloys [10], or catalysts [11] among others. In addition to predicting
and optimizing material properties, AI has also been applied in the
discovery of new materials. Through the use of ML (ML) algorithms,
researchers have been able to identify promising material candidates
from vast databases of known materials and predict their potential
properties. Overall, the state of the art in AI use in materials science and
engineering is rapidly advancing, with significant progress being made
in the ability to predict, optimize, and discover new materials. This has
the potential to greatly impact the field, leading to more efficient and
cost-effective materials design and discovery.
Industry 4.0, also known as the Fourth Industrial Revolution, refers
to the current trend of automation and data exchange in manufacturing technologies, including the use of AI and the Internet of Things
(IoT) [12]. In Industry 4.0, AI is used to automate and optimize various
tasks in Manufacturing Processes (MP), such as predicting equipment
failures, optimizing production schedules, and identifying inefficiencies
in the production line. AI is also used to analyze large amounts of
data generated by the IoT sensors in manufacturing plants, to identify
patterns and trends that can improve efficiency and reduce costs [13].
Some examples of AI applications in Industry 4.0 include predictive
maintenance (using AI to analyze data from sensors on equipment to
predict when maintenance is required [14], reducing downtime and
improving efficiency), quality control (inspection and classification of
products on the production line, improving the accuracy and speed of
During the recent years, there have been breakthroughs of AI methods used in a range of fields. One of the first fields to experience
said breakthroughs was Computer Vision technology (CV). The use
of Convolutional Neural Networks (CNNs) enabled the creation of
models that can be used for tasks such as image recognition, object
detection, semantic segmentation and image generation. Apart from CV
which has been dominating the interest in terms of AI methods use
and development, the emergence of large Natural Language Processing
(NLP) models marked an era of rapid advances in various tasks, namely
language modeling, machine translation, sentiment analysis and text
generation, during the recent years. Nowadays there is a vast range of
major industries that have seen huge advancements with the use of AI
methods, including healthcare, logistics and transportation, commerce,
food technology, banking and entertainment.
The use of Artificial Intelligence (AI) in materials science and engineering has been rapidly increasing in recent years, with advancements
in ML techniques and the growing availability of high-quality data.
This has led to significant progress in the ability to predict material
properties, optimize material design, and accelerate discovery of new
materials [1,2]. Another key area where AI has been applied is in the
prediction of material properties. ML algorithms have been used to
predict a wide range of mechanical, thermal, and electrical properties
of materials, such as the Young’s modulus [3,4], melting temperature [5,6], and conductivity [7]. Recently, AI has been also utilized
is in the optimization of material design. By incorporating constraints
such as cost and environmental impact, ML algorithms have been able
∗ Corresponding author.
E-mail address: i.papadimitriou@iti.gr (I. Papadimitriou).
https://doi.org/10.1016/j.commatsci.2024.112793
Received 27 July 2023; Received in revised form 10 December 2023; Accepted 5 January 2024
Available online 19 January 2024
0927-0256/© 2024 Elsevier B.V. All rights reserved.
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
Fig. 1. Conceptual overview of AI/ML methods in Materials Science.
quality control processes and process optimization (data analysis from
the production line and bottleneck/inefficiency identification, allowing
for more efficient use of resources). Before delving deeper, a conceptual
overview of AI/ML methods and techniques discussed in the paper, as
well as their applications in materials science is shown in Fig. 1.
main limitations of using ML in materials design is the availability of
high-quality data [22]. The accuracy of ML models depends heavily
on the quality and quantity of data used for training. Henceforth,
when the term ML models is used, it will denote an umbrella typically
consisting of classic algorithms such as logistic regression, decision
trees, Support Vector Machines (SVM) or K-Nearest-Neighbors (KNN).
In addition, these methods are typically black boxes, which means it
can be challenging to interpret their predictions and understand how
they arrive at their conclusions.
DL is a subset of ML that has shown great promise in materials design. DL models can learn complex patterns in data and make
predictions with high accuracy [23–26]. DL has been used to predict
material properties, discover new materials, and optimize the synthesis
of materials. One of the key benefits of using DL in materials design is
its ability to handle large and complex datasets. DL models can learn
complex correlations between material properties and their underlying
atomic structures. DL has also shown promising results in the design
of new materials with desired properties, where it can generate new
materials based on a set of desired properties. However, one of the
main limitations of using DL in materials design is its high computational cost. DL models require significant computational resources, and
training large models can be time-consuming and expensive.
In addition to ML and DL, other AI techniques have been used in
materials design. For example, evolutionary algorithms (EA) have been
used to optimize the synthesis of materials [27]. EA can be used to
search for the optimal synthesis conditions for a given material, such
as temperature and pressure. One of the key benefits of using EA in
materials design is its ability to handle complex optimization problems
with a large number of variables. However, one of the main limitations
2. Materials design and discovery
The discovery of new materials with desired properties is a key
challenge in many scientific and engineering fields. The traditional
trial-and-error approach to materials design is time-consuming and
resource-intensive, and there is a growing interest in using artificial
intelligence (AI) techniques to accelerate the discovery of new materials [15]. In this survey paper, we review recent developments in
the application of AI to materials design and discovery, including
ML, Deep Learning (DL), and other AI techniques. In recent years,
there have been several studies on the applications of AI in materials
science, including empirical interatomic potential development, MLbased potential, property predictions, and molecular discoveries using
Generative Adversarial Networks (GAN) [16].
ML has shown tremendous potential in accelerating the discovery
and design of new materials. ML models can be trained using a large
dataset of known materials and their properties to predict the properties of new materials [17–21]. One of the major benefits of using
ML in materials design is that it can significantly reduce the time and
cost involved in experimental synthesis and characterization. ML can
also help identify new materials with desired properties that have not
yet been synthesized. This can lead to the discovery of new materials
that have potential applications in various fields. However, one of the
2
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
to selection of best practice parameters for the calculation/simulation.
Epistemic uncertainty can be reduced through further research and
experimentation, or by using more accurate models and better data.
On the other hand, aleatory uncertainty is inherent to the system being
studied and arises from the inherent randomness and variability of
the system. Even when conditions and materials are identical, each
instantiation will differ microscopically. This type of uncertainty cannot
be reduced through further research or better models and must be accounted for in the analysis of results, while a potential treatment would
be for example the description in terms of a probability distribution,
rather than the traditional average behavior.
of using EA in materials design is its inability to handle uncertainty
and noise in the data. EA can be sensitive to small changes in the input
parameters, and it can be challenging to find the global optimum for
complex problems.
Another AI technique that has been used in materials design is
Natural Language Processing (NLP) [28]. NLPs can be used to analyze
scientific literature and extract information about materials and their
properties [29]. This information can then be used to train ML models
or generate new materials with desired properties. One of the benefits
of using an NLP in materials design is its ability to handle unstructured
data such as scientific literature. NLPs can also help identify new
research directions and potential applications of materials. However,
one of the main limitations of using NLPs in materials design is the
quality of the available literature. NLP models depend heavily on the
quality and quantity of data used for training, and the accuracy of the
models can be affected by biases and errors in the data.
Utilizing AI for research purposes is gaining momentum, with platforms such as OpenAI’s ChatGPT playing a pivotal role in this transformation [30]. This shift is evident at the University of Wisconsin–
Madison, where researchers have harnessed ML to streamline the discovery of new materials [31]. Moreover, they have expanded the
application of ChatGPT to include the extraction of valuable data from
extensive scientific literature, a task traditionally known for its laborious nature. The research team, led by researchers Morgan and Polak,
devised a strategy whereby ChatGPT would scrutinize scientific papers
at a sentence level, isolating data of relevance. The gathered information was subsequently organized in a tabular format, a process that
delivered an accuracy rate of close to 90%. To achieve full automation,
the researchers embarked on ‘prompt’ engineering, a method involving
a series of questions that instructed ChatGPT to extract and subsequently validate the data. This approach prompted the AI to revisit the
information, enabling it to identify any potential inaccuracies. Morgan
and Polak’s pioneering approach marks a departure from traditional
programming methods. It emphasizes the use of natural language as
the primary tool for AI interaction, thereby making the technology
more accessible and reducing the need for deep technical knowledge.
Importantly, the adoption of AI in research does not render researchers
obsolete but rather equips them with the means to undertake previously
unachievable projects due to limitations in time, funding, or personnel.
Thus, the integration of AI, as exemplified by Morgan and Polak,
not only boosts efficiency but also paves the way for new scientific
exploration opportunities.
Further research is needed to fully realize the potential of AI in
materials design, and to address the challenges and limitations discussed in the literature. Some potential areas for future research include
the development of new AI algorithms and methods for materials
design, the integration of AI with physical models and simulations, the
application of AI to a wider range of materials and applications, and
the development of strategies for addressing ethical and societal issues
in the use of AI in materials design.
AI has become an important tool for materials design, and it offers
significant potential for advancing materials science and engineering.
However, there are also challenges and limitations to the use of AI in
materials design, including the need for high-quality training data, the
difficulty of interpreting and explaining the results of AI algorithms,
and the potential for bias. These challenges will need to be addressed
in order to fully realize the potential of AI in materials design. In this
comprehensive review, we will present the state-of-the-art applications
of artificial intelligence in the field of materials discovery, exploring
the methodologies incorporated in these practices, and addressing the
observed challenges.
Finally, uncertainty is a key concept in modeling. There exist two
types of uncertainty: epistemic and aleatory. Epistemic uncertainty
arises from a lack of knowledge or understanding of the system being
studied, such as incomplete or inaccurate data or an incomplete understanding of the underlying physical phenomena, pertaining mainly
2.1. Materials informatics
Materials informatics is an emerging multidisciplinary field that
combines materials science with information science to facilitate the
discovery, development, and manufacture of new materials [32]. This
convergence of disciplines is enabled by the significant advancements
in computational power, ML, and data storage capabilities seen in
recent years. The field aims to extract valuable insights from large
and complex datasets generated in materials science, effectively transforming the conventional trial-and-error approach into a systematic,
data-driven one [32]. Historically, the development of new materials has been a labor-intensive and time-consuming process. It often
involves numerous experimental iterations to understand the relationships between the material’s structure, processing methods, and properties. Materials informatics has the potential to revolutionize this
process by leveraging computational tools to model these relationships,
thereby reducing the time and cost associated with materials development [32]. Furthermore, materials informatics plays a critical role
in the context of Industry 4.0, which envisions the digital transformation of manufacturing. The ability to predict material properties and
behaviors can significantly enhance the efficiency and quality of MPs,
leading to improved product performance and reduced environmental
impact [32].
The foundation of materials informatics lies in the collection, analysis, and interpretation of massive datasets related to materials [33]. It
fundamentally involves three interconnected components: data generation, feature extraction, and handling the complexity and heterogeneity
of materials data. Data Generation in Materials Science refers to the
collection and preparation of data for analysis. Data can come from various sources such as experimental studies, computational simulations,
and even scientific literature. With advancements in experimental and
computational techniques, materials science is now able to produce vast
amounts of data spanning different scales, from atomic to macroscopic,
and different properties, from mechanical to optical [34].
Feature extraction and selection is a critical step in materials informatics that involves the identification and selection of features (also
known as descriptors) that characterize the materials of interest. Features could include physical, chemical, and structural attributes. The
chosen features should capture the essential properties of the materials
and their relationships with the properties of interest. ML algorithms
are typically used for this purpose [15]. Materials data is often highdimensional, meaning it involves a large number of features for each
sample. This high dimensionality can present challenges in data analysis and interpretation, including issues related to overfitting and the
curse of dimensionality. Techniques such as dimensionality reduction
can help mitigate these problems [35]. Dealing with heterogeneity
and complexity of materials data: Materials data is typically heterogeneous, with different types of data (e.g., text, images, numerical
data) needing to be integrated for analysis. Moreover, the data often
captures complex, non-linear relationships among variables. Addressing
this complexity and heterogeneity requires advanced data integration,
data mining, and ML techniques [32].
In materials informatics, a multitude of techniques are leveraged
to analyze and interpret data. These include various ML strategies,
3
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
transfer learning, natural language processing, and data mining. ML
provides a powerful set of tools for predicting material properties and
discovering underlying patterns in data. These tools are supervised,
unsupervised and semi-supervised learning, transfer learning and NLP.
Materials informatics offers powerful tools and methods that can
be applied to various aspects of materials science and engineering.
By using ML models to predict material properties based on various
features, materials informatics can significantly speed up the process
of materials discovery and design. This approach allows for the exploration of a larger design space in a shorter amount of time, leading
to the discovery of materials with novel or superior properties [17].
Furthermore, materials informatics can be used to develop predictive
models that can forecast the properties of materials based on their
composition and structure. This capability can be leveraged to optimize
material selection for specific applications, potentially improving the
performance and efficiency of the resulting products [32]. Moreover,
by combining domain knowledge with ML techniques, materials informatics can help design experiments that are more likely to yield
useful results. This approach can lead to more efficient use of resources
and faster achievement of research objectives [36]. Finally, in the
manufacturing setting, materials informatics can be used to monitor
the production process and detect any anomalies or deviations from
the expected quality standards. This application can help prevent the
production of defective products, reduce waste, and improve overall
manufacturing efficiency.
Materials informatics represents a transformative approach in the
field of materials science and engineering. By leveraging advanced computational techniques, including ML and data mining, this emerging
discipline provides an efficient pathway for accelerating materials discovery and development. Central to materials informatics are principles
such as data generation, feature extraction, and the management of
high-dimensionality and complexity in materials data. Utilized in areas
such as predictive modeling of materials properties, optimal experiment
design, and quality control in manufacturing, materials informatics
showcases its potential to revolutionalise the materials science field.
(MD) algorithms serve as a more practical approach in larger scales.
Finally, when at the macroscopic scale, continuum methods that essentially approximate the conservation relations between mass, energy
an momentum, offer a more expeditious but complimentary view [38].
Sophisticated calculations such as DFT are computationally intensive
and require massive amounts of computing power (manifesting in the
use of high performance computers, requiring large scale collaborations
at times). This is a very apt example of the expeditious effect the AI
can have in materials discovery. AI has the capacity to rationalize
large amounts of data and detect the underlying correlations between
them with relatively few data points (compared to the traditional approaches), thus make valuable predictions and help navigate uncharted
composition space landscapes. With the increasing interest in developing AI tools for DFT modeling, there has been a growing emphasis on
the importance, primarily, of data quality.
Only recently Merchant et al. a research team from Google’s Deepmind unveiled a novel deep learning tool, Graph Networks for Materials
Exploration (GNoME), designed to accelerate and expedite the identification of stable crystals for potential applications in various technologies [39]. The researchers use two frameworks, involving structural
modifications and compositional predictions, generating over 109 candidates, which are then filtered, ranked, and evaluated using GNoME,
ultimately assessed through DFT computations for stability and precision 2. GNoME’s predictions encompassed 2.2 million new crystals,
with 381,000 newly recognized (almost an order of magnitude larger
than previous work) as highly stable and suitable for experimental synthesis for a total of 421,000 stable crystals. This breakthrough, equivalent to 800 years’ worth of conventional knowledge, involved materials
with potential uses in superconductors, supercomputers, advanced batteries, and more. GNoME utilized AI to forecast material stability
with an unprecedented scale and accuracy. External researchers have
already experimentally realized 736 of the predicted structures, highlighting the tool’s efficacy. The discovered materials exhibit promise
for the development of sustainable technologies, including enhanced
electric vehicle batteries and more efficient computing.
Data quantity is also crucial for properly coupling AI with DFT.
Property mapping and high throughput screening by combining DFT
with AI is an approach that has gained significant attention in recent
years. This involves predicting the properties of materials based on
their electronic structures using ML models. The ML models are trained
on a large dataset of DFT calculations to learn the relationship between
the electronic structure and the material properties of interest. Once
trained, these models can be used to rapidly predict the properties of
new materials without the need for computationally expensive DFT
calculations. A study by Meredig et al. [40] used an ML-based approach
to screen over 1.6 million compounds and reveal 4500 previously
unknown ternary compounds, in which the AI algorithm required no
knowledge of crystal structure and operated at six orders of magnitude
lower computational expense. This technique would lay the foundation
for entirely new phase diagrams (or suggest additional contributions
to existing ones), accelerate the process of interatomic potentials construction (as will be discussed below), and provide a large list of
interesting new compositions that may now be mined for technological
applications. Overall, property mapping and screening in DFT with AI
have the potential to greatly accelerate the discovery and design of
new materials with desired properties, especially for applications where
high-throughput screening of large databases is required.
Apart from expeditious interpolation and high-throughput screening, AI can help first principles methods in a more fundamental level.
Density Functional Theory (DFT) has been a cornerstone for ab initio
computational studies, predicting a wide range of material properties
by calculating the electronic structure of a material. There are many
different types of DFT functionals, each with different levels of accuracy
and computational cost. The choice of functional depends on the specific system being studied and the desired level of accuracy. Functionals
are used to calculate the energy and other properties of a system based
2.2. Density functional theory
Density functional theory (DFT) is a widely used method in computational chemistry and physics for predicting the electronic structure
and properties of materials and molecules. It involves solving the
Schrödinger equation for a system of interacting electrons, using approximations to account for the many-body nature of the electron
interactions. In recent years, there have been efforts to use artificial
intelligence (AI) techniques to improve the efficiency and accuracy of
DFT calculations. Other approaches involve using AI to optimize the
choice of functional (approximation) used in DFT calculations, or to
learn the exchange–correlation functional directly from data. Overall,
the use of AI in DFT has the potential to significantly reduce the computational cost of predicting the electronic structure and properties of
materials and molecules, and could enable the study of larger and more
complex systems than is currently possible. However, these methods are
still an active area of research and development, and there is ongoing
work to improve their accuracy and applicability.
As DFT is essentially a first principles calculation approach, it is
very sensitive to the parameters used in order to achieve the optimal balance between accuracy and computation intensiveness. When
enough data are available the latter can be alleviated by interpolating
between known structures and their properties using AI techniques.
A well documented example is the case of diffusion modeling. It is a
fundamental phenomenon, present in all states of matter and consists
of the transport of atoms and molecules from high to low concentration
regions [37]. As is the case with materials discovery in most fields, for
many years, the discovery of novel materials was an arduous process of
trial and error where luck was a decisive factor [38]. At the electronic
level, DFT calculations play a central role, while Molecular Dynamics
4
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
Fig. 2. A summary of the GNoME-based discovery.
Source: The image is reproduced under the CC-BY
4.0 License from Ref. [39].
computational cost of DFT calculations, pseudopotentials are used to
replace the core electrons with an effective potential that accurately
models their behavior. This allows for more efficient calculations while
still capturing the important electronic behavior of the material. There
are different types of pseudopotentials, including ultrasoft [45] and
projector-augmented wave (PAW) pseudopotentials [46]. The use of AI
in DFT pseudopotential development stems from ML algorithms, which
can predict materials properties by training on large datasets. The use
of AI techniques, such as neural networks, has shown promising results
in predicting pseudopotential parameters and enhancing transferability. Woo et al. introduced a deep neural network-based method for
generating highly transferable local pseudopotentials (LPPs) for rapid
quantum simulations, overcoming limitations of traditional LPPs by
explicitly incorporating norm-conserving conditions in a loss function
and demonstrating superior performance and transferability across sand p-block elements [47].
To summarize, as AI continues to advance, it is expected to play
a more prominent role in materials modeling and the development of
more accurate and efficient DFT methods. Fast interpolation and highthroughput screening, as well as faster pseudopotential and functional
development can help bypass the constraints of ab initio techniques
and its trade-off between time consumption and materials modeling
accuracy, leading to the faster discovery of novel materials tailored to
specific sets of properties.
Fig. 3. The xc energy density is constructed using the fully-connected NN, which takes
the spatially local descriptors.
Source: The image is reproduced under the CC-BY 4.0 License from Ref. [43].
on its electron density. In other words, they determine the total energy
of a system as a functional of its electron density. This is done by
solving the Kohn–Sham equations [41], which are a set of equations
that describe the behavior of the electrons in the system. The first proof
of principle for AI use in DFT appeared ten years ago, when a simple
kernel ridge regression method was used to find an approximation of
the functional by training on results from accurate numerical calculations for a simple problem, such as the kinetic energy of non interacting
fermions in a one-dimensional box [42]. Nagai et al. demonstrated that
deep learning techniques, applied to a limited dataset of molecules,
could facilitate the creation of a Meta-Generalized Gradient Approximation (Meta-GGA) exchange–correlation functional. Remarkably, this
newly designed functional could be extended to accurately describe
hundreds of molecules containing first- and second-row elements, maintaining the precision of conventional functionals. This accomplishment was made possible by employing a versatile feed-forward neural
network to link density and energy (see Fig. 3) [43].
Kirkpatrick et al. [44] have introduced a novel approach to density
functional development by addressing the fractional electron problem.
By incorporating AI into their methodology, they have made significant
strides in creating more accurate density functionals. The authors used
a combination of ML algorithms and quantum Monte Carlo calculations
to develop new functionals capable of accurately describing fractional
electron systems. Their approach has demonstrated potential for the
design of advanced materials and development of more accurate DFT
methods. AI has had a significant impact on the development of DFT
pseudopotentials, leading to improvements in accuracy and efficiency.
The work of Kirkpatrick et al. [44] on solving the fractional electron
problem has pushed the frontiers of density functional development by
leveraging AI techniques.
Pseudopotentials, which simplify the calculation of the electronic
structure, play a crucial role in DFT simulations. They are used to model
the behavior of the core electrons in a material. Core electrons are the
electrons in an atom’s innermost shells, and they are responsible for
the atom’s chemical properties. However, they are much more tightly
bound to the nucleus than the valence electrons, which are the electrons
in the outermost shell. As a result, they require a lot of computational resources to accurately calculate their behavior. To reduce the
2.3. Molecular dynamics
Molecular dynamics (MD) is a powerful computational technique
for simulating the behavior of molecules and materials over time.
However, accurate modeling of the interactions between atoms and
molecules is crucial for obtaining reliable results from MD simulations.
Traditional force field development requires a combination of experimental data and empirical rules, which can be time-consuming and
computationally expensive. AI and ML techniques have shown great
potential in the development of accurate and efficient force fields for
MD simulations.
One approach to AI-assisted MD simulations is the use of computer
vision methods. These methods can be used to extract information from
large datasets of molecular structures, such as images or videos of
molecules. For example, the CNN can be used to identify and classify
different types of molecules based on their structural features. The CNN
has been successfully applied to the classification of small molecules,
as well as the prediction of protein-ligand binding affinities [48,49].
Other computer vision methods, such as autoencoders and generative
adversarial networks (GANs), can be used for the generation of new
molecular structures with desirable properties [50,51].
In addition to computer vision methods, AI and ML algorithms can
also be used for the development of more accurate and efficient force
fields. For example, neural networks can be used to learn the underlying relationships between atomic positions and forces, allowing for the
5
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
Fig. 4. Concept of the surrogate FEA approach. On the left, the numerical predictions before applying the ML approach. After the ML algorithm is applied, the numerical prediction
is a better representation of the actual system.
Source: The image is reproduced under the CC-BY 4.0 License from Ref. [59].
development of more accurate force fields [52]. These ML models have
been employed for optimizing potentials, leading to more accurate and
efficient simulations [53]. Support vector machines (SVMs) can be used
to predict the thermodynamic properties of materials based on their
atomic structures [54]. An example of the successful application of AI
in MD simulations is the work of Gastegger et al. [55], who developed
a neural network potential (NNP) for the simulation of molecular
dynamics. The NNP was trained on a dataset of over 20,000 organic
molecules and was shown to accurately predict molecular structures,
energies, and forces. The use of the NNP significantly reduced the
computational cost of the MD simulations, while maintaining a high
level of accuracy. Moreover, deep learning algorithms, such as graph
neural networks, can be used to model complex chemical reactions and
dynamics [56,57].
AI-assisted MD simulations have also been used for the discovery of
new materials and molecules. The use of AI and ML algorithms can accelerate the screening of large databases of molecules and materials for
their properties and potential applications. For example, the Materials
Project database uses AI and ML techniques to predict the properties
of thousands of materials, including their mechanical, thermal, and
electronic properties [58]. The database has been used for the discovery
of novel materials with desirable properties.
In conclusion, AI and ML techniques have shown great promise
in the field of MD simulations, particularly in the development of
accurate and efficient force fields and the discovery of new materials
and molecules. The combination of computer vision methods, neural
networks, SVMs, and deep learning algorithms can provide a powerful
tool for advancing the field of materials science. However, further
research is needed to address the challenges of large and diverse
datasets, appropriate training algorithms, and the interpretation of
results obtained from AI-assisted MD simulations.
Another application of AI in FEA is in the sphere of mesh refinement
optimization [62]. The incorporation of AI algorithms has empowered
FEA software to intuitively adapt the mesh resolution to the complexity
of the problem. This adaptation not only enhances the accuracy of the
analysis but also significantly reduces the computational cost, leading
to a more efficient process. This adaptive capability of AI in FEA has
proven to be particularly valuable in complex engineering problems
where the mesh resolution plays a critical role in determining the
accuracy of the analysis.
Furthermore, AI has also been used in failure prediction and design
optimization in FEA [63]. By using predictive models and optimization
algorithms, AI can help engineers anticipate potential failures in the
design phase and optimize the design to improve its performance and
durability. This can significantly reduce the cost and time required for
prototyping and testing, making the design process more efficient and
effective.
Overall, the state of the art in AI usage in FEA is rapidly progressing,
with substantial advancements being made in areas like material property prediction, mesh refinement optimization, failure prediction, and
design optimization. This amalgamation of AI and FEA holds promising
potential and is poised to bring about a significant transformation in
the field of engineering analysis, leading to more accurate, efficient,
and cost-effective FEA.
2.5. PINNs
The integration of physical laws into ML models, a concept known
as PINNs, was first introduced in 2018 by Raissi et al. [64]. This idea
stemmed from the realization that incorporating known physics into the
architecture and loss function of neural networks could solve complex
differential equations. The initial concept of combining physics-based
knowledge with ML models, however, originated in the 1980s [65].
PINNs are a type of Deep Neural Networks (DNNs) that are trained
to solve supervised learning tasks while adhering to any given law of
physics represented by general nonlinear Partial Differential Equations
(PDEs) [64]. The training of these networks is performed by constructing a loss function composed of a data-fidelity term, which measures
the model’s data fit, and a physics-informed term, which penalizes
predictions not adhering to the physical laws [64].
The primary function of PINNs is to approximate solutions to PDEs.
The difference between the predicted and actual outputs is minimized
by adjusting the weights and biases of the neural network through
backpropagation. The distinctive feature of PINNs is the integration
of known physical laws as auxiliary constraints in the network’s loss
function. This means that the training process not only minimizes
the discrepancy between model predictions and observed data but
also ensures that the model predictions conform to known physical
laws [64].
Traditionally, numerical methods have been used to solve complex
differential equations in materials science, which involves studying the
properties and behavior of materials. These equations can describe a
variety of phenomena, from the behavior of materials under stress to
the diffusion of atoms or the flow of heat. However, these traditional
methods can be computationally expensive and may not provide a
global solution. PINNs can revolutionize materials science by offering
2.4. Finite elements analysis
AI has been making significant strides in diverse sectors, including
Finite Element Analysis (FEA). As we advance technologically, the integration of ML into FEA has noticeably amplified in recent years [59].
This symbiosis of AI and FEA is largely propelled by the development
of sophisticated ML techniques and the availability of high-quality, diverse datasets. This integration has catalyzed significant advancements
in improving both the precision and efficiency of FEA by seamlessly
incorporating ML algorithms into the analytical process [60].
A pivotal domain where AI’s role in FEA is quite conspicuous is in
the prediction of material properties [61]. ML algorithms have been
ingeniously employed to predict a wide range of properties including
mechanical, thermal, and electrical characteristics of materials [6]. This
includes attributes such as Young’s modulus [3,4], melting temperature [5,6], and conductivity [7] among others. Surrogate FEA models
have been used to successfully provide a mid-fidelity estimate of the
material’s response at any time or location of interest, by employing
supervised ML algorithms (Fig. 4) [59]. These algorithms have proven
their worth by displaying a high degree of accuracy in their predictions.
High precision has led to more accurate and efficient FEA, largely diminishing the need for exhaustive experimental testing, thereby saving
valuable resources and time.
6
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
Lastly, CV is a very useful area of research to manufacturing as it can
enable smarter, faster and more efficient quality control [71]. Also, it is
an integral part of the Collaborative Robots (Cobots) design [72]. These
are specifically designed for direct interaction with a human within
a defined collaborative workspace; CV plays an important role in the
process of creating collaborative environments between humans and
robots.
efficient and accurate solutions to these problems. By solving the
governing PDEs and respecting the underlying physics, PINNs can
predict the behavior of materials under various conditions. This enables
rapid screening of material properties and conditions, accelerating
materials design and discovery. Moreover, PINNs can be instrumental
in discovering new materials by predicting their properties based on
governing physical laws. This capacity to design materials with desired
properties can significantly impact various industries, from electronics
to aerospace [66].
In addition to their ability to solve complex differential equations,
PINNs can also provide interpretability, a significant advantage over
conventional neural networks. Because PINNs incorporate physical
laws into their architecture, it is possible to trace how these laws
influence the network’s predictions. This transparency can yield insights into the relationship between input variables and the predicted
output, enhancing our understanding of the physical phenomena being
modeled [67]. Furthermore, by incorporating physical laws, PINNs
can better generalize and make reliable predictions even in regions
of the input space where data might be scarce or unavailable. This
interpretability and generalizability make PINNs particularly powerful
tools in scientific research and industry [67]. Overall, PINNs represent
a groundbreaking tool in the intersection of computational physics and
ML. By integrating physical laws into neural networks, they offer robust
and reliable predictions, significantly advancing materials design and
discovery.
3.1. Manufacturing process optimization techniques
AI is a rapidly growing field with numerous applications in various
industries. One such industry is manufacturing, where AI has the potential to revolutionize the entire process [73]. In simple terms, AI can be
defined as the ability of machines to perform tasks that would normally
require human intelligence, such as recognizing patterns or making
decisions. The role of AI in MP is to optimize the production cycle
by automating tasks, predicting outcomes, and improving efficiency.
This technology can be used in a wide range of applications, from
optimizing the supply chain to predicting equipment failures. With
AI, manufacturers can reduce waste, increase production speed, and
improve quality, all while lowering costs [74].
The importance of AI in MP is increasing day by day. This is because
of the numerous benefits it offers, such as increased productivity, lower
costs, and improved quality. By using AI, manufacturers can make
more informed decisions, reduce errors, and increase the accuracy of
predictions [74]. This has the potential to transform MP and make them
more efficient and effective. These processes have traditionally been
based on well-established techniques that have been refined over the
years to improve efficiency and reduce errors, while they are typically
designed to produce high-quality products at a reasonable cost [74].
Before the advent of AI, conventional optimization techniques were
used to improve these processes. Some of the most widely used techniques include statistical process control (SPC) and Design of Experiments (DoE) [75]. SPC involves monitoring and controlling the quality
of a product or process by using statistical methods. DoE, on the other
hand, is a structured method of identifying and testing the critical
factors that affect the quality of a product or process. While these conventional optimization techniques have been successful in improving
MP, they also have their limitations. SPC can be time-consuming and
requires a large amount of data to be effective. DoE, on the other hand,
can be expensive and may not be suitable for all types of MP.
Evolutionary algorithms, such as genetic algorithms and particle
swarm optimization, have been employed in MP optimization [76].
These algorithms are inspired by natural selection and mimic the
evolutionary process to optimize the MP. Genetic algorithms are used
for optimization problems with discrete variables, while particle swarm
optimization is applied to continuous variable optimization problems.
While these AI techniques have shown great potential in MP optimization, they also have limitations. The performance of these techniques
is highly dependent on the quality and quantity of data, and their
interpretability can be a challenge.
Another limitation of conventional optimization techniques is that
they do not always account for the complex interactions between
different factors that affect the quality of a product or process. By
using AI algorithms and predictive analytics, AI can identify patterns
and correlations in large datasets that conventional techniques may
miss. ML techniques have gained significant attention in MP optimization [77]. Supervised learning, unsupervised learning, and RL are the
three major types of ML techniques used in MP. Supervised learning is
utilized for classification and prediction, while unsupervised learning
is used for clustering and dimensionality reduction. Furthermore, RL
is applied in MP for decision-making and control. Deep learning is a
type of ML that uses neural networks with multiple layers to model
complex relationships in data. CNNs and RNNs are two popular types
of deep learning techniques used in MP. CNNs are used for image and
3. Process engineering, manufacturing and Industry 4.0
Manufacturing also stands to gain from next level advancements
in AI and the improved efficiency they bring in predictive maintenance, production robotization and optimization and quality control.
The recent Industry 4.0 paradigm shift has integrated AI as a leading
component of the industrial and manufacturing transformation. AIassisted industrial production systems can outperform humans, while
in some cases robots can complete tasks that a human could not or
would not perform, such as manipulation of hazardous materials.
ML is a subset of AI that enables learning from data via pattern
recognition. ML methods are typically categorized in unsupervised,
supervised, semi-supervised and RL approaches. Approaches such as
Support Vector Machines (SVM), Principal Component Analysis (PCA),
Gaussian Mixture Models (GMMs), K-Nearest Neighbors (KNN) and
Artificial Neural Networks (ANNs), as well as the fusion of the aforementioned methods [18] have been used for descriptive, predictive and
prescriptive maintenance and analytics in Industry 4.0 [19]. Methods
such as Decision Trees (DTs), Gradient Boosting (GB), clustering, Kmeans, Density-Based Spatial Clustering of Applications with Noise
(DBSCAN) and ANNs have been utilized for early failure detection [20]
and intrusion detection [21].
Deep Learning (DL) is a subfield of ML that consists of multi-layered
models that are typically trained on large datasets, which allows the
model to learn patterns and relationships within the data. These networks are typically composed of multiple layers of interconnected
nodes, which process and transform the input data as it flows through
the network. Convolutional (CNNs) and Recurrent Neural Networks
(RNNs), autoencoders and Deep Belief Networks are amongst the algorithms that have been used in Industry 4.0 (maintenance, analytics,
early fault detection and fault diagnosis) [23–26]
Language models enable a system or a machine to understand the
human language and extract insights from it. The advent of attentionbased methods [68] and transformers [69] has marked the era of
data-driven agents with systems such as GPT-3 [70]. NLP systems have
recently started finding their way into the manufacturing industry as
well. Attention-based models can be used for the interpretation of
natural language to structure machine language (e.g. SQL [28]) and
vice versa.
7
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
algorithms can adjust the printing parameters in real-time, ensuring the
optimal quality of the part. In 3D printing, AI is also being used to
develop new materials and alloys, enabling the creation of stronger,
more durable, and more versatile parts [83]. By analyzing the properties of different materials and their performance in different printing
processes, AI algorithms can develop new alloys and composites that
are optimized for specific applications, such as aerospace or medical
implants.
In nanomanufacturing, AI has revolutionized the field by enabling
the precise control and manipulation of materials at the nanoscale [84].
Nanomanufacturing involves the production of materials and devices
with dimensions on the scale of nanometers, where the properties of the
material can differ significantly from those at the macroscopic scale.
AI algorithms can control the properties of materials and manipulate
them to form complex structures and devices with unprecedented
precision. One of the applications of AI in nanomanufacturing is in the
development of new sensors and devices that can detect and respond
to stimuli at the nanoscale. These devices have the potential to revolutionize fields such as biomedicine and environmental monitoring,
enabling the detection and treatment of diseases and pollutants with
high precision. The integration of AI in additive manufacturing and
nanomanufacturing has enabled the development of new materials,
processes, and devices that were previously unattainable. AI algorithms
can optimize the design of parts and structures, monitor and control
the printing process, and develop new materials and alloys. These advancements have the potential to revolutionize fields such as aerospace,
biomedicine, and environmental monitoring.
Looking to the future, the implementation of AI in MP is expected
to increase significantly, bringing with it new opportunities and challenges. Among these challenges are the need for skilled workers, the
ethical implications of using AI in decision-making, and ensuring the
security of manufacturing systems. Overall, the use of AI in MP has
the potential to revolutionize the industry by improving efficiency,
reducing costs, and increasing productivity. By embracing these new
technologies and leveraging them effectively, manufacturers can gain
a competitive advantage in the global marketplace.
signal processing, while RNNs are utilized for time-series analysis and
sequence data.
AI can be used for decision support in MP, particularly in areas
such as production planning, scheduling, and inventory management.
AI techniques such as RL, decision trees, and Bayesian networks can
be used to analyze data, identify patterns, and make predictions [77].
These techniques can help manufacturers make informed decisions
about production, reduce waste, and optimize inventory levels.
One of the main benefits of AI in MP is the ability to automate tasks
that are repetitive, time-consuming, and prone to error. AI-powered
robots and machines can perform tasks such as assembly, material
handling, and quality control. This can help reduce costs, improve efficiency, and enhance safety in the MP. AI techniques such as linear programming, optimization algorithms, and predictive maintenance can
be used to identify areas of improvement and implement cost-saving
measures [77].
Quality control is an essential aspect of MP. AI can be used to
improve the accuracy and efficiency of quality control processes, such
as defect detection and classification, using techniques such as image recognition and machine vision [78]. AI-powered quality control
can help manufacturers improve product quality, reduce defects, and
increase customer satisfaction.
The impact of AI has been significant in various traditional manufacturing sectors, such as the automotive industry, aerospace industry,
and consumer goods industry. The automotive industry has been one
of the primary beneficiaries of AI in recent years. AI has helped
manufacturers to optimize their MP and improve the quality of their
products. For instance, the use of AI in predictive maintenance has
helped manufacturers to reduce downtime and improve production
efficiency [77]. AI-powered robots have also been used in various
automotive MP such as welding and painting, resulting in improved
accuracy and speed [79]. Furthermore, AI has been used in the development of self-driving cars, which have the potential to revolutionize
the transportation industry. Furthermore, the aerospace industry has
also benefited significantly from the use of AI. AI has helped aerospace
manufacturers to optimize their production processes and improve the
safety and performance of their products [80]. For example, AI has been
used in the design and development of aircraft, resulting in improved
aerodynamics and fuel efficiency. AI-powered robots have also been
used in various aerospace MP such as drilling and riveting, improving
accuracy and reducing lead times [81]. Additionally, AI has been
used in predictive maintenance of aircraft engines, helping to reduce
maintenance costs and improve safety. Finally, the consumer goods
industry has also been impacted by the use of AI in MP. AI has helped
manufacturers to optimize their production processes and improve
product quality. For instance, AI has been used in the development of
new products, resulting in improved design and functionality [82]. AI
has also been used in supply chain management, helping manufacturers
to improve their inventory management and reduce costs.
Additive manufacturing, also known as 3D printing, is an advanced
MP that has seen a surge in popularity over the past decade due to its
ability to produce complex geometries in a cost-effective manner. The
integration of AI in additive manufacturing has enabled the production
of parts with greater precision, accuracy, and consistency. AI has also
facilitated the development of new materials and processes that were
previously unfeasible or impractical to manufacture [83]. One of the
significant benefits of AI in additive manufacturing is the ability to
optimize the design of parts and structures. AI algorithms can analyze
the performance of different designs, and use this information to generate optimized designs that are more efficient, lighter, and cheaper
to manufacture. This optimization can lead to the development of new
products that were previously unattainable, such as complex lattices
and intricate geometries. Moreover, AI can be used to monitor and
control the printing process, ensuring that the part is printed correctly
and according to the desired specifications. By monitoring various
parameters, such as temperature, humidity, and material flow rate, AI
3.2. Active learning
Active learning is a ML method that involves actively selecting the
most informative examples to label, with the goal of improving the
performance of a ML model. In the context of materials design, active
learning can be used to improve the performance of ML algorithms that
are used to explore the space of possible materials and optimize their
properties [85,86].
One of the key challenges in materials design is the vastness of the
materials space, which includes an enormous number of possible combinations of chemical elements, structures, and properties. Traditional
methods for exploring this space, such as experiments and simulations,
are time-consuming and expensive, and they are limited by the human
ability to generate and test hypotheses. ML algorithms can help to
overcome these limitations by enabling the rapid and efficient exploration of the materials space, but they require large amounts of labeled
data in order to learn the relationships between materials properties
and their performance in specific applications. Active learning can be
used to address this challenge by enabling ML algorithms to select
the most informative examples to label, based on their potential to
improve the performance of the model [87]. This can be achieved by
using an active learning algorithm to select the examples that are most
uncertain or most informative, according to a selected criterion, and
then labeling these examples and adding them to the training dataset.
By iteratively selecting and labeling the most informative examples,
active learning can help to improve the performance of a ML model
for materials design, and it can also reduce the amount of labeled data
that is required to achieve a given level of performance.
8
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
Overall, active learning is a powerful and effective method for
improving the performance of ML algorithms for materials design.
By enabling ML algorithms to actively select the most informative
examples to label, active learning can help to reduce the amount of
labeled data that is required, and it can also improve the accuracy,
interpretability, and generalizability of the learned models [86].
The use of active learning in industrial processes has been growing in recent years, with advancements in ML techniques and the
growing availability of data. Active learning refers to the ability of a
ML algorithm to actively select the most informative data points for
training, rather than relying on a fixed dataset. This has the potential to
greatly improve the efficiency and effectiveness of industrial processes,
by reducing the amount of data required for training and enabling
real-time adaptation to changing conditions.
One key area where active learning has been applied is in predictive
maintenance for industrial equipment. By actively selecting data points
for training, ML algorithms have been able to accurately predict when
equipment is likely to fail, allowing for proactive maintenance and
reducing downtime. This has been particularly useful in industries
such as manufacturing, where equipment failures can have significant
financial and operational impacts.
Another area where active learning has been utilized is in process
control and optimization. By actively selecting data points for training,
ML algorithms have been able to adapt to changing process conditions
and optimize control parameters in real-time. This has been particularly
useful in industries such as chemical manufacturing and oil refining,
where process conditions can vary widely and require precise control.
Overall, the state of the art in active learning for industrial processes
is rapidly advancing, with significant progress being made in the ability
to adapt to changing conditions and optimize control parameters. This
has the potential to greatly impact industrial processes, leading to
improved efficiency and effectiveness.
has been employed for real-time optimization of injection molding
processes, leading to reduced waste and improved product quality [99].
AI has been applied in self-correcting processing across various industries, with a particular focus on materials manufacturing. AI can be
used to optimize production processes, identify defects, and maintain
consistent product quality [100]. For example, AI-based self-correcting
systems can optimize injection molding processes, reducing waste and
improving product quality [101]. AI-driven techniques can assist in defect detection [102], process parameter optimization [103] and online
predictive maintenance [14].
Despite these benefits, challenges associated with AI in selfcorrecting processing include the need for high-quality data, the complexity of AI models, and concerns about trust and transparency [104].
The future of AI in self-correcting processing holds great promise
for improving efficiency, reliability, and safety across industries. As
AI algorithms and computational power advance, it is expected that
AI-based self-correcting systems will become more capable and versatile, enabling them to tackle a wider range of complex tasks and
systems. In materials manufacturing, this could lead to significant
advancements in production efficiency, product quality, and sustainability. Despite the potential benefits of AI in self-correcting processing,
several challenges need to be addressed. High-quality and diverse
data is crucial for training accurate and robust AI models. The lack
of such data can lead to biased or ineffective AI systems [105]. AI
models, particularly deep learning algorithms, can be complex and
difficult to interpret. Ensuring that AI-driven self-correcting systems
are transparent and understandable is essential for building trust and
facilitating human oversight [106]. Integrating AI-based self-correcting
systems with existing infrastructure and processes can be challenging,
requiring modifications to existing workflows and the development of
new interfaces and protocols [107].
3.3. Digital twins
3.2.1. Self-correcting processing and smart manufacturing
Self-correcting processing is a technology that enables systems to
monitor, analyze, and adjust their operations in real-time to maintain
desired outcomes [88]. These systems identify deviations from expected
performance and automatically make corrections to ensure optimal
functioning [89]. Self-correcting processing is particularly relevant to
materials manufacturing, where maintaining efficient and reliable operations is crucial for productivity, safety, and sustainability [90]. This
technology is also applicable to various other industries, including
transportation, healthcare, and energy.
Before the advent of AI, self-correcting processing techniques relied on traditional control algorithms, such as proportional–integral–
derivative (PID) controllers, adaptive control, and model predictive
control [91]. These techniques utilized mathematical models to predict
system behavior and make adjustments based on predefined rules and
feedback loops [92]. However, these approaches often required extensive domain knowledge and manual tuning, and they were limited in
their ability to handle complex, nonlinear, and dynamic systems [93].
AI techniques have emerged as powerful tools for self-correcting
processing, offering greater adaptability, accuracy, and efficiency compared to traditional methods. Supervised and RL are three major categories of ML that have been utilized in self-correcting techniques in
manufacturing to optimize processes, enhance product quality, and
reduce waste [94]. Supervised learning, where an algorithm learns
to map inputs to outputs based on labeled training data, has been
extensively applied in manufacturing for self-correcting processes [95].
For instance, Cannizzaro et al. [96] employed supervised learning to
develop a predictive model for identifying and correcting defects in additive MP. RL [97], a method where an agent learns to make decisions
by interacting with its environment and receiving feedback in the form
of rewards or penalties, has shown promise in self-correcting techniques
and predictive maintenance in manufacturing [98]. Additionally, RL
Digital twin technology has been gaining popularity in various
industries, including materials science and engineering. A digital twin
is a virtual replica of a physical object or system that can be used to
simulate and predict its behavior in the real world. In materials design
and discovery, digital twins can be used to model and optimize the
properties of materials, predict their behavior under different conditions, and accelerate the discovery of new materials. This chapter will
explore the concept of digital twins in materials design and discovery,
their advantages, and practical applications.
The foundational concepts and frameworks needed to formulate and
continuously update both the form and function of the digital twin
of a selected material physical twin have been established [108]. The
Materials–Information Twin Tetrahedra (MITT) framework captures a
holistic perspective of materials science and engineering in the presence
of modern digital tools and infrastructures [109]. The digital twin
is primarily introduced in Chandhana et al. [110], along with its
advantages and practical applications in different sectors.
Digital twins can provide a range of advantages in materials design
and discovery. They can be used to model and optimize the properties
of materials, predict their behavior under different conditions, and
accelerate the discovery of new materials. Digital twin theory can store
and represent detailed information of objects, such as their materials,
textures, and 2D and 3D images [111]. The concept of digital twins
has already highlighted its potential in industries such as construction,
manufacturing, automotive, and agriculture [112–115].
Digital twins can be used in various practical applications in materials design and discovery. For example, they can be used to simulate
and optimize the properties of materials for specific applications, such
as aerospace or biomedical engineering. They can also be used to
predict the behavior of materials under different conditions, such as
temperature, pressure, or stress. Digital twins can also be used to
accelerate the discovery of new materials by simulating and testing
9
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
with advancements in AI and ML, are expected to contribute to accelerated materials discovery and the development of novel materials for a
wide range of applications. Various types of data are found in materials
science, including experimental, computational, and simulation data.
Experimental data is typically obtained through laboratory testing of
materials, while computational data is generated using mathematical
models and simulations [22]. Simulation data is particularly important,
as it can help predict the properties and behavior of materials under
different conditions, leading to the development of new materials with
desired properties [122].
Several open access materials science databases and repositories are
available and are shown in Table 1.
AI and ML have played a significant role in materials science, particularly in data processing, analysis, and simulation. ML algorithms have
been used to predict material properties, optimize synthesis processes,
and accelerate materials discovery [1,2]. Furthermore, AI techniques
have been employed to analyze large datasets and identify patterns or
correlations that can inform material design [8–10,10,11]. Challenges
in collecting, storing, and accessing materials science data include
data heterogeneity, interoperability, and accessibility. To address these
challenges, efforts have been made to standardize data formats, develop
ontologies for materials science, and promote open access to data.
The future of data in materials science is expected to be shaped by
the continued integration of AI and ML, as well as the development
of new experimental techniques and computational methods. Highthroughput experimentation, which involves the rapid synthesis and
characterization of a large number of materials, will likely generate
vast amounts of data, facilitating the discovery of new materials with
unique properties.
their properties in a virtual environment [116]. The modeling and
implementation method of a digital twin based on a physics simulation
for a material handling system is described in Jeong et al. [117].
Digital twin technology has the potential to revolutionize materials
design and discovery by providing a virtual environment for simulating
and optimizing the properties of materials. The foundational concepts
and frameworks needed to formulate and continuously update both the
form and function of the digital twin of a selected material physical
twin have been established. The advantages and practical applications
of digital twins in materials design and discovery are numerous and
varied. As digital twin technology continues to evolve, it is likely that
its use in materials science and engineering will become increasingly
widespread.
3.4. Characterization and computer vision
CV techniques have been increasingly used in materials design and
manufacturing in recent years due to their ability to process large
amounts of data and identify patterns and features within that data.
These capabilities make them particularly useful for tasks such as
image classification, object recognition, and anomaly detection. In the
field of materials design and manufacturing, CNNs have been applied
to a wide range of applications, including the prediction of material
properties, the optimization of MP, and the identification of defects in
materials [118–121].
One important application of CV in materials design is the prediction of material properties. Accurate predictions of material properties such as strength, ductility and fracture toughness are critical for
the design and development of new materials. CV have been shown
to be effective at predicting these properties based on input data
such as microstructures, compositions, and processing conditions. CNNs
have been used to predict the mechanical properties of metallic alloys [10] based on their microstructure, predict the corrosion resistance
of coatings based on their composition and processing conditions.
Another application of CV techniques in materials design and manufacturing is the optimization of MP. In manufacturing, it is important
to optimize processes in order to improve efficiency, reduce costs, and
improve product quality. CNNs have been used to optimize a variety of
MP, including casting, forging, and welding. For example, CNNs have
been used to predict the final shape of a casting based on the mold
geometry and casting conditions, and to optimize the welding process
in order to minimize defects and improve the strength of the welded
joint.
In addition to these applications, CV techniques have also been
used to identify defects in materials. Defects in materials can have a
significant impact on their performance and durability, and the early
identification of defects is important in order to prevent failures and
improve product quality. CNNs have been used to detect defects in
a variety of materials, including metals, ceramics, and polymers. For
example, CNNs have been used to identify cracks in metal structures,
and to detect defects in ceramic tiles and polymeric materials.
Overall, CV has proven to be a valuable tool in materials design
and manufacturing, with a wide range of applications in the prediction
of material properties, the optimization of MP, and the identification
of defects in materials. As the use of CNNs continues to grow in these
fields, it is likely that CV will play an increasingly important role in the
development of new materials and the optimization of MP.
4.2. Explainability
The proliferation of big data in materials science has opened new avenues for predictive modeling, optimization, and discovery. However,
the sheer volume and complexity of data pose significant challenges.
One of the pressing issues is making sense of intricate models that act
as "black boxes’’, providing predictions without explanations. This has
led to an increasing emphasis on the need for not just robust but also
explainable AI algorithms. Explainability in AI ensures that complex
ML models can be understood, interpreted, and trusted by domain
experts. This is particularly crucial in fields like materials informatics,
where a misinterpretation can lead to suboptimal designs or even safety
hazards [137].
Shapley Analysis based on the SHAP method, originating from
cooperative game theory, have found applications in ML to attribute the
contribution of each feature to the prediction for a particular instance.
In the context of materials science, Shapley Analysis can be employed
to understand how different material properties, influence the overall
performance or suitability of a material for a specific application [138].
A case study is shown in Fig. 5. Shapley values work by approximating
the output of a model with a local linear explanation model. The
coefficients of this explanation model quantify the local effect of each
feature on the output. The coefficients can be aggregated to get global
feature contributions (Fig. 5(a)). Then model is built to relate the
physio-chemical descriptors of a capping layer of lead halide perovskite
solar cells. Afterwards, a ML model is trained to predict the onset time
of degradation of the solar cells under ambient conditions. Shapley
analysis allows the identification of the dominant descriptors in the
model, shown in the figures as a distribution of local Shapley values.
Top polar surface area and H-bond donor have the most significant
impact the output’s prediction and are demonstrated to have dominant
importance by additional experimentation (Fig. 5(b)).
Inverse design stands as another pillar in the realm of explainable
AI in materials science. Traditional material discovery often involves
forward simulations where properties are calculated based on given
4. Big data and explainability
4.1. Data in materials science
Data in materials science plays a critical role in advancing our understanding of material properties and driving innovation in the field.
The availability and accessibility of various types of data, combined
10
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
Table 1
Materials science databases.
Database
Description
Materials Project [123]
A database focused on computational materials science data, including
crystal structures, electronic properties, and thermodynamic data.
Crystallography Open Database (COD) [124]
An open-access repository of crystal structures for various materials.
Inorganic Crystal Structure Database (ICSD) [125]
A database for completely identified inorganic crystal structures.
MatWeb [126]
A comprehensive database of material properties, including mechanical,
thermal, and electrical properties.
Cambridge Structural Database [127]
The Cambridge Structural Database (CSD) is a highly curated and
comprehensive resource. Established in 1965, the CSD is the world’s
repository for small-molecule organic and metal–organic crystal
structures.
International Centre for Diffraction Data [128]
The ICDD is an authority on X-ray powder diffraction and related
diffraction methods. It provides a suite of databases designed for use in
materials identification and characterization.
GDB Databases [129]
The GDB databases are generated by systematic enumeration of all
possible molecules following simple construction rules. These databases
are used for structure prediction, drug discovery, or materials design.
ZINC Database [130]
ZINC is a free database of commercially-available compounds for
virtual screening. It is provided by the Shoichet Laboratory in the
Department of Pharmaceutical Chemistry at the University of
California, San Francisco (UCSF).
Automatic FLOW for Materials Discovery LIBrary [131]
AFLOWLIB is an ab-initio, high-throughput computational materials
repository from Duke University. It is part of the Materials Project, a
multi-institution effort to compute the properties of all known
inorganic materials.
Open Quantum Materials Database [132]
The OQMD is a high-throughput database aimed at the discovery of
new materials. It contains density functional theory (DFT) calculated
properties for millions of materials.
Harvard Clean Energy Project [133]
The Harvard Clean Energy Project used computational chemistry and
the power of distributed computing to find new materials for solar
cells. The project aimed to identify promising organic photovoltaics.
TE Design Lab [134]
The TE Design Lab is a database for thermoelectric materials. It
provides resources for the design and optimization of new
thermoelectric materials.
NREL Materials Database [135]
The National Renewable Energy Laboratory (NREL) Materials Database
is a resource for data on materials used in renewable energy
technologies.
Materials Cloud [136]
Materials Cloud (MC) is a web-based portal for materials science. It
offers educational, research, and archiving tools, and promotes the
sharing and reuse of data generated in the field of materials science.
delve into uncharted territory with confidence and consistency, there
are several challenges that must be addressed.
Firstly, there is the issue of common frameworks to work within
and more specifically, standardization. With regard to the latter, Materials Data Analytics (MODA) and Characterization Data (CHADA) have
emerged as important frameworks. MODA focuses on documenting
materials modeling, encompassing use cases, models, solvers, and other
processing steps. However, it has limitations in its scope, notably a
rigid focus on a static set of physical equations. Work that extends
beyond the predefined theoretical or methodological categories either
has to be forcibly aligned with the closest existing category, or it
cannot be included at all [141]. CHADA, on the other hand, aims
for standardization in material characterization but often results in
representations that lack clear insights into the experiments conducted.
The outcome is a representation that, while understandable to humans,
is frequently ambiguous and often provides less insight than a detailed
research paper explaining the experimental procedures [141]. The lack
of consistent standards extends beyond MODA and CHADA, encompassing data formats, evaluation metrics, and methodologies. This lack of
standardization makes it challenging to compare or integrate different
studies, calling for more adaptable and inclusive standards.
Another issue that arises in the realm of materials science is the
trade-off between model accuracy and interpretability, especially when
employing complex ML models like deep neural networks [137]. These
models are capable of delivering highly precise predictions; however,
material structures. In contrast, inverse design methods work backwards: they start with desired properties and use ML algorithms to
suggest possible material structures or compositions that might exhibit
those properties [139]. Inverse design not only accelerates the material
discovery process but also adds a layer of explainability. By mapping
desired properties back to material compositions or structures, it provides a clear rationale for why a particular material might be suitable
for a specific application [139].
Other methods such as LIME (Local Interpretable Model-agnostic
Explanations) and counterfactual explanations are also gaining traction
in the field of materials informatics. These methods offer localized
explanations for model predictions and can be particularly useful in scenarios where only instance-specific interpretability is required [140].
The need for robust and explainable AI in materials science is not
just a technological requirement but a pragmatic necessity. As the
field continues to be data-rich, the models grow more complex. The
integration of explainability methods like Shapley Analysis and inverse
design serves to make these complex models interpretable, thereby
making the data-driven discovery process in materials science more
transparent, reliable, and actionable.
5. Challenges
While AI techniques have achieved significant strides in material
design, discovery, and process optimization, several areas within materials science remain relatively underexplored. In order for the field to
11
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
efforts will be key in unlocking the full potential of AI in this rapidly
evolving field.
6. Conclusions
In this comprehensive review, the transformative impact of AI
across various domains of materials science has been highlighted. From
Materials Informatics, Density Functional Theory (DFT), Molecular
Dynamics (MD), to Finite Elements Analysis (FEA), the breadth and
depth of AI’s applicability in materials science have been demonstrated.
Beyond the scope of material design, the role of AI in revolutionizing
manufacturing through process optimization techniques is discussed.
Active learning and self-correcting systems are shown to represent
a new era in smart manufacturing. Additionally, Digital Twins are
presented as a groundbreaking tool for simulation and optimization
in industrial settings. The necessity for both explainable and robust
AI systems is emphasized, thereby fulfilling a critical requirement for
the broader adoption of AI in materials science. This focus aligns
with the discussion on Big Data, where the potential of data-driven
methodologies is illuminated.
However, this review also identifies several challenges and gaps that
exist in the current landscape of AI applications in materials science.
These range from the urgent need for standardization, highlighted by
initiatives like MODA and CHADA, to the trade-offs between model
accuracy and interpretability. Questions surrounding the computational
costs of AI algorithms are raised, pointing to concerns about scalability
and environmental impact. The democratization of AI and the importance of open data are also discussed as critical factors for the future
development of the field.
In conclusion, as AI technologies continue to evolve and find deeper
integration into materials science, addressing these multifaceted challenges through open dialogue, rigorous scientific scrutiny, and collaborative research is imperative. Such a comprehensive approach is
expected to mitigate existing challenges and unlock new avenues for
discovery and innovation in materials science.
Fig. 5. Interpretability with Shapley values. (a) Shapley values are a generalization of
various black-box explainability methods. (b) Case study of Shapley analysis in material
science.
Source: The image is reproduced under the CC-BY 4.0 License from Ref. [137].
their intricate architecture often acts as a "black box’’, making it
difficult to understand the underlying mechanisms that led to a particular decision. In materials science, this lack of interpretability can
be a significant drawback. Understanding the ‘why’ and ‘how’ behind
predictions is essential for gaining insights into material properties and
behaviors, which in turn is crucial for the iterative process of material
design and discovery.
Computational efficiency is another critical challenge in applying
AI to materials science. Advanced algorithms, especially deep learning models, often require significant computational resources for both
training and inference. This resource intensity raises serious concerns
about the scalability of these techniques. Specifically, it calls into question their applicability in real-time scenarios, where quick decisionmaking is essential, or in settings where computational resources are
constrained. The computational burden not only limits the immediate
utility of these algorithms but also poses long-term questions about the
energy efficiency and environmental impact of large-scale deployments.
Another aspect worth considering is the democratization of AI in
the field of materials science. The advent of open-source platforms
and community-driven resources is lowering the entry barrier, making
it possible for a wider range of researchers, including those from
smaller institutions, to engage in cutting-edge research [123]. This
democratization has the potential not only to accelerate innovation but
also to diversify the sources of breakthroughs in the field. Rather than
waiting for advancements to trickle down from a few well-resourced
organizations with extensive data centers, democratization allows for
a more collective, widespread contribution to the discovery and design
of new materials. The same goes for open access to high-quality, standardized datasets not only fosters collaboration but also enables more
robust ML models [123]. As AI continues to advance, the importance of
open data becomes even more critical for validating models, ensuring
quality control, and accelerating innovation in the field.
Addressing these multifaceted challenges and opportunities is crucial for the sustainable and responsible advancement of AI in materials
science. Open dialogue, rigorous scientific scrutiny, and collaborative
CRediT authorship contribution statement
Ioannis Papadimitriou: Conceptualization, Data curation, Investigation, Methodology, Writing – original draft, Writing – review &
editing. Ilias Gialampoukidis: Writing – review & editing. Stefanos
Vrochidis: Resources. Ioannis Kompatsiaris: Supervision.
Declaration of competing interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to
influence the work reported in this paper.
Data availability
No data was used for the research described in the article.
Acknowledgment
This work was supported by the Horizon Europe Framework Programme and EC-funded project DiMAT, EU under grant agreement No
101091496.
Appendix
See Table 2.
12
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
Table 2
Table of acronyms in order of appearance.
Acronym
Full form
AI
CV
CNN
NLP
ML
IoT
MP
DL
GAN
PINN
PDE
SVM
KNN
EA
NLP
DFT
MD
PAW
GAN
NNP
FEA
MODA
CHADA
RL
PCA
GMM
ANN
DT
GB
DBSCAN
RNN
GPT-3
SQL
Cobots
SPC
DoE
PID
RL
MITT
XAI
COD
ICSD
CSD
ICDD
GDB
UCSF
AFLOWLIB
OQMD
HCEP
TE
NREL
MC
XAI
Artificial Intelligence
Computer Vision
Convolutional Neural Network
Natural Language Processing
Machine Learning
Internet of Things
Manufacturing Processes
Deep Learning
Generative Adversarial Network
Physics-Informed Neural Network
Partial Differential Equations
Support Vector Machines
K-Nearest-Neighbors
Evolutionary Algorithms
Natural Language Processing
Density Functional Theory
Molecular Dynamics
Projector-augmented Wave
Generative Adversarial Network
Neural Network Potential
Finite Element Analysis
Materials Data Analytics
Characterization Data
Reinforcement Learning
Principal Component Analysis
Gaussian Mixture Model
Artificial Neural Network
Decision Trees
Gradient Boosting
Density-Based Spatial Clustering of Applications with Noise
Recurrent Neural Network
Generative Pretrained Transformer 3
Structured Query Language
Collaborative Robots
Statistical Process Control
Design of Experiments
Proportional-Integral-Derivative
Reinforcement Learning
Materials–Information Twin Tetrahedra
Explainable AI
Crystallography Open Database
Inorganic Crystal Structure Database
Cambridge Structural Database
International Centre for Diffraction Data
Generated Database
University of California, San Francisco
Automatic FLOW for Materials Discovery LIBrary
Open Quantum Materials Database
Harvard Clean Energy Project
ThermoElectric
National Renewable Energy Laboratory
Materials Cloud
Explainable AI
[7] L. Chen, H. Tran, R. Batra, C. Kim, R. Ramprasad, Machine learning models
for the lattice thermal conductivity prediction of inorganic materials, Comput.
Mater. Sci. 170 (2019) 109155.
[8] C. Lv, X. Zhou, L. Zhong, C. Yan, M. Srinivasan, Z.W. Seh, C. Liu, H. Pan,
S. Li, Y. Wen, et al., Machine learning: an advanced platform for materials
development and state prediction in lithium-ion batteries, Adv. Mater. 34 (25)
(2022) 2101474.
[9] G.J. Adekoya, O.C. Adekoya, U.K. Ugo, E.R. Sadiku, Y. Hamam, S.S. Ray, A
mini-review of artificial intelligence techniques for predicting the performance
of supercapacitors, Mater. Today: Proc. (2022).
[10] Z.-W. Xu, X.-M. Liu, K. Zhang, Mechanical properties prediction for hot
rolled alloy steel using convolutional neural network, IEEE Access 7 (2019)
47068–47078.
[11] K. McCullough, T. Williams, K. Mingle, P. Jamshidi, J. Lauterbach, Highthroughput experimentation meets artificial intelligence: a new pathway to
catalyst discovery, Phys. Chem. Chem. Phys. 22 (20) (2020) 11174–11196.
[12] S. Vaidya, P. Ambad, S. Bhosle, Industry 4.0–a glimpse, Proc. Manuf. 20 (2018)
233–238.
[13] H. Yang, S. Kumara, S.T. Bukkapatnam, F. Tsung, The internet of things for
smart manufacturing: A review, IISE Trans. 51 (11) (2019) 1190–1216.
[14] C. Gianoglio, E. Ragusa, P. Gastaldo, F. Gallesi, F. Guastavino, Online predictive
maintenance monitoring adopting convolutional neural networks, Energies 14
(15) (2021) 4711.
[15] J. Cai, X. Chu, K. Xu, H. Li, J. Wei, Machine learning-driven new material
discovery, Nanoscale Adv. 2 (8) (2020) 3115–3130.
[16] Y. Hong, B. Hou, H. Jiang, J. Zhang, Machine learning and artificial neural
network accelerated computational discoveries in materials science, Wiley
Interdiscip. Rev. Comput. Mol. Sci. (2019).
[17] Y. Liu, T. Zhao, W. Ju, S. Shi, Materials discovery and design using machine
learning, J. Materiom. 3 (3) (2017) 159–177.
[18] A. Diez-Olivan, J. Del Ser, D. Galar, B. Sierra, Data fusion and machine learning
for industrial prognosis: Trends and perspectives towards industry 4.0, Inf.
Fusion 50 (2019) 92–111.
[19] I. Ahmed, G. Jeon, F. Piccialli, From artificial intelligence to explainable
artificial intelligence in industry 4.0: A survey on what, how, and where, IEEE
Trans. Ind. Inform. 18 (8) (2022) 5031–5042.
[20] J. Carvajal Soto, F. Tavakolizadeh, D. Gyulai, An online machine learning
framework for early detection of product failures in an Industry 4.0 context,
Int. J. Comput. Integr. Manuf. 32 (4–5) (2019) 452–465.
[21] S.-T. Park, G. Li, J.-C. Hong, A study on smart factory-based ambient intelligence context-aware intrusion detection system using machine learning, J.
Ambient Intell. Humaniz. Comput. 11 (4) (2020) 1405–1412.
[22] P. Patrone, A. Kearsley, A. Dienstfrey, The role of data analysis in uncertainty quantification: Case studies for materials modeling, in: 2018 AIAA
Non-Deterministic Approaches Conference, 2018, p. 0927.
[23] L. Li, K. Ota, M. Dong, Deep learning for smart industry: Efficient manufacture
inspection system with fog computing, IEEE Trans. Ind. Inform. 14 (10) (2018)
4665–4673.
[24] B. Luo, H. Wang, H. Liu, B. Li, F. Peng, Early fault detection of machine tools
based on deep learning and dynamic identification, IEEE Trans. Ind. Electron.
66 (1) (2018) 509–518.
[25] J. Pan, Y. Zi, J. Chen, Z. Zhou, B. Wang, LiftingNet: A novel deep learning
network with layerwise feature learning from noisy mechanical data for fault
classification, IEEE Trans. Ind. Electron. 65 (6) (2017) 4973–4982.
[26] Y. Tao, X. Wang, R.-V. Sánchez, S. Yang, Y. Bai, Spur gear fault diagnosis using
a multilayer gated recurrent unit approach with vibration signal, IEEE Access
7 (2019) 56880–56889.
[27] P.C. Jennings, S. Lysgaard, J.S. Hummelshøj, T. Vegge, T. Bligaard, Genetic
algorithms for computational materials discovery accelerated by machine
learning, NPJ Comput. Mater. 5 (2019) 1–6.
[28] S. Ferreira, G. Leitão, I. Silva, A. Martins, P. Ferrari, Evaluating human-machine
translation with attention mechanisms for industry 4.0 environment SQL-based
systems, in: 2020 IEEE International Workshop on Metrology for Industry 4.0
& IoT, IEEE, 2020, pp. 229–234.
[29] F. Kuniyoshi, J. Ozawa, M. Miwa, Analyzing research trends in inorganic
materials literature using NLP, in: ECML/PKDD, 2021.
[30] OpenAI, https://openai.com/, Accessed: 2010-09-30.
[31] M.P. Polak, D. Morgan, Extracting accurate materials data from research papers
with conversational language models and prompt engineering – example of
ChatGPT, 2023, arXiv:2303.05352.
[32] R. Ramprasad, R. Batra, G. Pilania, A. Mannodi-Kanakkithodi, C. Kim, Machine learning in materials informatics: recent applications and prospects, NPJ
Comput. Mater. 3 (1) (2017) 54.
[33] K. Rajan, Materials informatics: The materials ‘‘gene’’ and big data, Annu. Rev.
Mater. Res. 45 (2015) 153–169.
[34] L. Himanen, A. Geurts, A.S. Foster, P. Rinke, Data-driven materials science:
status, challenges, and perspectives, Adv. Sci. 6 (21) (2019) 1900808.
[35] S. Samudrala, K. Rajan, B. Ganapathysubramanian, Data dimensionality
reduction in materials science, Inform. Mater. Sci. Eng. (2013) 97–119.
References
[1] J. Wei, X. Chu, X.-Y. Sun, K. Xu, H.-X. Deng, J. Chen, Z. Wei, M. Lei, Machine
learning in materials science, InfoMat 1 (3) (2019) 338–358.
[2] T. Mueller, A.G. Kusne, R. Ramprasad, Machine learning in materials science:
Recent progress and emerging applications, Rev. Comput. Chem. 29 (2016)
186–273.
[3] K. Yang, X. Xu, B. Yang, B. Cook, H. Ramos, N. Krishnan, M.M. Smedskjaer, C.
Hoover, M. Bauchy, Predicting the Young’s modulus of silicate glasses using
high-throughput molecular dynamics simulations and machine learning, Sci.
Rep. 9 (1) (2019) 1–11.
[4] Y.-J. Hu, G. Zhao, M. Zhang, B. Bin, T. Del Rose, Q. Zhao, Q. Zu, Y. Chen, X.
Sun, M. de Jong, et al., Predicting densities and elastic moduli of SiO2-based
glasses by machine learning, NPJ Comput. Mater. 6 (1) (2020) 1–13.
[5] N. Qu, Y. Liu, M. Liao, Z. Lai, F. Zhou, P. Cui, T. Han, D. Yang, J. Zhu,
Ultra-high temperature ceramics melting temperature prediction via machine
learning, Ceram. Int. 45 (15) (2019) 18551–18555.
[6] V. Venkatraman, S. Evjen, H.K. Knuutila, A. Fiksdahl, B.K. Alsberg, Predicting
ionic liquid melting points using machine learning, J. Mol. Liq. 264 (2018)
318–326.
13
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
[64] M. Raissi, P. Perdikaris, G.E. Karniadakis, Physics-informed neural networks: A
deep learning framework for solving forward and inverse problems involving
nonlinear partial differential equations, J. Comput. Phys. 378 (2019) 686–707.
[65] J.R. Quinlan, Induction of decision trees, Mach. Learn. 1 (1986) 81–106.
[66] K.T. Butler, D.W. Davies, H. Cartwright, O. Isayev, A. Walsh, Machine learning
for molecular and materials science, Nature 559 (7715) (2018) 547–555.
[67] C. Rudin, Stop explaining black box machine learning models for high stakes
decisions and use interpretable models instead, Nat. Mach. Intell. 1 (5) (2019)
206–215.
[68] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A.N. Gomez, Ł.
Kaiser, I. Polosukhin, Attention is all you need, Adv. Neural Inf. Process. Syst.
30 (2017).
[69] T. Wolf, L. Debut, V. Sanh, J. Chaumond, C. Delangue, A. Moi, P. Cistac,
T. Rault, R. Louf, M. Funtowicz, et al., Transformers: State-of-the-art natural
language processing, in: Proceedings of the 2020 Conference on Empirical
Methods in Natural Language Processing: System Demonstrations, 2020, pp.
38–45.
[70] T. Brown, B. Mann, N. Ryder, M. Subbiah, J.D. Kaplan, P. Dhariwal, A.
Neelakantan, P. Shyam, G. Sastry, A. Askell, et al., Language models are
few-shot learners, Adv. Neural Inf. Process. Syst. 33 (2020) 1877–1901.
[71] M. Nagy, G. Lăzăroiu, Computer vision algorithms, remote sensing data fusion
techniques, and mapping and navigation tools in the Industry 4.0-based Slovak
automotive sector, Mathematics 10 (19) (2022) 3543.
[72] A. Pazienza, N. Macchiarulo, F. Vitulano, A. Fiorentini, M. Cammisa, L. Rigutini,
E. Di Iorio, A. Globo, A. Trevisi, A novel integrated industrial approach
with cobots in the age of industry 4.0 through conversational interaction and
computer vision., in: CLiC-It, 2019.
[73] S. Fahle, C. Prinz, B. Kuhlenkötter, Systematic review on machine learning
(ML) methods for manufacturing processes–Identifying artificial intelligence (AI)
methods for field application, Proc. CIRP 93 (2020) 413–418.
[74] E.S. Meieran, Intelligent manufacturing systems, in: Proceedings of 15th
IEEE/CHMT International Electronic Manufacturing Technology Symposium,
1993, pp. 323–327.
[75] S. Haridy, S.A. Gouda, Z. Wu, An integrated framework of statistical process
control and design of experiments for optimizing wire electrochemical turning
process, Int. J. Adv. Manuf. Technol. 53 (2011) 191–207.
[76] C. Dimopoulos, A.M.S. Zalzala, Recent developments in evolutionary computation for manufacturing optimization: problems, solutions, and comparisons,
IEEE Trans. Evol. Comput. 4 (2000) 93–113.
[77] S. Fahle, C. Prinz, B. Kuhlenkötter, Systematic review on machine learning
(ML) methods for manufacturing processes – identifying artificial intelligence
(AI) methods for field application, Proc. CIRP 93 (2020) 413–418.
[78] V. Piuri, F. Scotti, M. Roveri, Computational intelligence in industrial quality
control, in: IEEE International Workshop on Intelligent Signal Processing, 2005,
2005, pp. 4–9.
[79] X. Wang, X. Zhou, Z. Xia, X. Gu, A survey of welding robot intelligent path
optimization, J. Manuf. Process. (2020).
[80] M.A.A. Oroumieh, S.M.-B. Malaek, M. Ashrafizaadeh, S. Taheri, Aircraft design
cycle time reduction using artificial intelligence, Aerosp. Sci. Technol. 26 (2013)
244–258.
[81] H. Kihlman, I. Eriksson, M. Ennis, Robotic orbital drilling of structures for
aerospace applications, tech. rep., SAE Technical Paper, 2002.
[82] N. Nozaki, E. Konno, M. Sato, M. Sakairi, T. Shibuya, Y. Kanazawa, S.
Georgescu, Application of artificial intelligence technology in product design,
Fujitsu Sci. Tech. J. 53 (4) (2017) 43–51.
[83] Y. Wang, P. Zheng, T. Peng, H. Yang, J. Zou, Smart additive manufacturing:
Current artificial intelligence-enabled methods and future perspectives, Sci.
China Technol. Sci. (2020) 1–12.
[84] G.M. Sacha, P. Varona, Artificial intelligence in nanotechnology, Nanotechnology 24 (2013).
[85] D. Cohn, L. Atlas, R. Ladner, Improving generalization with active learning,
Mach. Learn. 15 (1994) 201–221.
[86] B. Settles, Active learning literature survey, tech. rep., University of
Wisconsin-Madison Department of Computer Sciences, 2009.
[87] S. Sun, D.R. Hardoon, Active learning with extremely sparse labeled examples,
Neurocomputing 73 (16–18) (2010) 2980–2988.
[88] E.F. Camacho, C.B. Alba, Model Predictive Control, Springer science & business
media, 2013.
[89] K.J. Åström, R.M. Murray, Feedback Systems: An Introduction for Scientists and
Engineers, Princeton University Press, 2021.
[90] J. Lee, H.D. Ardakani, S. Yang, B. Bagheri, Industrial big data analytics and
cyber-physical systems for future maintenance & service innovation, Proc. Cirp
38 (2015) 3–7.
[91] K.J. Åström, T. Hägglund, K.J. Astrom, Advanced PID Control, Vol. 461, ISAThe Instrumentation, Systems, and Automation Society Research Triangle Park,
2006.
[92] Y. Huang, J. Na, X. Wu, G.-B. Gao, Y. Guo, Robust adaptive control for vehicle
active suspension systems with uncertain dynamics, Trans. Inst. Meas. Control
40 (4) (2018) 1237–1249.
[93] K.S. Narendra, A.M. Annaswamy, Stable Adaptive Systems, Courier Corporation,
2012.
[36] R. Aggarwal, M. Demkowicz, Y. Marzouk, Information-driven experimental
design in materials science, in: Information Science for Materials Discovery and
Design, Springer, 2015, pp. 13–44.
[37] H. Mehrer, Diffusion in Solids: Fundamentals, Methods, Materials, DiffusionControlled Processes, Vol. 155, Springer Science & Business Media,
2007.
[38] Y. Elbaz, D. Furman, M. Caspary Toroker, Modeling diffusion in functional
materials: from density functional theory to artificial intelligence, Adv. Funct.
Mater. 30 (18) (2020) 1900778.
[39] A. Merchant, S. Batzner, S.S. Schoenholz, M. Aykol, G. Cheon, E.D. Cubuk,
Scaling deep learning for materials discovery, Nature (2023) 1–6.
[40] B. Meredig, A. Agrawal, S. Kirklin, J.E. Saal, J.W. Doak, A. Thompson, K.
Zhang, A. Choudhary, C. Wolverton, Combinatorial screening for new materials
in unconstrained composition space with machine learning, Phys. Rev. B 89
(2014) 094104, URL https://link.aps.org/doi/10.1103/PhysRevB.89.094104.
[41] W. Kohn, L.J. Sham, Self-consistent equations including exchange and
correlation effects, Phys. Rev. 140 (4A) (1965) A1133.
[42] J.C. Snyder, M. Rupp, K. Hansen, K.-R. Müller, K. Burke, Finding density
functionals with machine learning, Phys. Rev. Lett. 108 (25) (2012) 253002.
[43] R. Nagai, R. Akashi, O. Sugino, Completing density functional theory by
machine learning hidden messages from molecules, NPJ Comput. Mater. 6 (1)
(2020) 43.
[44] J. Kirkpatrick, B. McMorrow, D.H. Turban, A.L. Gaunt, J.S. Spencer, A.G.
Matthews, A. Obika, L. Thiry, M. Fortunato, D. Pfau, et al., Pushing the frontiers
of density functionals by solving the fractional electron problem, Science 374
(6573) (2021) 1385–1389.
[45] D. Vanderbilt, Soft self-consistent pseudopotentials in a generalized eigenvalue
formalism, Phys. Rev. B 41 (11) (1990) 7892.
[46] G. Kresse, D. Joubert, From ultrasoft pseudopotentials to the projector
augmented-wave method, Phys. Rev. B 59 (3) (1999) 1758.
[47] J. Woo, H. Kim, W.Y. Kim, Neural network-based pseudopotential: development
of a transferable local pseudopotential, Phys. Chem. Chem. Phys. 24 (34) (2022)
20094–20103.
[48] I. Wallach, M. Dzamba, A. Heifets, AtomNet: a deep convolutional neural
network for bioactivity prediction in structure-based drug discovery, 2015,
arXiv preprint arXiv:1510.02855.
[49] D.C. Elton, Z. Boukouvalas, M.D. Fuge, P.W. Chung, Deep learning for molecular
design—a review of the state of the art, Mol. Syst. Des. Eng. 4 (4) (2019)
828–849.
[50] R. Gómez-Bombarelli, J.N. Wei, D. Duvenaud, J.M. Hernández-Lobato, B.
Sánchez-Lengeling, D. Sheberla, J. Aguilera-Iparraguirre, T.D. Hirzel, R.P.
Adams, A. Aspuru-Guzik, Automatic chemical design using a data-driven
continuous representation of molecules, ACS Central Sci. 4 (2) (2018) 268–276.
[51] Z. Yao, B. Sánchez-Lengeling, N.S. Bobbitt, B.J. Bucior, S.G.H. Kumar, S.P.
Collins, T. Burns, T.K. Woo, O.K. Farha, R.Q. Snurr, et al., Inverse design of
nanoporous crystalline reticular materials with deep generative models, Nat.
Mach. Intell. 3 (1) (2021) 76–86.
[52] J. Behler, M. Parrinello, Generalized neural-network representation of highdimensional potential-energy surfaces, Phys. Rev. Lett. 98 (14) (2007)
146401.
[53] L. Zhang, J. Han, H. Wang, R. Car, E. Weinan, Deep potential molecular
dynamics: a scalable model with the accuracy of quantum mechanics, Phys.
Rev. Lett. 120 (14) (2018) 143001.
[54] A.P. Bartók, S. De, C. Poelking, N. Bernstein, J.R. Kermode, G. Csányi, M.
Ceriotti, Machine learning unifies the modeling of materials and molecules,
Sci. Adv. 3 (12) (2017) e1701816.
[55] M. Gastegger, P. Marquetand, Molecular Dynamics with Neural Network
Potentials, Springer, 2020.
[56] D.K. Duvenaud, D. Maclaurin, J. Iparraguirre, R. Bombarell, T. Hirzel, A.
Aspuru-Guzik, R.P. Adams, Convolutional networks on graphs for learning
molecular fingerprints, Adv. Neural Inf. Process. Syst. 28 (2015).
[57] Y. Xu, K. Lin, S. Wang, L. Wang, C. Cai, C. Song, L. Lai, J. Pei, Deep learning
for molecular generation, Future Med. Chem. 11 (6) (2019) 567–597.
[58] A. Jain, S. Ong, G. Hautier, W. Chen, W. Richards, S. Dacek, S. Cholia, D.
Gunter, D. Skinner, G. Ceder, et al., The materials project: A materials genome
approach to accelerating materials innovation, APL Mater. (2013).
[59] P. Vurtur Badarinath, M. Chierichetti, F. Davoudi Kakhki, A machine learning
approach as a surrogate for a finite element analysis: Status of research and
application to one dimensional systems, Sensors 21 (5) (2021) 1654.
[60] J. Mackerle, K. Orsborn, Expert systems for finite element analysis and design
optimization—a review, Eng. Comput. (1988).
[61] J. Huang, J. Liew, A. Ademiloye, K.M. Liew, Artificial intelligence in materials
modeling and design, Arch. Comput. Methods Eng. 28 (2021) 3399–3413.
[62] J. Yang, T. Dzanic, B. Petersen, J. Kudo, K. Mittal, V. Tomov, J.-S. Camier,
T. Zhao, H. Zha, T. Kolev, et al., Reinforcement learning for adaptive mesh
refinement, in: International Conference on Artificial Intelligence and Statistics,
PMLR, 2023, pp. 5997–6014.
[63] H.E. Balcıoğlu, A.Ç. Seçkin, Comparison of machine learning methods and finite
element analysis on the fracture behavior of polymer composites, Arch. Appl.
Mech. 91 (2021) 223–239.
14
Computational Materials Science 235 (2024) 112793
I. Papadimitriou et al.
[116] S. Fu, Z. Wan, W. Lu, H. Liu, P. Zhang, J. Tan, F. Pan, Z. Liu, High-accuracy
virtual testing of air conditioner’s digital twin focusing on key material’s
deformation and fracture behavior prediction, Sci. Rep. (2022).
[117] D.-Y. Jeong, M.-S. Baek, T.-B. Lim, Y.-W. Kim, S.-H. Kim, Y.-T. Lee, W.-S. Jung,
I.-B. Lee, Digital twin: Technology evolution stages and implementation layers
with technology elements, Ieee Access (2022).
[118] A. Mann, S.R. Kalidindi, Development of a robust CNN model for capturing
microstructure-property linkages and building property closures supporting
material design, Virtual Mater. Des. (2022) 879614107.
[119] S. Stieber, N. Schröter, A. Schiendorfer, A. Hoffmann, W. Reif, FlowFrontNet:
improving carbon composite manufacturing with CNNs, in: Joint European
Conference on Machine Learning and Knowledge Discovery in Databases,
Springer, 2021, pp. 411–426.
[120] W.A. Saidi, W. Shadid, I.E. Castelli, Machine-learning structural and electronic
properties of metal halide perovskites using a hierarchical convolutional neural
network, NPJ Comput. Mater. 6 (1) (2020) 1–7.
[121] D.-D. Chen, Y. Lin, F. Wu, A design framework for optimizing forming processing parameters based on matrix cellular automaton and neural network-based
model predictive control methods, Appl. Math. Model. 76 (2019) 918–937.
[122] T.D. Huan, A. Mannodi-Kanakkithodi, C. Kim, V. Sharma, G. Pilania, R.
Ramprasad, A polymer dataset for accelerated property prediction and design,
Sci. Data 3 (1) (2016) 1–10.
[123] Materials Project, https://materialsproject.org/, Accessed: 2010-09-30.
[124] Crystallography Open Database, http://www.crystallography.net/cod/, Accessed: 2010-09-30.
[125] Inorganic Crystal Structure Database, https://icsd.fiz-karlsruhe.de, (Accessed:
2010-09-30).
[126] MatWeb, http://www.matweb.com/, Accessed: 2010-09-30.
[127] Cambridge Structural Database, https://ccdc.cam.ac.uk, Accessed: 2010-09-30.
[128] International Centre for Diffraction Data, https://www.icdd.com, Accessed:
2010-09-30.
[129] GDB Databases, https://gdb.unibe.ch, Accessed: 2010-09-30.
[130] ZINC Database, https://zinc15.docking.org, Accessed: 2010-09-30.
[131] Automatic FLOW for Materials Discovery LIBrary, http://aflowlib.org, Accessed:
2010-09-30.
[132] Open Quantum Materials Database, http://oqmd.org, Accessed: 2010-09-30.
[133] Harvard Clean Energy Project, (website currently being migrated), Accessed:
2010-09-30.
[134] TE Design Lab, http://tedesignlab.org, Accessed: 2010-09-30.
[135] NREL Materials Database, https://materials.nrel.gov, Accessed: 2010-09-30.
[136] Materials Cloud, https://www.materialscloud.org/v, Accessed: 2010-09-30.
[137] F. Oviedo, J.L. Ferres, T. Buonassisi, K.T. Butler, Interpretable and explainable
machine learning for materials science and chemistry, Account. Mater. Res. 3
(6) (2022) 597–607.
[138] N.T.P. Hartono, J. Thapa, A. Tiihonen, F. Oviedo, C. Batali, J.J. Yoo, Z. Liu, R.
Li, D.F. Marrón, M.G. Bawendi, et al., How machine learning can help select
capping layers to suppress perovskite degradation, Nat. Commun. 11 (1) (2020)
4172.
[139] J. Wang, Y. Wang, Y. Chen, Inverse design of materials by machine learning,
Materials 15 (5) (2022) 1811.
[140] X. Zhong, B. Gallagher, S. Liu, B. Kailkhura, A. Hiszpanski, T. Han, Explainable
machine learning in materials science, NPJ Comput. Mater. 8 (1) (2022) 1–19.
[141] M.T. Horsch, B. Schembera, H.A. Preisig, European standardization efforts from
FAIR toward explainable-AI-ready data documentation in materials modelling,
Proc. ICAPAI (2023).
[94] C.M. Bishop, N.M. Nasrabadi, Pattern Recognition and Machine Learning, Vol.
4, Springer, 2006.
[95] J.D. Kelleher, B. Mac Namee, A. D’Arcy, Fundamentals of machine learning for
predictive data analytics: algorithms, in: Worked Examples, and Case Studies,
2015.
[96] D. Cannizzaro, A.G. Varrella, S. Paradiso, R. Sampieri, Y. Chen, A. Macii, E.
Patti, S. Di Cataldo, In-situ defect detection of metal additive manufacturing: an
integrated framework, IEEE Trans. Emerg. Top. Comput. 10 (1) (2021) 74–86.
[97] R.S. Sutton, A.G. Barto, Reinforcement Learning: An Introduction, MIT Press,
2018.
[98] J. Shahrabi, M.A. Adibi, M. Mahootchi, A reinforcement learning approach to
parameter estimation in dynamic job shop scheduling, Comput. Ind. Eng. 110
(2017) 75–82.
[99] W. Gu, S. Liu, Z. Zhang, Y. Li, A distributed physical architecture and data-based
scheduling method for smart factory based on intelligent agents, J. Manuf. Syst.
65 (2022) 785–801.
[100] Y. Lu, K.C. Morris, S. Frechette, et al., Current Standards Landscape for
Smart Manufacturing Systems, 8107, (3) National Institute of Standards and
Technology, NISTIR, 2016.
[101] M. Charest, R. Finn, R. Dubay, Integration of artificial intelligence in an
injection molding process for on-line process parameter adjustment, in: 2018
Annual Ieee International Systems Conference (Syscon), IEEE, 2018, pp. 1–6.
[102] K. Paraskevoudis, P. Karayannis, E.P. Koumoulos, Real-time 3D printing remote
defect detection (stringing) with computer vision and artificial intelligence,
Processes 8 (11) (2020) 1464.
[103] L. Monostori, Z.J. Viharos, Hybrid, AI-and simulation-supported optimisation of
process chains and production plants, CIRP Ann. 50 (1) (2001) 353–356.
[104] E.I. Vlahogianni, M.G. Karlaftis, J.C. Golias, Short-term traffic forecasting:
Where we are and where we’re going, Transp. Res. C 43 (2014) 3–19.
[105] P. Domingos, A few useful things to know about machine learning, Commun.
ACM 55 (10) (2012) 78–87.
[106] Z.C. Lipton, The mythos of model interpretability: In machine learning, the
concept of interpretability is both important and slippery, Queue 16 (3) (2018)
31–57.
[107] A. Kusiak, Smart manufacturing must embrace big data, Nature 544 (7648)
(2017) 23–25.
[108] S.R. Kalidindi, M. Buzzy, B.L. Boyce, R. Dingreville, Digital twins for materials,
Front. Mater. (2022).
[109] M. Deagen, L.C. Brinson, R.A. Vaia, L.S. Schadler, The materials tetrahedron
has a ‘‘digital twin’’, Mrs Bull. (2022).
[110] T. Chandhana, A. Balija, S.R. Kumaran, B. Singh, Digital twins-enabling
technologies including AI, sensors, cloud, and edge computing, in: Handbook
of Research on Applications of AI, Digital Twin, and Internet of Things for
Sustainable Development, IGI Global, 2023, pp. 306–331.
[111] K. Feng, T. Xi, The information visualization design of smart museums based
on the digital twin theory, Usability User Exp. 39 (39) (2022).
[112] R. Al-Sehrawy, B. Kumar, Digital twins in architecture, engineering, construction and operations. a brief review and analysis, in: Proceedings of the 18th
International Conference on Computing in Civil and Building Engineering:
ICCCBE 2020, Springer, 2021, pp. 924–939.
[113] R. Rosen, G. Von Wichert, G. Lo, K.D. Bettenhausen, About the importance of
autonomy and digital twins for the future of manufacturing, Ifac-papersonline
48 (3) (2015) 567–572.
[114] D. Piromalis, A. Kantaros, Digital twins in the automotive industry: The road
toward physical-digital convergence, Appl. Syst. Innov. 5 (4) (2022) 65.
[115] W. Purcell, T. Neubauer, Digital twins in agriculture: A state-of-the-art review,
Smart Agri. Technol. 3 (2023) 100094.
15
Download