Artificial Intelligence
Artificial Intelligence
Artificial
Intelligence in
Manufacturing
Enabling Intelligent, Flexible and
Cost-Effective Production Through AI
Artificial Intelligence in Manufacturing
John Soldatos
Editor
Artificial Intelligence
in Manufacturing
Enabling Intelligent, Flexible and
Cost-Effective Production Through AI
Editor
John Soldatos
Netcompany-Intrasoft S.A
Märel Luxembourg, Luxembourg
© The Editor(s) (if applicable) and The Author(s) 2024. This book is an open access publication.
Open Access This book is licensed under the terms of the Creative Commons Attribution 4.0 Inter-
national License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adaptation,
distribution and reproduction in any medium or format, as long as you give appropriate credit to the
original author(s) and the source, provide a link to the Creative Commons license and indicate if changes
were made.
The images or other third party material in this book are included in the book’s Creative Commons
license, unless indicated otherwise in a credit line to the material. If material is not included in the book’s
Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the
permitted use, you will need to obtain permission directly from the copyright holder.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors, and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
For over a decade, Artificial Intelligence (AI) technologies and applications are
proliferating in a rapid pace. The rise of AI is driven by a variety of factors including
the unprecedented improvements in hardware and software, and the explosion in the
amount of generated data. These advances enable the development of sophisticated
AI models (e.g., deep learning models, deep reinforcement learning models, large
language models), as well as their deployment and execution in realistic settings.
This is also the reason why modern manufacturers are undertaking significant
investments in AI solutions as part of their digital transformation journey. As a
result, AI is rapidly transforming the manufacturing industry, through enabling
tangible improvements in the efficiency, quality, and productivity of industrial
organizations.
A variety of AI-based use cases are nowadays deployed in Industry 4.0 pro-
duction lines. Some of the most prominent examples of such AI-enabled use
cases can be found in the areas of predictive maintenance, quality control, supply
chain optimization, production planning, process automation, and safety monitor-
ing. For instance, a variety of machine learning models are nowadays used to
make quality control more practical and more intelligent, by automating product
quality inspection, enabling timely detection of defects, and identifying production
configurations that could lead into production problems. As another example, deep
learning algorithms are commonly used to predict and anticipate machine failures
before they occur, based on predictive and accurate estimations of the Remaining
Useful Life (RUL) of the machinery. Likewise, there are AI systems that enable the
timely detection of anomalies in products and production processes.
These use cases are some of the most disruptive solutions of the Industry 4.0
era, which is transforming manufacturing enterprises by means of Cyber Physical
Production Systems (CPPS). In this direction, most AI use cases for Industry 4.0
emphasize the training, development, and deployment of accurate and effective
machine learning systems. The latter are integrated with Industrial Internet of
Things (IIoT) systems in the scope of scalable and secure cloud/edge environments.
To this end, industrial solution integrators leverage structuring principles and
blueprints specified in standards-based reference architectures for Industry 4.0
v
vi Preface
systems. Nevertheless, the AI’s potential for manufacturing is still largely underex-
ploited. State of the art systems are usually limited to the extraction of data-driven
AI-based insights for improving production processes and related decision making.
These insights are based on quite simple models about the production processes and
hardly combine capabilities of multiple AI systems and algorithms. To alleviate
these limitations, there are research initiatives that explore the integration and
collaboration of multiple AI systems in the scope of production processes. In this
direction, there is on-going research on:
• Multi-agent systems that foster enhanced collaborative intelligence based on the
interaction and the development of synergies across different autonomous AI
agents.
• Solutions for AI interoperability across diverse systems. These solutions leverage
advanced knowledge models (e.g., Semantic Knowledge Graphs (SKGs) and
embeddings that capture the relationships between different entities) to enable the
development of sophisticated AI systems that span entire multi-stage production
processes beyond simple ML-based state machines.
During the last couple of years, Industry 4.0 is evolving to a direction where
AI serves manufacturing workers, while at the same time interacting closely with
them in a variety of human-in-the-loop scenarios such as human-robot collaboration
(HRC) scenarios. At the same time, AI use cases are increasingly aiming at increas-
ing production sustainability to ensure that the manufacturing sector contributes to
strategic targets such as the European Green Deal (EGD) of the European Union
(EU). Sustainability and human-centricity are driving the transition of Industry 4.0
digital manufacturing systems to the Industry 5.0 era, which emphasizes human-
centricity and environmental performance.
The advent of Industry 5.0 systems is increasing the functional sophistication and
integration complexity of AI systems in manufacturing. It also asks for an evolution
of AI in a human-centered dimension, where AI systems operate in trustworthy and
reliable manner. Specifically, the evolution of AI systems toward the Industry 5.0
era asks for:
• Novel AI architectures for Industry 5.0: Industry 5.0 system comprises multiple
AI components (e.g., robots, machine learning models, Natural Language Pro-
cessing (NLP)) that must safely and effectively interact with humans in industrial
environments. The development and deployment of such systems requires novel
architectures and structuring principles, beyond classical architectures of Big
Data, AI, and Industrial Internet of Things (IIoT) platforms.
• Knowledge Modelling and Representation: HRC use cases are usually deployed
in highly dynamic environments involving humans, robots, and AI systems
that interact with each other. The implementation of advanced and automated
reasoning in such an environment asks for novel ways for representing processes
in ways that capture the complex interrelationships between the different actors.
• Models and Learning paradigms for Human-Robot Collaboration: Industry 5.0
introduces a need for deploying novel learning paradigms that foster the interplay
Preface vii
between humans and AI actors. Such paradigms include, for example, active
learning and intelligent Multi-Agent Systems (MAS). They enable outcomes
that combine the speed of AI systems with the credibility of human judgment.
Likewise, Industry 4.0 solutions like Digital Twins are currently transformed
to account for the context of the human workers, i.e., they are evolving toward
human-centric digital twins.
• Explainability, transparency, and regulatory compliance: Industry 5.0 systems
pose their own unique transparency and safety requirements. They involve
humans in the loop that must be able to understand the decisions and operation
of AI system. Hence, AI use cases cannot be developed based on black-box
AI models. Rather, AI systems should be transparent, explainable, trusted, and
understandable to humans. Manufacturers must also ensure that their AI systems
adhere to the mandates of emerging AI regulations such as the AI Act in Europe.
The aim of this book is to shed light on the limitations of existing solutions for
AI in manufacturing and to introduce novel solutions that:
• Improve the functional capabilities and technical performance of state-of-the-art
AI systems for manufacturing in a variety of production processes like production
scheduling and quality control
• Enhance the human centricity, the trustworthiness, and the overall social perfor-
mance of AI systems in line with the requirements and concepts of the Industry
5.0 era
The book comprises 27 chapters that present innovative AI systems and solutions
spanning both state-of-the-art Industry 4.0 use cases and emerging, human-centric
Industry 5.0 use cases. The chapters are contributed by Y EU-funded projects,
which are closely collaborated in the context of the AI4Manufacturing Cluster
of European projects, as well as in the scope of the activities of the European
Factories of the Future Research Association (EFFRA). The contributing projects
focus on the development, deployment, and operation of AI systems for production
lines. Each of the project addresses a set of unique challenges of AI in Industry
4.0 and/or Industry 5.0 use cases, such as the development and deployment of
effective MAS systems, the development of trusted and explainable AI systems, the
specification and implementation of knowledge models and semantics for Industry
5.0 applications, as well as the development of novel forms of digital twin systems
and applications (e.g., human-centric digital twins).
Specifically, the book is structured in the following three parts:
Part I: Architectures and Knowledge Modelling for AI
This part presents architectures for AI-based Industry 5.0 systems and solutions,
ranging from high-level reference architecture models to architecture of specific
AI platforms and solutions’ marketplaces. The presented architectures illustrate the
structure of both Industry 4.0 and Industry 5.0 use cases with emphasis on the
structuring principles that drive the integration of AI and ML models with industrial
systems. Moreover, this part of the book includes several chapters that illustrate
viii Preface
This book has received funding from the European Union’s Horizon 2020
research and innovation program under grant agreements No. 956573 (STAR),
No. 957204 (MAS4AI), No. 957362 (XMANAI), No. 101000165 (ASSISTANT),
No. 957331 (knowlEdge), and No. 957402 (Teaming. AI), which are part of the
AI4Manufacturing Cluster of projects. Several contributions of the book have been
also supported financially from other EU projects (e.g., No. 952119 (KITT4SME)
No. 870092 (DIMOFAC), No. 869963 (MERGING)) and various national projects
as indicated in the acknowledgement sections of each chapter.
The editor and the chapter co-authors acknowledge valuable support from
partners of the above-listed EU projects.
Disclaimer: The contents of the book reflect only the contributors’ and co-
authors’ view. The European Commission is not responsible for any use that may
be made of the information it contains.
ix
Contents
xi
xii Contents
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 503
Editor and Contributors
xv
xvi Editor and Contributors
Contributors
Elmar Kiesling WU, Institute for Data, Process and Knowledge Management,
Vienna, Austria
Timotej Klemenčič University of Ljubljana, Ljubljana, Slovenia
Kosmas Alexopoulos Laboratory for Manufacturing Systems & Automation
(LMS), Department of Mechanical Engineering & Aeronautics, University of
Patras, Rio-Patras, Greece
Franz Krause University of Mannheim, Data and Web Science Group, Mannheim,
Germany
Kabul Kurniawan WU, Institute for Data, Process and Knowledge Management,
Vienna, Austria
Austrian Center for Digital Production (CDP), Vienna, Austria
Fenareti Lampathaki Suite5 Data Intelligence Solutions, Limassol, Cyprus
Giuseppe Landolfi University of Applied Science of Southern Switzerland,
Switzerland
Eleni Lavasa Athena Research Center, Marousi, Greece
Maria Chiara Leva Technological University Dublin, School of Environmental
Health, Dublin, Ireland
Ainhoa Etxabarri Llana UNIMETRIK S.A., Legutiano,Álava, Spain
Afra Maria Petrusa Llopis AIMEN Technology Centre, Smart Systems and
Smart Manufacturing Group, Pontevedra, Spain
Alberto Botana López AIMEN Technology Centre, Smart Systems and Smart
Manufacturing Group, Pontevedra, Spain
Pedro Lopez Fundacion TECNALIA R&I, Madrid, Spain
Andreas Louca Suite5 Data Intelligence Solutions, Limassol, Cyprus
Paul Lukowicz German Research Center for Artificial Intelligence (DFKI),
Kaiserslautern, Germany
Department of Computer Science, RPTU Kaiserslautern-Landau, Kaiserslautern,
Germany
Daniel Gordo Martín AIMEN Technology Centre, Smart Systems and Smart
Manufacturing Group, Pontevedra, Spain
Andrea Fernández Martínez AIMEN Technology Centre, Smart Systems and
Smart Manufacturing Group, Pontevedra, Spain
Jorge Martinez-Gil Software Competence Center Hagenberg GmbH, Hagenberg,
Austria
Pablo A. Martin High Performance and Artificial Intelligence, Barcelona Super-
computing Center, Barcelona, Spain
xx Editor and Contributors
Jawad Masood AIMEN Technology Centre, Smart Systems and Smart Manufac-
turing Group, Pontevedra, Spain
Antonello Meloni Mathematics and Computer Science Department, University of
Cagliari, Cagliari, Italy
Markku Mikkola VTT Technical Research Centre of Finland Ltd., Espoo, Finland
Elena Minisci CRIT S.R.L., Vignola, Italy
Dunja Mladenić Jožef Stefan Institute, Ljubljana, Slovenia
Elias Montini University of Applied Science of Southern Switzerland, Switzerland
Politecnico di Milano, Milan, Italy
David Monzo Tyris AI, Valencia, Spain
Bernhard Moser Software Competence Center Hagenberg GmbH, Hagenberg,
Austria
William Motsch Deutsches Forschungszentrum für Künstliche Intelligenz GmbH
(DFKI), Kaiserslautern, Germany
Serafeim Moustakidis AIDEAS OU, Tallinn, Estonia
Santiago Muiños-Landin AIMEN Technology Centre, Smart Systems and Smart
Manufacturing Group, Pontevedra, Spain
Marco Murgia Mathematics and Computer Science Department, University of
Cagliari, Cagliari, Italy
Jose Angel Segura Muros AIMEN Technology Centre, Smart Systems and Smart
Manufacturing Group, Pontevedra, Spain
Linda Napoletano Deep Blue, Rome, Italy
Nikolaos Nikolakis Laboratory for Manufacturing Systems & Automation (LMS),
Department of Mechanical Engineering & Aeronautics, University of Patras, Rio-
Patras, Greece
Nikoletta Nikolova Data Science, Netherlands Organisation for Applied Scientific
Research (TNO), Den Haag, The Netherlands
Alexandros Nizamis Centre for Research and Technology Hellas, Information
Technologies Institute (CERTH/ITI), Thessaloniki, Greece
Parsha Pahlevannejad German Research Center for Artificial Intelligence
(DFKI), Kaiserslautern, Germany
Technologie-Initiative SmartFactory, Kaiserslautern, Germany
Panagiotis Mavrothalassitis Laboratory for Manufacturing Systems & Automa-
tion (LMS), Department of Mechanical Engineering & Aeronautics, University of
Patras, Rio-Patras, Greece
Editor and Contributors xxi
xxiii
xxiv Abbreviations
1 Introduction
For over a decade, manufacturing enterprises are heavily investing in their digital
transformation based on Cyber-Physical Production Systems (CPPS) that enable
the digitization of production processing such as production scheduling, products’
assembly, physical assets’ maintenance, and quality control. The deployment and
operation of CPPS in the manufacturing shopfloor is the main enabler of the fourth
industrial revolution (Industry 4.0) [1], which boosts automation and efficiency
toward improving production speed and quality [2], while lowering production costs
and enabling novel production models such as lot-size-one manufacturing and mass
customization.
Industry 4.0 applications are usually developed based on advanced digital
technologies such as Big Data, Internet of Things (IoT), and Artificial Intelligence
(AI), which are integrated with CPPS systems in the manufacturing shopfloor
and across the manufacturing value chain. In cases of nontrivial Industry 4.0
systems, this integration can be challenging, given the number and the complexity
of the systems and technology involved. For instance, sophisticated Industry 4.0
use cases are likely to comprise multiple sensors and automation devices, along
with various data analytics and AI modules that are integrated in digital twins
(DTs) systems and applications. To facilitate such challenging integration tasks,
industrial automation solution providers are nowadays offered with access to various
reference architecture models for Industry 4.0 applications. These models illustrate
the functionalities and technological building blocks of Industry 4.0 applications,
while at the same time documenting structuring principles that facilitate their
integration and deployment in complete systems and applications. Some of these
reference architecture models focus on specific aspects of Industry 4.0 such as data
collection, data processing, and analytics, while others take a more holistic view that
addresses multiple industrial functionalities. Moreover, several architecture models
address nonfunctional requirements as well, such as the cybersecurity and safety of
industrial systems.
During the last couple of years, there is a surge of interest on Industry 4.0
applications that emphasize human-centered industrial processes, i.e., processes
with the human in the loop, as well as the achievement of ambitious sustainability
and resilience objectives. The latter are at the very top of the policy agenda of the
European Union, as reflected in the European Green Deal (EGD) and Europe’s
Circular Economy Action Plan (CEAP). This has led to the introduction of the term
Industry 5.0, which evolves Industry 4.0 in a direction that complements efficiency
and productivity goals with societal targets, notably contributions to sustainability
and the workers’ well-being [3]. Hence, Industry 5.0 targets a sustainable, human-
centric, and resilient industry [4]. In this direction, Industry 4.0 systems must be
enhanced with human-centric technologies that put the worker at the center of the
production process, while at the same time fostering security, safety, transparency,
and trustworthiness. For instance, the shift from Industry 4.0 to Industry 5.0 asks for
the deployment and use of transparent, interoperable, and explainable AI systems
[5], beyond black-box systems (e.g., deep neural networks) that are typically used in
Industry 4.0 deployments. As another example, Industry 5.0 applications comprise
technological paradigms that foster the collaboration between humans and industrial
systems (e.g., co-bots), rather than systems that aim at replacing the human toward
hyper-automation (e.g., fully autonomous industrial robots). Likewise, the scope
of digital twins in Industry 5.0 tends to differ from Industry 4.0, as simulations
and what-if analysis account for human parameters (e.g., physical characteristics,
emotional status, skills) as well. Also, Industry 5.0 pays greater emphasis on
nonfunctional requirements such as data protection, security, and safety when
compared to Industry 4.0 that prioritizes industrial performance and accuracy.
Despite these technological differences between Industry 5.0 and Industry 4.0
systems, there is still a lack of standards, formal guidelines, and blueprints for devel-
oping, deploying, and operating Industry 5.0 systems. In most cases, manufacturers
and providers of industrial automation solutions make use of conventional Industry
4.0 and blueprints, which they enhance with the required features and functionalities
of their Industry 5.0 use cases at hand. We argue that this is a considerable
misstep in the process of designing and implementing Industry 5.0 solutions, as
it deprives architects and developers of industrial systems of the opportunity to
consider Industry 5.0 functionalities and features from the early stages of the
industrial systems’ development. State-of-the-art approaches to developing Industry
5.0 systems start from Industry 4.0 reference architectures and address human-
Reference Architecture for AI-Based Industry 5.0 Applications 5
2 Relevant Work
the context of the human user toward customizing the industrial functionalities
accordingly.
The Industrial Internet Security Framework (IISF) complements the IIRA with a
security viewpoint for industrial systems [9]. One of the main objectives of the IISF
is to prescribe the functions needed for the development, deployment, and operation
of trusted IIoT. These functions are also essential for ensuring the trustworthiness
of Industry 5.0 systems and their AI components in industrial environments. Thus,
the structure and functions of IISF provided inspiration about how to support AI
trustworthiness for industrial use cases. The IISF specifies functionalities that secure
all the different elements of an industrial system such as the various communication
endpoints of the system. Most of these functions can be used to boost the security
and trustworthiness of Industry 5.0 systems as well, as they safeguard the operation
of the networks, the data, and the data processing functions of Industry 5.0 systems.
Specifically, the IISF is concerned with the five main characteristics that affect
the trustworthiness of IIoT deployments, i.e., security, safety, reliability, resilience,
and privacy. The framework specifies a functional viewpoint that is destined to
secure IIoT systems compliant to the IIRA. To this end, the functional viewpoint
specifies six interacting and complementary building blocks, which are organized in
a layered fashion. The top layer comprises four security functions, namely endpoint
protection, communications and connectivity protection, security monitoring and
analysis, and security configuration management. Likewise, a data protection layer
and a system-wide security model and policy layer are specified. Each one of
the functional building blocks of the IISF can be further analyzed in more fine-
grained functions such as monitoring functionalities, data analytics functionalities,
and actuation functionalities. Each of these three types of functionalities include
security-related functions.
One more reference architecture for industrial systems, notably for fog comput-
ing systems, was introduced by the OpenFog Consortium prior to its incorporation
within the Industrial Internet Consortium in 2019 [10]. The OpenFog RA specifies
the structure of large-scale fog computing system with emphasis on how fog nodes
are connected to enhance the intelligence and to boost the efficiency of Industrial
IoT systems. The OpenFog RA specifies some cross-cutting functionalities, which
are characterized as “perspectives.” One of these perspectives deals with the security
functionalities, which implies that security is applicable to all layers and use
scenarios from the hardware device to the higher software layers of the architecture.
As already outlined, such security functions are key to the development and
deployment of trusted industrial systems of the Industry 5.0 era.
The Big Data Value Association (BDVA) has specified the structure of big data
systems based on the introduction of a reference model for big data systems [11].
The model illustrates a set of modules that are commonly used in big data systems
along with structuring principles that drive their integration. The BDVA reference
model consists of the following layers:
• Horizontal layers that illustrate the modules and the structure of data processing
chains. The modules of data processing chains support functions such as data
8 J. Soldatos et al.
collection, data ingestion, data analytics, and data visualization. The horizontal
layers do not map to a layered architecture, where all layers must coexist in
the scope of a system. For instance, it is possible to have a data processing
chain that leverages data collection and visualization collection functions without
necessarily using data ingestion and data analytics functionalities. Hence, the
BDVA horizontal layers can be used as building blocks to construct data pipelines
for AI systems.
• Vertical layers that deal with cross-cutting issues such as cybersecurity and trust.
The latter are applicable to all functionalities of the horizontal layers. Vertical
layers can be also used to specify and address nontechnical aspects such as the
ever important legal and regulatory aspects of AI systems.
The horizontal and vertical layers of the reference model are used to produce
concrete architectures for big data systems. There are clearly many commonalities
between big data and AI systems as many AI systems (e.g., deep learning systems)
are data-intensive and process large amounts of data. The BDVA RA does not,
however, address functionalities that foster the development of Industry 5.0 systems,
such as data quality and AI model explainability functionalities. As such it is
mostly appropriate for architecting AI systems without special provisions for their
trustworthiness and human centricity.
Standards-based functionalities for AI systems are also specified by the ISO/IEC
JTC 1/SC 42 technical committee on Artificial Intelligence [12]. The committee
has produced several standards that cover different aspects of AI systems, such
as data quality for analytics and machine learning (ML) (i.e., ISO/IEC DIS 5259-
1), transparency taxonomy of AI systems (i.e., ISO/IEC AWI 12792), a reference
architecture of knowledge engineering (i.e., ISO/IEC DIS 5392), functional safety
and AI systems (i.e., ISO/IEC CD TR 5469), as well as objectives and approaches
for explainability of ML models and AI systems (i.e., ISO/IEC AWI TS 6254).
As evident from the above-listed descriptive titles, the ISO/IEC JTC 1/SC 42
technical committee addresses human centricity (e.g., safety, explainability) and
trustworthiness (e.g., data quality, explainability) issues for Industry 5.0 systems.
Nevertheless, most of the relevant standards are under development and not yet
available for practical applications and use.
In recent years, the IEEE (Institute of Electrical and Electronics Engineers) SA
(Standards Association) is developing a suite of standards (i.e., standards of the
IEEE 7000 family) that deal with the ethical aspects of AI systems. For instance,
the IEEE 7000-2021 standard titled “IEEE Standard Model Process for Addressing
Ethical Concerns during System Design” [13] specifies a process that organizations
can follow to ensure that their AI systems adhere to ethical values and integrate
ethical AI concepts within their systems’ development lifecycle. These standards
can facilitate the development of Industry 5.0. However, they are mostly focused
on development, deployment, and operational processes rather on how to structure
Industry 5.0 systems.
Overall, there is a still a lack of standards-based architectures and blueprints
for the development of Industry 5.0 systems. Hence, AI developers, deployers, and
Reference Architecture for AI-Based Industry 5.0 Applications 9
engineers have no easy ways to structure, design, and build nontrivial trustworthy
human-centric AI systems [14].
The safety domain of the reference model outlines functionalities that are key to
ensuring the safe operation of AI systems (e.g., robots) in the scope of Industry 5.0
scenarios. It identifies the following indicative but important functionalities:
• Object localization and tracking: This functionality aims at identifying the
location of objects within industrial environments, notably of moving objects
such as mobile robots. The localization and tracking of such functionalities are
integral elements of applications that safeguard the safe operation of robotics and
AI systems in industrial environments.
• Safety zones detection: Automation systems that move within an industrial
environment (e.g., shopfloor, plant floor) must follow safe trajectories that
minimize the risk of collisions between automation systems, humans, and other
stationary elements of the workplace. In this direction, the detection of safety
zones based on AI technologies (e.g., reinforcement learning [23]) can increase
the safety of the AI systems and minimize related risks.
• Safe automatic mobile robots: This functionality is a placeholder for systems
that ensure the safe movement of automatic mobile robots. The implementation
of this functionality can benefit from other functionalities of this domain such as
the detection of safety zones.
• Worker safety: Apart from ensuring the safe operation and movement of robotic
systems, it is important to ensure the safety of the workers. Workers’ safety is at
the heart of Industry 5.0 system that emphasize human centricity. The respective
functionalities ensure that workers act within safe environments and that the
emotional and physical context of the human is properly considered in the design,
development, deployment, and operation of AI systems.
Reference Architecture for AI-Based Industry 5.0 Applications 13
The main modules and building blocks of the architecture are illustrated in the
following subparagraphs.
The architecture enables the development of secure, safe, and trusted AI systems
in production lines. To this end, systems compliant to the STAR-RA collect and
process data from AI-based systems in the shopfloor, including machines, robotic
cells, AMRs, and other digital manufacturing platforms. Industry 5.0 systems
comprise various CPPS systems and digital manufacturing platforms that serve as
data sources for other logical modules. The latter may also consume data from
other data sources in the shopfloor such as business information systems (e.g.,
ERP (Enterprise Resource Planning)) and manufacturing databases (e.g., historian
systems).
This building block provides a secure data collection solution that offers a real time
data collection, transformation, filtering, and management services to facilitate data
consumers (e.g., the AI cyber-defense module and the Security Policies Manager)
in accessing the required data. For example, it can be used to collect security-related
data (e.g., network, system, and solution proprietary) from monitored IoT systems
and store them to detect patterns of abnormal behavior by applying simple (i.e.,
filtering) or more elaborate (i.e., deep learning) data processing mechanisms. The
solution features specialized probes that may be deployed within the monitored
IoT/CPPS system or poling services for acquiring data from shopfloor sources such
as CPPS systems and digital manufacturing platforms. The module belongs to the
cybersecurity domain of the high-level reference architecture.
16 J. Soldatos et al.
The DPT module belongs to the cybersecurity domain and provides the means
for tracking and tracing industrial data. It interfaces to the data probes to acquire
information about the industrial data of the shopfloor such as information about
data types, volumes, and timestamps. Accordingly, it records this information (i.e.,
the metadata) about the acquired data to facilitate the detection of data abuse and
data tampering attempts. Specifically, data ingested in the DPT can be queried by
other modules to facilitate the validation of datasets and to ensure that the data
they consume have not been falsified. In this way, the DPT module reinforces the
reliability and the security of the industrial data that flow through the system.
There are different ways for implementing a DPT infrastructure for industrial data.
The STAR-RA promotes a decentralized approach, which leverages the benefits of
distributed ledger technologies, i.e., a blockchain protocol. Distributed ledger infras-
tructures offer some advantages for industrial data provenance, such as immutable
and tamper-resistant records. They also provide a comprehensive and auditable
trail that records the history of data transactions, including creation, modification,
and transfer events. In addition, blockchains enable the implementation of Smart
Contracts (SC) over the distributed ledger infrastructure, notably SCs that are used
to validate the metadata of the industrial datasets that are recorded in the blockchain.
SCs enable decentralized applications that provide information about the metadata
to interested modules such as the cyber-defense strategies module.
This module implements cyber-defense strategies for AI systems, i.e., strategies that
protect AI systems against adversarial attacks. These strategies operate based on
access to industrial data from:
• The AI systems (including ML systems) that must be protected from cybersecu-
rity attacks.
• The CPPS and digital manufacturing platforms that act as data sources.
• The metadata of the industrial data that are managed by the DPT module and its
blockchain implementation.
• The explainable AI (XAI) module, which implements explainable AI models that
illustrate and interpret the operation of various AI systems and algorithms.
The module materializes different strategies in response to attacks against AI sys-
tems. For instance, it implements cyber-defense strategies for poisoning and evasion
attacks. Nevertheless, additional cyber-defense strategies can be implemented and
integrated with the rest of the modules (i.e., secure networked data probes, DPT).
Reference Architecture for AI-Based Industry 5.0 Applications 17
A data integration infrastructure (e.g., based on a data bus pattern) can be used as a
data exchange middleware infrastructure to facilitate data transfer and data sharing
across different modules involved in the detection of a cybersecurity attacks, i.e.,
the DPT, the ACDS, and the SNP.
This module implements the security risk assessment and mitigation service of the
STAR-RA in-line with the cybersecurity domain of the high-level architecture. The
module assesses risk for assets associated with AI-based systems in manufacturing
lines. In this direction, it interacts with the AI cyber-defense strategies modules as
follows: (1) the defense strategies communicate to the RAME information about
identified risks for AI assets; and (2) the RAME consumes information from the
DPT to assess risks. It also offers mitigation actions for the identified risks such as
changes to the configuration of a probe via the SNP module.
This module defines and configures security policies that govern the operation
of the DPT, AI cyber-defense, and the RAME modules. Specifically, the module
specifies security policies that provide information about the probes and data
sources to be integrated, the configurations of the probes, as well as the cyber-
defense strategies to be deployed. By changing the applicable policies, the SPM
changes the configuration and the operation of other modules of the cybersecurity
domain (e.g., DPT, RAME, ACDS). The operation of the SPM is supported by a
Security Policies Repository (SPR), where policy files are persisted. Furthermore,
the SPM offers a GUI (graphic user interface) to the security officers of the factory
(e.g., members of CERT (computer emergency response teams)).
This module provides and executes XAI models and algorithms. It provides the
means for executing different types of XAI algorithms such as algorithms for
explaining deep neural networks and general-purpose algorithms (e.g., LIME –
Local Interpretable Model-Agnostic Explanations) [24] that explain the outcomes
of AI classifiers. As such the module is a placeholder of XAI techniques. The
XAI module provides its services to several other modules that leverage explainable
algorithms, such as the AI cyber-defense strategies module and the simulated reality
(SR) modules.
This module provides a placeholder for machine learning paradigms that foster
HRC, i.e., modules of the HRC domain of the high-level reference architecture.
Such paradigms include active learning and neurosymbolic learning, which help
robots and AI systems to benefit from human expertise in the context of human in
the loop industrial processes. These machine learning techniques for HRC fall in the
scope of the HRC domain of the high-level architecture.
This module consolidates domain knowledge about the production processes of the
manufacturing environment. It is used for inferencing by the other modules such as
the AL and neurosymbolic learning modules. The latter modules can interact and
update the module with knowledge acquired by the humans in the scope of human-
in-the-loop processes. Therefore, it also falls in the scope of the HRC domain.
Reference Architecture for AI-Based Industry 5.0 Applications 19
AMR Safety
This module comprises RL techniques that boost the safety of AMRs in industrial
environments such as manufacturing shopfloors. It provides insights on the safe
placement of robots in a manufacturing environment. To this end, it incorporates
functionalities such as objective localization and safety zones detection of the safety
domain of the high-level architecture of Fig. 1.
This module implements a digital twin that factors human-centered parameters (e.g.,
fatigue, emotional status of the worker). It is a placeholder for digital twins of
human-centered processes, including AI-based processes that have the human in the
loop. It interacts with the analytics platforms, the workers, and the humans’ digital
models.
The HDT offers a centralized access point to exploit a wide set of workers’
related data. It leverages a digital representation of the workers, which is seamlessly
integrated with production system DTs. The latter can be exploited by AI-based
modules to compute complex features that, enriching the HDT, enable better
decisions, and dynamically adapt automation systems behavior toward improving
production performance, workers’ safety, and well-being.
This module persists and manages data about the human worker toward supporting
the construction, deployment, and operation of HDTs. They provide the means for
creating and using digital representations of the workers.
This module provides a GUI interaction modality between factory workers and AI
systems. It comprises visualization elements (e.g., dashboards), while enabling users
to interact with the AI-based modules (e.g., provide form-based input).
This module enables NLP interactions between the factory users and relevant AI
modules (e.g., AL modules). It is a placeholder for different NLP implementations
and interfaces.
20 J. Soldatos et al.
Feedback Module
This module coordinates the provision of feedback from the human worker to
the AI system. It is particularly important for the implementation of human–AI
systems interactions (e.g., HRC scenarios). The feedback module interfaces to some
interaction module (e.g., GUI or NLP) that enables the transferring of user data to
the feedback module and vice versa.
This module leverages sensors and IoT devices (e.g., electroencephalography (EEG)
sensors) to collect information about the worker’s fatigue. The collected information
is transferred to other modules such as the human models and the HDT.
The STAR-RA for Industry 5.0 applications can be used to support the imple-
mentation of popular secure and trustworthy data-driven use cases in industrial
environments. In this direction, selected functional modules of the STAR-RA can
be deployed and operated. The specification of the modules and the information
flows that can support specific HRC, cybersecurity, and safety solutions in Industry
5.0 context can be defined as blueprints over the introduced architecture. Each
blueprint provides a proven way to implement trusted data processing and AI
functionalities for industrial applications. Rather than having to read, browse and
understand the entire STAR-RA and its low-level technical details, interested parties
(e.g., solution integrators, manufacturers, researchers in industrial automation, and
digital manufacturing) could consult blueprints as practical ways for enhancing
the trustworthiness and regulatory compliance of their work. Following paragraphs
illustrate blueprints for popular technical solutions and for the adherence of Industry
5.0 to the AI regulation proposal of the European Parliament and the Council of
Europe.
Trigger (Re-)Training
Process
Model check
Acquire Training Data
Training Data
Analytics
Detection of
malformed
instances
Regulatory compliance blueprints illustrate how the STAR-RA and its component
could be leveraged to boost the adherence of AI solutions to the AI regulation
22 J. Soldatos et al.
proposal of the European Commission. In April 2021, the European Parliament and
the Council of Europe presented an initial proposal for the regulation of AI systems
[6]. This proposal is the first organized and structured effort to regulate AI systems
worldwide. Its importance for systems deployed within Europe is particularly high,
given that it lays a basis for future laws within the various EU member states. The
proposal establishes a technology-neutral definition of AI systems in EU law, while
presenting a risk-based classification of AI systems. The classification proposes to
categorize AI systems in four general classes, ranging from unacceptable risk to
no risk (i.e., risk free) systems. It also outlines the requirements and obligations
associated with the deployment of systems from each one of the envisaged risk
levels. For instance, “high-risk” AI systems can be authorized if and only if they
meet requirements spanning the areas of transparency, explainability, data quality,
Reference Architecture for AI-Based Industry 5.0 Applications 23
and more. These obligations are significantly lower for medium- and low-risk
systems.
STAR-RA includes technical components that can help AI deployers and
operators to meet regulatory requirements and obligations. Different components
can be used to support systems belonging to the different risk classes of the AI
Act. For example, the AI Act specifies that minimal-risk systems (e.g., ML-based
calculations and visualization of information about physical assets) can be deployed
without essential restrictions. There are no mandatory obligations for minimal risk
systems. Compliance to AI code of conduct for them is recommended, yet optional.
Deployers may therefore choose to deploy one or more STAR components from
the different domains of the platform (cybersecurity, human–robot collaboration,
safety), as well as explainable AI components as illustrated in the relevant regulatory
blueprint of Table 3.
On the other hand, when deploying a limited risk system, AI deployers must
ensure that they are meeting transparency obligations. In this direction, humans must
be notified of the existence of an AI system component in the loop of the industrial
process. This concerns industrial processes with the human in the loop, where AI
systems and human interact. It is, for example, the case of some HDT applications
where industrial systems collect information about the status of the worker and
adapt their operations to it. The STAR-RA includes XAI components that can
help deployers meet the requirements of limited risk deployments. Deployers can
optionally use other STAR components to increase the safety, security, and overall
trustworthiness of the AI system (see Table 4).
Many AI systems in manufacturing and other industrial environments can be
classified as being high risk. This is, for example, the case with systems involving
AMRs and other types of industrial robots. In the case of high-risk systems,
24 J. Soldatos et al.
deployers and operators must comply with a longer list of requirements, including
more stringent specifications regarding explainability, transparency, data quality,
and more. To support the qualification, deployment and use of such high-risk
systems, STAR-RA offers many relevant components that support data reliability,
AI algorithms’ reliability, increased cybersecurity, safe human–robot collaboration,
and more. The use of these systems in a high-risk AI context becomes mandatory
rather than optional. This is illustrated in Table 5.
5 Conclusions
important for companies that seek to comply with the AI Act and to demonstrate
regulatory readiness. The inclusion of this chapter in this open access book aspires
to raise awareness about both the technical and the regulatory aspects of trustworthy
and human-centered solutions for manufacturing.
Acknowledgments This work has been carried out in the H2020 STAR project, which has
received funding from the European Union’s Horizon 2020 research and innovation program under
grant agreement No. 956573. The authors acknowledge valuable support from all partners of the
project.
References
1. Soldatos, J., Lazaro, O., Cavadini, F. (eds.): The Digital Shopfloor: Industrial Automation
in the Industry 4.0 Era River Publishers Series in Automation, Control and Robotics. River
Publishers, Gistrup (2019) ISBN: 9788770220415, e-ISBN: 9788770220408
2. Christou, I., Kefalakis, N., Soldatos, J., Despotopoulou, A.: End-to-end industrial IoT platform
for Quality 4.0 applications. Comput. Ind. 137, 103591. ISSN 0166-3615 (2022). https://
doi.org/10.1016/j.compind.2021.103591
3. European Commission.: Industry 5.0 – what this approach is focused on, how it will be
achieved and how it is already being implemented. European Commission. https://research-
and-innovation.ec.europa.eu/research-area/industrial-research-and-innovation/industry-50_en
(2023). Last accessed 26 June 2023
4. Nahavandi, S.: Industry 5.0—a human-centric solution. Sustainability. 11, 4371 (2019). https:/
/doi.org/10.3390/su11164371
5. Dwivedi R., Dave D., Naik H., Singhal S., Omer R., Patel P., Qian B., Wen Z., Shah T., Morgan
G., Ranjan R.: Explainable AI (XAI): Core ideas, techniques, and solutions. ACM Comput.
Surv. 55, 9 (2023), Article 194 (September 2023), 33 pages. https://doi.org/10.1145/3561048
6. Proposal for a Regulation of The European Parliament and of the Council laying down
harmonized rules on Artificial Intelligence (Artificial Intelligence Act) and Amending Certain
Union Legislative Acts, COM/2021/206 final. https://eur-lex.europa.eu/legal-content/EN/TXT/
?uri=celex%3A52021PC0206. Last accessed 26 June 2023
7. Industrial Internet Consortium.: The Industrial Internet Reference Architecture v 1.9. Available
at: https://www.iiconsortium.org/IIRA.htm. Last accessed 26 June 2023
8. ISO/IEC/IEEE.: ISO/IEC/IEEE 42010:2011 Systems and software engineering – Architec-
ture description. http://www.iso.org/iso/catalogue_detail.htm?csnumber=50508 (2011). Last
accessed 26 June 2023
9. Industrial Internet Consortium.: The Industrial Internet Security Framework. Available at:
https://www.iiconsortium.org/IISF.htm. Last accessed 26 June 2023
10. IEEE Standard for Adoption of OpenFog Reference Architecture for Fog Computing. IEEE
Std 1934-2018, 1–176, 2 Aug 2018. https://doi.org/10.1109/IEEESTD.2018.8423800
11. European Big Data Value Strategic Research and Innovation Agenda, Version 4.0, October
2017. https://bdva.eu/sites/default/files/BDVA_SRIA_v4_Ed1.1.pdf. Last accessed 26 June
2023
12. ISO/IEC JTC 1/SC 42 Artificial intelligence. https://www.iso.org/committee/6794475.html.
Last accessed 26 June 2023
13. IEEE Standard Model Process for Addressing Ethical Concerns during System Design. IEEE
Std 7000-2021, 1–82, 15 Sept 2021. https://doi.org/10.1109/IEEESTD.2021.9536679
14. Rožanec, J., Novalija, I., Zajec, P., Kenda, K., Tavakoli, H., Suh, S., Veliou, E., Papamartzi-
vanos, D., Giannetsos, T., Menesidou, S., Alonso, R., Cauli, N., Meloni, A., Reforgiato,
R.D., Kyriazis, D., Sofianidis, G., Theodoropoulos, S., Fortuna, B., Mladenić, D., Soldatos,
26 J. Soldatos et al.
J.: Human-centric artificial intelligence architecture for industry 5.0 applications. Int. J. Prod.
Res. (2022). https://doi.org/10.1080/00207543.2022.2138611
15. Liu, Z., Wang, J., Gong, S., Tao, D., Lu, H.: Deep reinforcement active learning for
human-in-the-loop person re-identification. In: 2019 IEEE/CVF International Conference
on Computer Vision (ICCV), Seoul, Korea (South), pp. 6121–6130 (2019). https://doi.org/
10.1109/ICCV.2019.00622
16. Siyaev, A., Valiev, D., Jo, G.-S.: Interaction with industrial digital twin using neuro-symbolic
reasoning. Sensors. 23, 1729 (2023). https://doi.org/10.3390/s23031729
17. Díaz-Rodríguez, N., Lamas, A., Sanchez, J., Franchi, G., Donadello, I., Tabik, S., Filliat, D.,
Cruz, P., Montes, R., Herrera, F.: EXplainable Neural-Symbolic Learning (X-NeSyL) method-
ology to fuse deep learning representations with expert knowledge graphs: the MonuMAI
cultural heritage use case. Inf. Fusion. 79, 58–83., ISSN 1566-2535 (2022). https://doi.org/
10.1016/j.inffus.2021.09.022
18. Soldatos, J., Despotopoulou, A., Kefalakis, N., Ipektsidis, B.: Blockchain based data prove-
nance for trusted artificial intelligence. In: Soldatos, J., Kyriazis, D. (eds.) Trusted Artificial
Intelligence in Manufacturing: A Review of the Emerging Wave of Ethical and Human Centric
AI Technologies for Smart Production, pp. 1–29. Now Publishers, Norwell (2021). https://
doi.org/10.1561/9781680838770.ch1
19. Khurana, N., Mittal, S., Piplai, A., Joshi, A.: Preventing poisoning attacks on AI based threat
intelligence systems. In: In IEEE 29th International Workshop on Machine Learning for Signal
Processing (MLSP), pp. 1–6 (2019). https://doi.org/10.1109/MLSP.2019.8918803
20. Khorshidpour, Z., Hashemi, S., Hamzeh, A.: Learning a secure classifier against evasion attack.
In: IEEE 16th International Conference on Data Mining Workshops (ICDMW), pp. 295–302
(2016). https://doi.org/10.1109/ICDMW.2016.0049
21. Umbrico, A., Orlandini, A., Cesta, A.: An ontology for human-robot collaboration. Procedia
CIRP. 93, 1097–1102 (2020)
22. Montini, E., Cutrona, V., Bonomi, N., Landolfi, G., Bettoni, A., Rocco, P., Carpanzano, E.:
An IIoT platform for human-aware factory digital twins. Procedia CIRP. 107, 661–667., ISSN
2212-8271 (2022). https://doi.org/10.1016/j.procir.2022.05.042
23. Andersen, P., Goodwin, M., Granmo, O.: Towards safe reinforcement-learning in industrial
grid-warehousing. Inf. Sci. 537, 467–484., ISSN 0020-0255 (2020). https://doi.org/10.1016/
j.ins.2020.06.010
24. Ribeiro, M., Singh, S., Guestrin, C.: “Why should I trust you?”: explaining the predictions
of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference
on Knowledge Discovery and Data Mining (KDD ‘16), pp. 1135–1144. Association for
Computing Machinery, New York (2016). https://doi.org/10.1145/2939672.2939778
25. Soldatos, J., Kyriazis, K. (eds.): Trusted Artificial Intelligence in Manufacturing: A Review of
the Emerging Wave of Ethical and Human Centric AI Technologies for Smart Production. Now
Publishers, Boston-Delft (2021). https://doi.org/10.1561/9781680838770
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Designing a Marketplace to Exchange AI
Models for Industry 5.0
1 Introduction
Online marketplaces, locations on Internet where people can purchase and sale
services and goods, have highly increased in the last couple of decades. Recently,
various marketplaces for exchanging AI models have been introduced [1]. In these
marketplaces the AI models and machine learning (ML) algorithms have been
monetized and offered as products.
AWS Marketplace1 by Amazon enables its customers to find a large variety
of pre-built models and algorithms covering a wide range of use cases and
domains related to Business Analytics, Computer Vision, Healthcare, and Text and
Language Processing. A subscription-based model with charging per hour, day,
etc. is primarily adopted. A monthly based subscription model is also offered by
Akira.AI.2 This marketplace offers AI models especially for solutions related to
Text Analysis and Computer Vision along with access to processing, storage, and
network resources for enabling the execution of AI models. Pretrained models for a
1 https://aws.amazon.com/marketplace/solutions/machine-learning.
2 https://www.akira.ai.
wide variety of sectors are also available at the Gravity AI3 and Modelplace AI4
(specialized in Computer Vision) marketplaces. The latter enables the real-time
execution of models through web–browser interfaces. Other marketplaces5 move
a step further from live execution to shared building of models among developers
by providing software development kits (SDKs).
Some specific AI marketplaces for healthcare domain are also available such as
the Imaging AI Marketplace6 by IBM. It is a centralized marketplace that enables
healthcare providers to discover, purchase, and manage applications that provide the
latest AI-powered tools. In this marketplace, researchers and developers can reach a
large community of customers for their specific AI applications for eHealth domain
and take advantage of the provided infrastructure and deployment processes. To this
direction, Nuance Communications7 has introduced its marketplace for Healthcare
AI solutions as well providing similar functionalities as the IBM one.
In addition to word-leaders’ approaches and market-ready solutions, EC-funded
research projects have presented various marketplaces for listing or even trading
solutions including AI/ML algorithms. AI4EUROPE or AI on Demand [2] offers
a trustworthy open-source platform for the development, training, and sharing of
AI/ML models. However it is considered more as an open code repository as it lacks
business logic in comparison with commercial marketplaces. MARKET4.0 project
[3] develops a multi-sided business platform for plug and produce industrial product
service systems. The European Factory Foundation and EFPF project [4] offers a
Portal/Marketplace as part of its interoperable Data Spine that includes solutions
coming from previous EU projects and third-parties’ initiatives. However the
solutions are a mix of software solutions, products, and services. Other marketplaces
related to manufacturing domain and Industry 4.0 were provided by projects like
v-fos [5] (which offers an application marketplace with an embedded SDK) and
NIMBLE [6] that has introduced a federated interoperable eco-system for B2B
connectivity. Some other approaches [7] coming from research introduced AI as
enablers for marketplaces by combining virtual agents with semantics [8, 9] for
automated negotiations in manufacturing marketplaces [10]. In Boost 4.0 project
a common European Data Space for manufacturing was introduced instead of a
Marketplace. However, it contains AI services connected to available data sources
based on IDSA8 architecture. The latter supports also the establishment of AI
Marketplace9 that is a meeting place for AI providers and consumers. Recently,
3 https://www.gravity-ai.com.
4 https://modelplace.ai/.
5 https://c3.ai/c3-ai-application-platform/c3-ai-integrated-development-studio/c3-ai-
marketplace.
6 https://www.ibm.com/downloads/cas/6BWYDLDO.
7 https://www.nuance.com/healthcare.html.
8 https://internationaldataspaces.org/.
9 https://ki-marktplatz.com/en/.
Designing a Marketplace to Exchange AI Models for Industry 5.0 29
10 https://pop-machina.eu.
30 A. Nizamis et al.
to business logic and transactions, and the front-end part related to user-centric
services such as interfaces.
In particular, the User-Centric Services module provides a series of function-
alities related to user experience such as UIs, search functionalities, user profile
management, etc.
NFT-Based Monetization Framework for AI Models based on Smart Contracts
provides all the functionalities related to business logic based on blockchain. It
offers smart contracts (or chaincode) for fungible and non-fungible Tokens (FTs
and NFTs), marketplace, and mint notary. Furthermore, secure access services are
also part of this module.
AI Model Repository is responsible for the modeling of AI/ML algorithms, their
storage, and the provision of management services such as CRUD (Create, Read,
Update, and Delete) operations.
Various APIs have also been developed to enable the different module communi-
cation based on HTTP protocol.
All the core modules of the introduced marketplace are presented in the following
section.
In this section, the core modules of the knowlEdge Marketplace are described.
We start with the way, the data is stored in the knowlEdge AI Model Repository
Designing a Marketplace to Exchange AI Models for Industry 5.0 31
(Sect. 3.1), before we discuss the Monetization Framework (Sect. 3.2) and conclude
it with the User Interfaces (Sect. 3.3).
3.1.2 Ontology
The knowlEdge Ontology has been developed to ensure that a wide variety of
metadata are available for the knowlEdge repository. The Ontology consists of 12
different types of entities (see Fig. 3). In this entity–relationship diagram, the most
11 https://onnx.ai.
Designing a Marketplace to Exchange AI Models for Industry 5.0 33
important entities and relationships of the knowlEdge Ontology are given and show
the possibilities of the hierarchical structure of this Ontology. It shows the technical
ways the knowlEdge Repository stores the data according to it. These entities can be
split into user-related (User), model-related (Model, IO Vector, Model Specification
and Model Type), data-related (Task, Analysis Type, Application, Data, Property
and Property Type), and performance-related (Performance Evaluation) Entities.
The main entities/classes are as follows:
• User: A User is identified by its unique name and email address. Furthermore,
the timestamp of the time it was created is stored. A User can be owner of several
Models and creator of several Application, Data, and Property Type entities.
• Model: A Model contains its name, a description, and the timestamp of its
creation as well as a link to a Model Specification and the Model Type it
instantiates. It can have hierarchical children and a parent as well as multiple
IO Vectors as input and output.
• IO Vector: An IO Vector contains its name, dimensionality, and data type, which
can be of type integer, float, Boolean, or categorical. It can be input and output to
several Models.
• Model Specification: A Model Specification contains the actual model file, i.e.,
an ONNX or PMML file describing the model.
• Model Type: A Model Type has its name. It can be child to another Model Type
and parent to several Model Types. It can be instantiated by several Models.
• Task: A Task consists of its name, the timestamp of its creation, its Analysis
Type, Application, and Data. It is created by a single User, can have multiple
Tasks as children, and can have several Models trained for it. It may be child of
another Task.
• Analysis Type: An Analysis Type has a name and the timestamp of its creation.
It is part of a Task and can be child to one and parent to many Analysis Types.
• Application: An Application has a name, a description, and the timestamp of its
creation. It is part of a Task, created by a User, and can be child to one and parent
to many Applications.
• Data: A Data entity consists of its name and description, the Task it is part of, and
the timespan it was gathered during. It is created by a User, consists of several
Properties, and may inherit from one and be parent for several Data entities.
• Property: A Property consist of its name, the Property Type it instantiates, and
the Data entity it belongs to.
• Property Type: A Property Type consists of its name, creation time, and type,
which may be Boolean, integer, float, or categorical. It is created by a User and
may be instantiated by several Properties. A Property Type can be based on one
Property Type, and there can be multiple Property Types based on one Property
Type.
• Performance Evaluation: A Performance Evaluation represents the actual
performance of an AI Model on a Task. It is linked to both these entities and
contains the performance measurement as well as the information which measure
this represents.
34 A. Nizamis et al.
The model specifications in the Model Database will be stored in the ONNX or
PMML Format. These two formats offer compatibility with a wide range of different
machine learning techniques and frameworks to boost interoperability regarding AI
models in the proposed marketplace. While PMML focuses on traditional machine
learning methods, ONNX is specialized for exchanging deep neural networks. The
combination of both formats makes it possible to store a wide range of machine
learning models in an easy to use and deploy way.
Besides the Ontology for describing the Marketplace metadata and the reposi-
tory services for the management of AI models, a set of services regarding the
monetization-, security-, and business-related concepts was required for the delivery
of the proposed AI Model Marketplace.
Therefore, in the context of the knowlEdge Marketplace, a number of blockchain-
based services have been developed and deployed. An end-to-end decentralized
AI Model Marketplace that enables the monetization of AI models while ensuring
security, auditability, and verifiability has been implemented based on blockchain
technology. To guarantee ownership of AI models, each model is treated as a unique
asset on the distributed ledger, represented as non-fungible tokens (NFTs). The
use of NFTs provides additional functionalities, including ownership transfer. The
marketplace is based on the presupposition that participants share a common value
system, and fungible tokens are used as the equivalent of real-world fiat currency.
In Table 1, we provide the various actor roles and the means under which they
engage with the marketplace platform, i.e., their capabilities.
Note that the terms “AI Model Producer,” “AI Model Researcher,” and “AI
Model Developer” are used interchangeably to refer to the same actor. Similarly,
the terms “AI Model Consumer” and “Marketplace Customer” refer to the same
actor. Lastly, note that the same real-world entity can potentially enact in all of the
aforementioned roles, e.g., an AI model producer can also act as a consumer, or
marketplace customer for AI models produced by others.
The following functionalities are available in the DLT-based AI Model Market-
place, building on the core set of functionalities that were previously outlined. The
functionalities are as follows:
• AI model producers, represented as NFT owners, can advertise their willingness
to sell access to individual AI model binary files at a price of their choice.
• Each AI model producer can query the AI models they have advertised on the
marketplace.
• Each AI model producer can retract advertised AI models at any time.
• Any entity can query all AI models advertised on the marketplace.
• Interested AI model consumers can purchase access to any advertised AI model,
provided they have sufficient coin balance.
Designing a Marketplace to Exchange AI Models for Industry 5.0 35
• AI model consumers retain access rights to purchased AI model binary files, even
if the models are later retracted from the marketplace.
• Each consumer can query the marketplace for a list of all successful purchases.
• External entities, such as an AI Model repository, can securely verify that an
actor requesting access to an AI model binary file is a legitimate consumer who
has previously performed a successful purchase.
36 A. Nizamis et al.
A detailed diagram of the architecture can be found in Fig. 4. The diagram depicts
the involved users, the components of the blockchain infrastructure, and intuitive
descriptions of their interactions.
There are also several off-chain components that play different roles. For
example, we need to account for the physical storage of AI model files and provide
an identity and access management infrastructure for the actors and the services
they consume. Additionally, we need to include various integration and deployment-
related components such as API gateways and dashboards (user interfaces) for the
involved actors. The list of components along with a succinct description of their
function can be found below:
• Hyperledger Fabric (henceforth, HLF) Community Management (CM) API: A
federated identity authorization service that, apart from supporting all standard
OAuth 2.0 and OpenID Connect flows, encompasses, as part of the user
registration (onboarding) process, private key and X.509 digital certificate
generation, which are subsequently stored in an HLF-compliant secure wallet
store.
• HLF Wallet: Implementation of HLF’s Wallet interface employing MongoDB
as a storage medium for Hyperledger Fabric’s Go SDK. This is employed
internally by the HLF SC Gateway component (see below).
• HLF Smart Contract (SC) Gateway: a configurable microservice that exposes
functions of any smart contract deployed on any arbitrary Hyperledger Fabric
channel as HTTP endpoints
Designing a Marketplace to Exchange AI Models for Industry 5.0 37
• NFT Metadata Store: a REST API that exposes endpoints for storing and
retrieving metadata files associated with NFTs
• NFT Chaincode: the smart contract that implements the entire non-fungible
token-related functionality
• FT Chaincode: the smart contract that implements the entire fungible token-
related functionality
Besides the core back-end services and the corresponding modules that were
presented in the previous section, a module focused on the delivery of front-end-
related services is also included in knowlEdge Marketplace. This module does not
only include interfaces but also supports some user-related functionalities such as
search capabilities and user’s profile management. They are considered in the same
building block as they are strictly connected to a front-end theme that was used and
their functionality is derived from them.
Regarding the interfaces design, best practices were used. The design pillars that
were followed were the esthetic and minimalistic design, the use of common and
consistent UI elements, the adoption of widely used input controls and navigational
elements, the error prevention and good error messaging, etc. The UIs were imple-
mented as web-based interfaces using technologies such as Angular, Bootstrap, and
Nebular. The ngx-admin template was used to enable faster implementation as it
is a popular admin dashboard based on Angular and it is free and open source. It
is efficient as it is packed with a huge number of UI components and it is highly
customizable.
The UIs enable the exploration of various available AI models in different views
(see Fig. 5) based on the user’s preferences (grid view and list view). The search
functionalities provide various filters such as AI models owner, category of the
algorithm, price range, rating, etc. Text-based search is also supported, so the user
can type text related to the model’s name, keywords, and other metadata.
By selecting a model, the user is able to read details (see Fig. 6) such as
description, specifications of the model, and metadata such as rating, price, and
owner. Furthermore, any datasets connected to a model are also visible. All the
data available to UI are dynamically retrieved from Repository and Monetization
modules. The user can also select to add to cart a model in order to purchase it
based on the NFT monetization module.
Besides exploring and purchasing AI models, a user can act as a provider
and deploy his/her own AI model by using corresponding interfaces (Fig. 7) that
are available in a kind of a step wizard form. First, the user adds the dataset
details that were used for training a model. Then the general details regarding the
task/application that the model is related to (e.g., predictive maintenance) are added.
After that, the user adds AI model details such as the type, input and output, model
38 A. Nizamis et al.
format, and connections with other models and deploys the model itself (e.g., an
ONNX file, etc.).
4 Conclusions
and their metadata based on standards has been defined as a necessity. Moreover,
AI developers should be able to provide their models based on widely used
formats and standards in such marketplaces. Furthermore, services to enable trusted
transactions and sharing, along with security and protection of ownership in AI
marketplaces, have been found as core concepts that should be covered as well.
The use of blockchain technology for this kind of services has been proved as an
ideal candidate as it provides all the necessary concepts regarding monetization and
secure and trusted transactions. Moreover, user-friendly and easy-to-use interfaces
were another important factor that should be considered as in the end, as any other
marketplace, it is focused to end-users.
Regarding the next steps, the knowlEdge Marketplace focuses on further testing
and evaluation by domain experts targeting to final improvements so to be con-
sidered as a completely market-ready solution. The plan is for this marketplace to
be one of the core AI marketplaces in Europe related to AI models exchanged in
Industry 5.0 era.
Acknowledgments This work has received funding from the European Union’s Horizon 2020
research and innovation program under Grant Agreement No. 957331—KNOWLEDGE.
References
1. Kumar, A., Finley, B., Braud, T., Tarkoma, S., Hui, P.: Sketching an ai marketplace: Tech,
economic, and regulatory aspects. IEEE Access 9, 13761–13774 (2021)
2. Cortés, U., Cortés, A., Barrué, C.: Trustworthy AI. The AI4EU approach. In: Proceedings of
Science (2019)
40 A. Nizamis et al.
3. Mourtzis, D., Angelopoulos, J., Panopoulos, N.: A survey of digital B2B platforms and
marketplaces for purchasing industrial product service systems: a conceptual framework.
Procedia CIRP 97, 331–336 (2021)
4. Deshmukh, R.A., Jayakody, D., Schneider, A., Damjanovic-Behrendt, V.: Data spine: a
federated interoperability enabler for heterogeneous IoT platform ecosystems. Sensors 21(12),
4010 (2021)
5. Anaya, V., Moalla, N., Stellingwerff, L., Flores, J.L., Fraile, F.: vf-OS IO Toolkit. Enterprise
Interoperability: Smart Services and Business Impact of Enterprise Interoperability, pp. 91–97
(2018)
6. Gönül, S., Çavdaroğlu, D., Kabak, Y., Glachs, D., Gigante, F., Deng, Q.: A B2B marketplace
eCommerce platform approach integrating purchasing and transport processes. In: Interna-
tional Conference on Interoperability for Enterprise Systems and Applications, pp. 105–121.
Springer, Cham (2023)
7. Bonino, D., Vergori, P.: Agent marketplaces and deep learning in enterprises: the composition
project. In: 2017 IEEE 41st Annual Computer Software and Applications Conference (COMP-
SAC), vol. 1, pp. 749–754. IEEE, New York (2017)
8. Nizamis, A.G., Ioannidis, D.K., Kaklanis, N.T., Tzovaras, D.K.: A semantic framework for
agent-based collaborative manufacturing eco-systems. IFAC-PapersOnLine 51(11), 382–387
(2018)
9. Nizamis, A., Vergori, P., Ioannidis, D., Tzovaras, D.: Semantic framework and deep learning
toolkit collaboration for the enhancement of the decision making in agent-based marketplaces.
In: 2018 5th International Conference on Mathematics and Computers in Sciences and Industry
(MCSI), pp. 135–140. IEEE, New York (2018)
10. Mertens, C., Alonso, J., Lázaro, O., Palansuriya, C., Böge, G., . . . , Poulakidas, A.: A
framework for big data sovereignty: the European industrial data space (EIDS). In: Data
Spaces: Design, Deployment and Future Directions. pp. 201–226. Springer International
Publishing, Cham (2022)
11. Sarpatwar, K., Sitaramagiridharganesh Ganapavarapu, V., Shanmugam, K., Rahman, A.,
Vaculin, R.: Blockchain enabled AI marketplace: the price you pay for trust. In: Proceedings
of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)
12. Alvarez-Napagao, S., Ashmore, B., Barroso, M., Barrué, C., Beecks, C., Berns, F., Bosi, I.,
Chala, S.A., Ciulli, N., Garcia-Gasulla, M., Grass, A., Ioannidis, D., Jakubiak, N., Köpke, K.,
Lämsä, V., Megias, P., Nizamis, A., Pastrone, C., Rossini, R., Sànchez-Marrè, M., Ziliotti,
L.: knowlEdge Project–Concept, methodology and innovations for artificial intelligence in
industry 4.0. In: 2021 IEEE 19th International Conference on Industrial Informatics (INDIN),
pp. 1–7. IEEE, New York (2021)
13. Wajid, U., Nizamis, A., Anaya, V.: Towards Industry 5.0–A Trustworthy AI Framework for
Digital Manufacturing with Humans in Control. Proceedings https://ceur-ws. org ISSN, 1613,
0073 (2022)
14. Tzavaras, A., Mainas, N., Petrakis, E.G.M.: OpenAPI framework for the Web of Things.
Internet of Things 21, 100675 (2023)
15. Banker, K., Garrett, D., Bakkum, P., Verch, S.: MongoDB in action: covers MongoDB version
3.0. Simon and Schuster, New York (2016)
16. White, T.: Hadoop: the definitive guide. O’Reilly Media Inc., California (2012)
17. Guazzelli, A., Zeller M., Lin, W., Williams, G., et al.: PMML: an open standard for sharing
models. R J. 1(1), 60 (2009)
Designing a Marketplace to Exchange AI Models for Industry 5.0 41
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Human-AI Interaction for Semantic
Knowledge Enrichment of AI Model
Output
1 Introduction
2 Related Works
The human-AI collaboration system offers features for data and model selection.
Operators select models and data from the available list of options in order to
execute them for a specific scenario. Model and data selection are critical factors
in human-AI collaboration. Because the choice of model and data significantly
influences the performance, accuracy, and overall effectiveness of the AI system.
When considering human-AI collaboration, several key considerations come into
play. One aspect is determining the specific requirements of the task at hand,
understanding the problem domain, the type of input data, and the desired output.
This knowledge helps guide the selection of an appropriate model for the dataset.
Another aspect is understanding capabilities of the AI model because different AI
models and algorithms are suitable for the task. Considering factors such as the
model’s architecture, complexity, interpretability, and scalability affect the choice
of a model that aligns with the task requirements.
models to optimize their performance. Feedback of domain expert plays a vital role
in configuration adaptation as it provides valuable insights into the effectiveness and
suitability of the AI model’s behavior in that the AI system can learn and adjust its
configuration settings to improve its performance and align more closely with the
user’s requirements in response to incorporating domain expert feedback. Moreover,
when a model offers a need for specific configurations of machines that need to
be modified, operators/managers can adapt the configurations of machines so that
it suits to the model under consideration, for example, if new machines need to be
added to the human-AI collaboration system, their configuration should be extracted
and stored in such a way that they are accessible and usable to the modules.
model learns and accumulates more data that it can gradually require less feedback
from the domain expert.
Organizations can benefit from this system despite these challenges by defining
feasible objectives of the human-AI collaboration in the manufacturing setting
whereby they identify the specific areas where AI can enhance manufacturing pro-
cesses, such as quality control, predictive maintenance, or supply chain optimization
and establish key performance indicators (KPIs) to measure success. For example,
companies can utilize this approach to perform what-if analysis in order to explore
the potential implications of different scenarios and make more informed decisions
by combining the analytical capabilities of AI models with human judgment,
expertise, and contextual understanding. Domain experts can modify the input
parameters, adjust variables, or introduce new constraints to observe the potential
changes in the outcomes. The AI system then performs the simulations and presents
the results to the human collaborator.
An example scenario that shows the procedure for the human-AI interaction is
shown below:
• Fetch predicted labels form the output of automatic label detection of models.
• Present the data with predicted label to the domain expert.
• Present the domain expert with choices to (i) accept the predicted label and
(a) confirm the predicted label or (b) offer an alternative label or (ii) reject the
predicted label and offer the correct label.
– If the domain expert accepts and confirms the label, the process ends.
– If the domain expert accepts the predicted label and offers an alternative
label or a refinement of the label or rejects the predicted label altogether
and offers the correct label, the domain expert’s input will be sent as input
to retrain the model.
• Visualization of behavior of the model with/without the domain expert’s input
will be shown for comparison of the effect of the domain fusion.
• Human-AI collaboration system will expose an API of the visualization to the
DSS component through which the user will inspect the outputs of the model.
Figure 3 shows the process of data/model selection and parameter optimization
including data flow and UI mockup for model selection and parameter optimization
user interface through which the domain expert selects the model and parameter and
optimizes the parameter values. The UI presents visualization of processing results
for the selected model, parameter, and values. Once the domain expert determines
the model, parameter, and values, the UI then enables the domain expert to export
the result which will then be consumed by the Decision Support System (DSS).
The domain expert selects a section of the visualization and provides engineered
knowledge, i.e., manual labeling of data points. This helps the user to visually
inspect the dataset and enrich it with domain knowledge to boost the quality of
the data to be used as training dataset for better performance of the ML model. For
example, for an AI model built for anomaly detection, this is achieved by enabling
the user to select the data point on the visualization plot in order to display and
review (and when applicable, modify) the data that are marked by the system as
anomalies. This is implemented by providing point, box, or lasso [15] selection
where the user can select a single (or multiple data points on the graphs) and get the
corresponding data points back, to provide the domain knowledge.
As depicted in Figs. 2 and 3, the domain expert will load data and models from
the model repository, run the models on the data, observe the visualization, and
adjust parameters in order to achieve the desired behavior of the AI model. Once the
domain expert obtains satisfactory output from the model, she/he can then provide
feedback. The algorithm shown in Fig. 4 shows the detailed operations during the
domain knowledge enrichment.
Human-AI Interaction for Semantic Knowledge Enrichment of AI Model Output 51
Acknowledgments The research leading to these results has received funding from the Horizon
2020 Program of the European Commission under Grant Agreement No. 957331.
References
1. Arinez, J.F., Chang, Q., Gao, R.X., Xu, C., Zhang, J.: Artificial intelligence in advanced
manufacturing: current status and future outlook. J. Manuf. Sci. Eng. 142(11), 110804 (2020)
2. Aron, J.: How Innovative is Apple’s New Voice Assistant, Siri? (2011)
3. Baghernezhad-Tabasi, S., Druette, L., Jouanot, F., Meurger, C., Rousset, M.-C.: IOPE:
interactive ontology population and enrichment. In: Workshop on Ontology-Driven Conceptual
Modelling of Digital Twins co-located with Semantics 2021 (2021)
4. Fan, M., Yang, X., Yu, T., Liao, Q.V., Zhao, J.: Human-ai collaboration for UX evaluation:
effects of explanation and synchronization. Proc. ACM Hum.-Comput. Interact. 6(CSCW1),
1–32 (2022)
5. Fernández-Izquierdo, A., Cimmino, A., Patsonakis, C., Tsolakis, A.C., García-Castro, R.,
Ioannidis, D., Tzovaras, D.: OpenADR ontology: semantic enrichment of demand response
strategies in smart grids. In: International Conference on Smart Energy Systems and Technolo-
gies (SEST), pp. 1–6. IEEE, New York (2020)
6. Gunasekaran, A., Yusuf, Y.Y., Adeleye, E.O., Papadopoulos, T., Kovvuri, D., Geyi, D.G.: Agile
manufacturing: an evolutionary review of practices. Int. J. Prod. Res. 57(15–16), 5154–5174
(2019)
7. Idoudi, R., Ettabaa, K.S., Solaiman, B., Hamrouni, K.: Ontology knowledge mining for
ontology conceptual enrichment. Knowl. Manag. Res. Pract. 17(2), 151–160 (2019)
8. Khadpe, P., Krishna, R., Fei-Fei, L., Hancock, J.T., Bernstein, M.S.: Conceptual metaphors
impact perceptions of human-ai collaboration. Proc. ACM Hum.-Comput. Interact. 4(CSCW2),
1–26 (2020)
9. Li, T., Vorvoreanu, M., DeBellis, D., Amershi, S.: Assessing human-AI interaction early
through factorial surveys: a study on the guidelines for human-AI interaction. ACM Trans.
Comput.-Hum. Interact. 30(5), 45 (2023). https://doi.org/10.1145/3511605
10. LLC, C.T.: Time Series Data Labeling: A Complete Know-how for Efficient AI Implemen-
tation (2022). https://www.cogitotech.com/blog/time-series-data-labeling-a-complete-know-
how-for-efficient-ai-implementation/. Last accessed: November 16, 2023
11. López, G., Quesada, L., Guerrero, L.A.: Alexa vs. Siri vs. Cortana vs. Google assistant:
a comparison of speech-based natural user interfaces. In: Advances in Human Factors and
Systems Interaction: Proceedings of the AHFE 2017 International Conference on Human
Factors and Systems Interaction, July 17–21, 2017, The Westin Bonaventure Hotel, Los
Angeles, California, USA 8, pp. 241–250. Springer, Berlin (2018)
12. Mucha, H., Robert, S., Breitschwerdt, R., Fellmann, M.: Interfaces for explanations in human-
ai interaction: proposing a design evaluation approach. In: Extended Abstracts of the 2021 CHI
Conference on Human Factors in Computing Systems, CHI EA ’21, New York, NY, USA.
Association for Computing Machinery, New York (2021)
13. Nguyen, A.T., Kharosekar, A., Krishnan, S., Krishnan, S., Tate, E., Wallace, B.C., Lease,
M.: Believe it or not: designing a human-ai partnership for mixed-initiative fact-checking. In:
Proceedings of the 31st Annual ACM Symposium on User Interface Software and Technology,
UIST ’18, pp. 189–199. Association for Computing Machinery, New York (2018)
54 S. A. Chala and A. Graß
14. Stratogiannis, G., Kouris, P., Alexandridis, G., Siolas, G., Stamou, G., Stafylopatis, A.:
Semantic enrichment of documents: a classification perspective for ontology-based imbalanced
semantic descriptions. Knowl. Inf. Syst. 63(11), 3001–3039 (2021)
15. Tsang, M., Enouen, J., Liu, Y.: Interpretable artificial intelligence through the lens of feature
interaction. arXiv preprint arXiv:2103.03103 (2021)
16. Urbanowicz, R.J., Moore, J.H.: Exstracs 2.0: description and evaluation of a scalable learning
classifier system. Evol. Intell. 8, 89–116 (2015)
17. Wang, W., Barnaghi, P.M., Bargiela, A.: Probabilistic topic models for learning terminological
ontologies. IEEE Trans. Knowl. Data Eng. 22(7), 1028–1040 (2009)
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Examining the Adoption of Knowledge
Graphs in the Manufacturing Industry:
A Comprehensive Review
1 Introduction
We propose some research questions to provide specific insights into how KGs
are used in manufacturing. These Research Questions (RQs) consider that the two
most popular KG types are Resource Description Framework (RDFs) and Labeled
Property Graph (LPG). Our RQs are designed to cover the essential aspects of
bibliometric facts and application scenarios.
The purpose of RQ1 is to demonstrate the significance and relevance of the topic
by providing an overview of bibliometric facts from previously published studies on
the applications of KGs in manufacturing.
3.2 Dataset
The presented findings illustrate the prevailing domains explored in the literature
on applying KGs in manufacturing, as summarized in Fig. 2. To determine whether
a given paper pertains to the manufacturing domain, the North American Industry
Classification System (NAICS1 ) was employed. However, most of the examined
literature does not specify any particular application domain.
Machinery is identified as the second most frequently represented domain,
after which Materials, Chemistry, and Automotive follow. Furthermore, Additive
Manufacturing, Aerospace, Mining, Operations, and Textile, albeit less frequently
1 https://www.census.gov/naics/.
60 J. Martinez-Gil et al.
Business
Chemistry
Computer Science
Engineering
Materials Science
0 10 20 30 40 50
Percentage of papers published
Fig. 1 Research communities that have published the most research on KGs in manufacturing
investigated, are also observed in the literature. The identified domains highlight the
diverse industries that benefit from leveraging KGs.
Most reviewed works employ knowledge fusion techniques in general scenarios
where KGs combine data from multiple sources. Additionally, KGs are applied
to automate the merging of isolated production processes, generate digital twins
based on KGs, and utilize them for automated source code development. These
findings demonstrate the versatility of KGs in manufacturing and their potential to
revolutionize various aspects of industrial production.
2 https://www.w3.org/.
Knowledge Graphs in the Manufacturing Industry 61
Textile
Operations
Mining
Materials
Machinery
General
Chemistry
Automotive
Aerospace
Additive
0 10 20 30 40 50
Percentage of papers published
Fig. 2 Manufacturing domains that have carried out the most research work around KGs
RDF
LPG
0 20 40 60 80 100
Percentage of papers published
semantic interoperability. The semantic layer of the available statements, along with
the reasoning applied to it, forms the foundation of intelligent systems in the RDF
domain.
On the other hand, LPG representation primarily emphasizes the graph’s struc-
ture, properties, and relationships. This highlights the unique characteristics of
graph data, opening new opportunities for data analysis and visualization. It also
brings a window of opportunity for developing ML systems that use graphs to infer
additional information.
62 J. Martinez-Gil et al.
Knowledge-Driven
Data-Driven
0 20 40 60 80 100
Percentage of papers published
4 Insights
This section summarizes the results obtained from our analysis and highlights
potential areas for future research on the use KGs in the manufacturing domain.
The findings are structured according to the RQs addressed earlier in the study.
Based on our study, we can deduce the most active research communities in the field
of KGs. The answer to RQ1 (AQ1) is as follows:
AQ1. The majority of primary research in the field of KGs is conducted in the
discipline of Computer Science. Research in KGs is less common in other
areas of knowledge.
This could be because computer scientists have been developing new representa-
tion models since the beginning. Today, KGs are considered the natural progression
of such models to make them more adaptable to new platforms and emerging
methods for managing large amounts of data.
Regarding the answer to RQ2 (AQ2), it is unsurprising that the most common
case is the preference for proposing generic models that can be easily adapted to
various domains.
The domains related to machinery and materials are the next most represented,
followed by chemistry and material. Finally, some KGs have also been developed
in the aerospace, additive manufacturing, mining, operations, and textile fields.
Regarding the representation of KGs, the two most commonly used data models
are RDF and LPG. However, in answering RQ3 (AQ3), we seek to identify the
current prevailing choice for representing KGs.
64 J. Martinez-Gil et al.
AQ3. In the industrial domain, RDF is the preferred format for building.
KGs. This is due to RDF’s ability to represent complex data and relationships
in a structured and interoperable manner, which allows for the building of
integrated knowledge spaces for both humans and AI components.
In light of our study, the utilization of KGs within the manufacturing industry has
experienced substantial growth in recent years as manufacturers seek to enhance
their operational efficiency and decision-making capabilities. The structural design
of KGs facilitates a more intuitive and comprehensive representation of data than
traditional database models, rendering KGs well suited for the manufacturing
industry.
One of the primary reasons for this keen interest is that by modeling relationships
between suppliers, manufacturers, and customers, organizations can better under-
stand the flow of goods, services, and information through their supply chain. This,
in turn, can assist them in identifying bottlenecks, optimizing production processes,
and ensuring product delivery to customers.
Additional Lesson Learned #2. The majority of the studies examined have
been published in conference proceedings. In many instances, this indicates
that the subject of investigation is still in the developmental stages. The state
of the art is gradually maturing in almost every research area, leading to more
journal publications with archival significance.
KGs can aid manufacturers in enhancing their ability to predict and respond to
shifts in demand. This can help reduce waste, optimize production processes, and
boost efficiency. However, most of the research is a work in progress, and there is
still a long way to go to consolidate the results of archival value.
As a result of our study, we have identified several issues that limit the adoption
of KGs in manufacturing and production environments. Some of the most critical
issues are described below. The first issue concerns tabular data. This kind of data
is frequently represented in values separated by commas. It is typically one of the
most common input methods in industrial environments because it enables modeling
a wide variety of data associated with temporal aspects (timestamps) and spatial
aspects (coordinates). However, more optimal solutions still need to be proposed.
Problem 1 (Dealing with Tabular Data) Most solutions today are created
to deal with information that is predominately textual in its presentation.
Although this information category is crucial in the sector, it is not domi-
nant in manufacturing settings, which involve working with machinery and
equipment that generate numerical data in tabular form.
Another fact that is taken for granted by both researchers and practitioners is that
it is possible for KGs to effectively deal with information of varying types that may
arrive via a variety of channels and sources. However, our research has not found a
large number of papers concerned with the temporal component of processing KGs.
66 J. Martinez-Gil et al.
Last but not least, according to the results of our investigation, work still needs
to be done in compiling the best practices for manufacturing KGs. In this sense, we
miss work in the direction of design and proposal of best practices for the sector.
KGs are suitable for the manufacturing industry because they can provide
systems with contextual data to achieve efficient and effective solutions. This
contextual data includes human experience, environmental knowledge, technical
conventions, etc. Creating such solutions becomes critical when the influence on
human life is essential, as in the case of a factory that employs human workers.
5 Conclusion
In this chapter, we have seen how the amount of data generated in the industrial
sector at a high velocity is bringing new challenges. For example, this data emanates
from multiple sources, each utilizing distinct formats and standards. Consequently,
integrating these divergent pieces of information is not only essential but also
critical. Contextualizing data elements utilizing relevant relationships is imperative
to ensure consistency and high-quality data.
The study also examines KGs as multifaceted knowledge bases that capture
interlinked descriptions of entities. KGs facilitate the smooth integration and
structuring of information at large scale, even from heterogeneous sources. Unlike
other knowledge bases, KGs are not homogeneous and do not require rigid schemas.
This makes KGs highly scalable and suitable for integrating and connecting diverse
data representations.
Knowledge Graphs in the Manufacturing Industry 67
Acknowledgments We would like to thank the anonymous reviewers for their constructive
comments to improve this work. Our study is based on a sample of literature on KGs in
manufacturing. The primary sources used in the study are not cited in the text, but are listed in
the References section. SCCH co-authors have been partially funded by the Federal Ministry for
Climate Action, Environment, Energy, Mobility, Innovation, and Technology (BMK), the Federal
Ministry for Digital and Economic Affairs (BMDW), and the State of Upper Austria in the frame
of SCCH, a center in the COMET Program managed by Austrian FFG. All co-authors have also
received funding from Teaming.AI, a project supported by the European Union’s Horizon 2020
program, under Grant Agreement No. 957402.
References
1. Aggour, K.S., Kumar, V.S., Cuddihy, P., Williams, J.W., Gupta, V., Dial, L., Hanlon, T.,
Gambone, J., Vinciquerra, J.: Federated multimodal big data storage & analytics platform for
additive manufacturing. In: 2019 IEEE International Conference on Big Data (Big Data), pp.
1729–1738. IEEE, New York (2019)
2. Alam, M., Fensel, A., Martinez-Gil, J., Moser, B., Recupero, D.R., Sack, H.: Special issue
on machine learning and knowledge graphs. Future Gener. Comput. Syst. 129, 50–53 (2022).
https://doi.org/10.1016/j.future.2021.11.022
3. Bachhofner, S., Kiesling, E., Kurniawan, K., Sallinger, E., Waibel, P.: Knowledge graph
modularization for cyber-physical production systems. In: Seneviratne, O., Pesquita, C.,
Sequeda, J., Etcheverry, L. (eds.) Proceedings of the ISWC 2021 Posters, Demos and Industry
Tracks: From Novel Ideas to Industrial Practice co-located with 20th International Semantic
Web Conference (ISWC 2021), Virtual Conference, October 24–28, 2021. CEUR Workshop
Proceedings, vol. 2980. CEUR-WS.org (2021). https://ceur-ws.org/Vol-2980/paper333.pdf
4. Bachhofner, S., Kiesling, E., Revoredo, K., Waibel, P., Polleres, A.: Automated process
knowledge graph construction from BPMN models. In: Strauss, C., Cuzzocrea, A., Kotsis, G.,
Tjoa, A.M., Khalil, I. (eds.) Database and Expert Systems Applications—33rd International
Conference, DEXA 2022, Vienna, Austria, August 22–24, 2022, Proceedings, Part I. Lecture
Notes in Computer Science, vol. 13426, pp. 32–47. Springer, Berlin (2022). https://doi.org/10.
1007/978-3-031-12423-5_3
5. Bachhofner, S., Kurniawan, K., Kiesling, E., Revoredo, K., Bayomie, D.: Knowledge graph
supported machine parameterization for the injection moulding industry. In: Villazón-Terrazas,
B., Ortiz-Rodríguez, F., Tiwari, S., Sicilia, M., Martín-Moncunill, D. (eds.) Knowledge Graphs
and Semantic Web—4th Iberoamerican Conference and Third Indo-American Conference,
KGSWC 2022, Madrid, Spain, November 21–23, 2022, Proceedings. Communications in
Computer and Information Science, vol. 1686, pp. 106–120. Springer, Berlin (2022). https://
doi.org/10.1007/978-3-031-21422-6_8
68 J. Martinez-Gil et al.
6. Bader, S.R., Grangel-Gonzalez, I., Nanjappa, P., Vidal, M.E., Maleshkova, M.: A knowledge
graph for industry 4.0. In: The Semantic Web: 17th International Conference, ESWC 2020,
Heraklion, Crete, Greece, May 31–June 4, 2020, Proceedings 17, pp. 465–480. Springer, Berlin
(2020)
7. Banerjee, A., Dalal, R., Mittal, S., Joshi, K.P.: Generating digital twin models using knowledge
graphs for industrial production lines. In: Proceedings of the 2017 ACM on Web Science
Conference, pp. 425–430 (2017)
8. Buchgeher, G., Gabauer, D., Martinez-Gil, J., Ehrlinger, L.: Knowledge graphs in manufac-
turing and production: a systematic literature review. IEEE Access 9, 55537–55554 (2021).
https://doi.org/10.1109/ACCESS.2021.3070395
9. Chhetri, T.R., Aghaei, S., Fensel, A., Göhner, U., Gül-Ficici, S., Martinez-Gil, J.: Optimising
manufacturing process with Bayesian structure learning and knowledge graphs. In: Computer
Aided Systems Theory—EUROCAST 2022—18th International Conference, Las Palmas de
Gran Canaria, Spain, February 20–25, 2022, Revised Selected Papers. Lecture Notes in
Computer Science, vol. 13789, pp. 594–602. Springer, Berlin (2022). https://doi.org/10.1007/
978-3-031-25312-6_70
10. Dombrowski, U., Reiswich, A., Imdahl, C.: Knowledge graphs for an automated information
provision in the factory planning. In: 2019 IEEE International Conference on Industrial
Engineering and Engineering Management (IEEM). pp. 1074–1078. IEEE, New York (2019)
11. Duan, W., Chiang, Y.Y.: Building knowledge graph from public data for predictive analysis: a
case study on predicting technology future in space and time. In: Proceedings of the 5th ACM
SIGSPATIAL International Workshop on Analytics for Big Geospatial Data, pp. 7–13 (2016)
12. Eibeck, A., Lim, M.Q., Kraft, M.: J-park simulator: an ontology-based platform for cross-
domain scenarios in process industry. Comput. Chem. Eng. 131, 106586 (2019)
13. Freudenthaler, B., Martinez-Gil, J., Fensel, A., Höfig, K., Huber, S., Jacob, D.: Ki-net: Ai-based
optimization in industrial manufacturing—A project overview. In: Computer Aided Systems
Theory—EUROCAST 2022—18th International Conference, Las Palmas de Gran Canaria,
Spain, February 20–25, 2022, Revised Selected Papers. Lecture Notes in Computer Science,
vol. 13789, pp. 554–561. Springer, Berlin (2022). https://doi.org/10.1007/978-3-031-25312-
6_65
14. Garofalo, M., Pellegrino, M.A., Altabba, A., Cochez, M.: Leveraging knowledge graph
embedding techniques for industry 4.0 use cases. In: Cyber Defence in Industry 4.0 Systems
and Related Logistics and IT Infrastructures, pp. 10–26. IOS Press, New York (2018)
15. Grangel-González, I., Halilaj, L., Vidal, M.E., Lohmann, S., Auer, S., Müller, A.W.: Seamless
integration of cyber-physical systems in knowledge graphs. In: Proceedings of the 33rd Annual
ACM Symposium on Applied Computing, pp. 2000–2003 (2018)
16. Grangel-González, I., Halilaj, L., Vidal, M.E., Rana, O., Lohmann, S., Auer, S., Müller,
A.W.: Knowledge graphs for semantically integrating cyber-physical systems. In: International
Conference on Database and Expert Systems Applications, pp. 184–199. Springer, Berlin
(2018)
17. Haase, P., Herzig, D.M., Kozlov, A., Nikolov, A., Trame, J.: metaphactory: a platform for
knowledge graph management. Semantic Web 10(6), 1109–1125 (2019)
18. He, L., Jiang, P.: Manufacturing knowledge graph: a connectivism to answer production
problems query with knowledge reuse. IEEE Access 7, 101231–101244 (2019)
19. Hermann, M., Pentek, T., Otto, B.: Design principles for Industrie 4.0 scenarios. In: 2016 49th
Hawaii International Conference on System Sciences (HICSS), pp. 3928–3937. IEEE, New
York (2016)
20. Hoch, T., Heinzl, B., Czech, G., Khan, M., Waibel, P., Bachhofner, S., Kiesling, E., Moser,
B.: Teaming.ai: enabling human-ai teaming intelligence in manufacturing. In: Zelm, M., Boza,
A., León, R.D., Rodríguez-Rodríguez, R. (eds.) Proceedings of Interoperability for Enterprise
Systems and Applications Workshops co-located with 11th International Conference on
Interoperability for Enterprise Systems and Applications (I-ESA 2022), Valencia, Spain, March
23–25, 2022. CEUR Workshop Proceedings, vol. 3214. CEUR-WS.org (2022). https://ceur-ws.
org/Vol-3214/WS5Paper6.pdf
Knowledge Graphs in the Manufacturing Industry 69
21. Kalaycı, E.G., Grangel González, I., Lösch, F., Xiao, G., ul Mehdi, A., Kharlamov, E.,
Calvanese, D.: Semantic integration of Bosch manufacturing data using virtual knowledge
graphs. In: The Semantic Web–ISWC 2020: 19th International Semantic Web Conference,
Athens, Greece, November 2–6, 2020, Proceedings, Part II 19, pp. 464–481. Springer, Berlin
(2020)
22. Kattepur, A.: Roboplanner: autonomous robotic action planning via knowledge graph queries.
In: Proceedings of the 34th ACM/SIGAPP Symposium on Applied Computing, pp. 953–956
(2019)
23. Ko, H., Witherell, P., Lu, Y., Kim, S., Rosen, D.W.: Machine learning and knowledge graph
based design rule construction for additive manufacturing. Addit. Manuf. 37, 101620 (2021)
24. Kumar, A., Bharadwaj, A.G., Starly, B., Lynch, C.: FabKG: a knowledge graph of manufac-
turing science domain utilizing structured and unconventional unstructured knowledge source.
arXiv preprint arXiv:2206.10318 (2022)
25. Leijie, F., Yv, B., Zhenyuan, Z.: Constructing a vertical knowledge graph for non-traditional
machining industry. In: 2018 IEEE 15th International Conference on Networking, Sensing and
Control (ICNSC), pp. 1–5. IEEE, New York (2018)
26. Li, R., Dai, W., He, S., Chen, X., Yang, G.: A knowledge graph framework for software-
defined industrial cyber-physical systems. In: IECON 2019-45th Annual Conference of the
IEEE Industrial Electronics Society, vol. 1, pp. 2877–2882. IEEE, New York (2019)
27. Li, X., Chen, C.H., Zheng, P., Wang, Z., Jiang, Z., Jiang, Z.: A knowledge graph-aided concept–
knowledge approach for evolutionary smart product–service system development. J. Mech.
Des. 142(10), 101403 (2020)
28. Li, X., Zhang, S., Huang, R., Huang, B., Xu, C., Kuang, B.: Structured modeling of
heterogeneous cam model based on process knowledge graph. Int. J. Adv. Manuf. Technol.
96(9–12), 4173–4193 (2018)
29. Liebig, T., Maisenbacher, A., Opitz, M., Seyler, J.R., Sudra, G., Wissmann, J.: Building a
Knowledge Graph for Products and Solutions in the Automation Industry (2019)
30. Liu, M., Li, X., Li, J., Liu, Y., Zhou, B., Bao, J.: A knowledge graph-based data representation
approach for IIoT-enabled cognitive manufacturing. Adv. Eng. Inform. 51, 101515 (2022)
31. Martinez-Gil, J., Buchgeher, G., Gabauer, D., Freudenthaler, B., Filipiak, D., Fensel, A.:
Root cause analysis in the industrial domain using knowledge graphs: a case study on
power transformers. In: Longo, F., Affenzeller, M., Padovano, A. (eds.) Proceedings of
the 3rd International Conference on Industry 4.0 and Smart Manufacturing (ISM 2022),
Virtual Event/Upper Austria University of Applied Sciences—Hagenberg Campus—Linz,
Austria, 17–19 November 2021. Procedia Computer Science, vol. 200, pp. 944–953. Elsevier,
Amsterdam (2021). https://doi.org/10.1016/j.procs.2022.01.292
32. Meckler, S., Steinmüller, H., Harth, A.: Building a knowledge graph with inference for a
production machine using the web of things standard. In: Advances and Trends in Artificial
Intelligence. From Theory to Practice: 34th International Conference on Industrial, Engineer-
ing and Other Applications of Applied Intelligent Systems, IEA/AIE 2021, Kuala Lumpur,
Malaysia, July 26–29, 2021, Proceedings, Part II 34, pp. 240–251. Springer, Berlin (2021)
33. Nayak, A., Kesri, V., Dubey, R.K.: Knowledge graph based automated generation of test cases
in software engineering. In: Proceedings of the 7th ACM IKDD CoDS and 25th COMAD, pp.
289–295 (2020)
34. Noy, N.F., Gao, Y., Jain, A., Narayanan, A., Patterson, A., Taylor, J.: Industry-scale knowledge
graphs: lessons and challenges. Commun. ACM 62(8), 36–43 (2019). https://doi.org/10.1145/
3331166
35. Peroni, S., Vitali, F.: Interfacing fast-fashion design industries with semantic web technologies:
the case of imperial fashion. J. Web Semant. 44, 37–53 (2017)
36. Ringsquandl, M., Kharlamov, E., Stepanova, D., Lamparter, S., Lepratti, R., Horrocks, I.,
Kröger, P.: On event-driven knowledge graph completion in digital factories. In: 2017 IEEE
International Conference on Big Data (Big Data), pp. 1676–1681. IEEE, New York (2017)
70 J. Martinez-Gil et al.
37. Ringsquandl, M., Lamparter, S., Lepratti, R., Kröger, P.: Knowledge fusion of manufacturing
operations data using representation learning. In: IFIP International Conference on Advances
in Production Management Systems, pp. 302–310. Springer, Berlin (2017)
38. Rožanec, J.M., Zajec, P., Kenda, K., Novalija, I., Fortuna, B., Mladenić, D.: XAI-KG:
knowledge graph to support XAI and decision-making in manufacturing. In: Proceedings
of the Advanced Information Systems Engineering Workshops: CAiSE 2021 International
Workshops, Melbourne, VIC, Australia, June 28–July 2, 2021, pp. 167–172. Springer, Berlin
(2021)
39. Tushkanova, O., Samoylov, V.: Knowledge net: model and system for accumulation, represen-
tation, and use of knowledge. IFAC-PapersOnLine 52(13), 1150–1155 (2019)
40. Wang, Z., Zhang, B., Gao, D.: A novel knowledge graph development for industry design: a
case study on indirect coal liquefaction process. Comput. Ind. 139, 103647 (2022)
41. Yan, H., Yang, J., Wan, J.: KnowIME: a system to construct a knowledge graph for intelligent
manufacturing equipment. IEEE Access 8, 41805–41813 (2020)
42. Zhang, X., Liu, X., Li, X., Pan, D.: MMKG: an approach to generate metallic materials
knowledge graph based on DBpedia and Wikipedia. Comput. Phys. Commun. 211, 98–112
(2017)
43. Zhao, M., Wang, H., Guo, J., Liu, D., Xie, C., Liu, Q., Cheng, Z.: Construction of an industrial
knowledge graph for unstructured Chinese text learning. Appl. Sci. 9(13), 2720 (2019)
44. Zhao, Y., Liu, Q., Xu, W.: Open industrial knowledge graph development for intelligent man-
ufacturing service matchmaking. In: 2017 International Conference on Industrial Informatics-
Computing Technology, Intelligent Technology, Industrial Information Integration (ICIICII),
pp. 194–198. IEEE, New York (2017)
45. Zhou, B., Bao, J., Li, J., Lu, Y., Liu, T., Zhang, Q.: A novel knowledge graph-based
optimization approach for resource allocation in discrete manufacturing workshops. Robot.
Comput. Integr. Manuf. 71, 102160 (2021)
46. Zhou, X., Lim, M.Q., Kraft, M.: A Smart Contract-Based Agent Marketplace for the j-park
Simulator–a Knowledge Graph for the Process Industry (2020)
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Leveraging Semantic Representations via
Knowledge Graph Embeddings
1 Introduction
F. Krause ()
University of Mannheim, Data and Web Science Group, Mannheim, Germany
e-mail: [email protected]
K. Kurniawan
WU, Institute for Data, Process and Knowledge Management, Vienna, Austria
Austrian Center for Digital Production (CDP), Vienna, Austria
e-mail: [email protected]
E. Kiesling
WU, Institute for Data, Process and Knowledge Management, Vienna, Austria
e-mail: [email protected]
J. Martinez-Gil · T. Hoch · M. Pichler · B. Heinzl · B. Moser
Software Competence Center Hagenberg GmbH, Hagenberg, Austria
e-mail: [email protected]; [email protected]; [email protected];
[email protected]; [email protected]
of valid relation types defined in the underlying ontology. Thus, an edge in the form
of a triple .(s, p, o) ∈ E implies an outgoing relation from the subject .s ∈ V to the
object .o ∈ V via the predicate .p ∈ R. Given such a KG, embedding techniques aim
to exploit the topology of the graph to generate latent feature representations
γ :V →𝚪
. (1)
3 Representation Learning
desired task.
Representation learning can be performed in a supervised, unsupervised, or
self-supervised manner. One example of a supervised approach for learning latent
feature representations is the training of deep neural networks on labeled input data.
Namely, given an input feature .π(v) for some .v ∈ V , the hidden layer outputs
(and also the output layer) obtained from the forward pass of the network can be
considered as alternative representations .γ (v), as illustrated in Fig. 1.
Contrarily, unsupervised representation learning techniques can be utilized
for unlabeled representations .π(v). Methods like principal component analysis
or auto-encoders intend to reduce the dimensionality of high-dimensional input
features. Accordingly, the goal of these algorithms is to determine alternative, low-
dimensional representations without the consideration of any target feature except
the input feature .π(v) itself. For example, auto-encoders feed a representation
.π(v) ∈ R
d ' into a deep neural network and attempt to reconstruct it, i.e., .π(v)
also serves as the output feature. However, the hidden layers are assumed to be low-
dimensional to serve as alternative representations .γ (v) ∈ Rd of .v ∈ V with .d ⪡ d '
as depicted in Fig. 2.
Fig. 3 Extract from the abstract in [15]. The semantics of the word products is encoded within the
sentences that contain it
exemplified in Fig. 3. While state-of-the-art NLP models like BERT [13] usually
split words into frequently occurring subword tokens via subword segmentation
algorithms such as Wordpiece [14], the inherent methods can be applied analogously
to sets of complete words. In the course of training such NLP models, numerical
embeddings .γ (v) ∈ Rd are assigned to the domain variables .v ∈ V with respect to
their original representations .π(v). These alternative representations are optimized
by backpropagating the output of the LLM for at least one element of its initial
representation .π(v).
Analogously, most NLP techniques can be applied to KG structures .G = (V , E)
by characterizing directed graph walks .(v1 , p1 , v2 , p2 , v3 , . . . , vl−1 , pl−1 , vl ) of
depth .l − 1 ∈ N as sentences that are composed of edges .(vi , pi , vi+1 ) ∈ E. For
instance, the sample manufacturing KG depicted in Fig. 4 contains the 4-hop walk
(John, executes, Task 1, output, Product 1, input, Task 2, output, Product 2).
One of these transfer approaches is RDF2Vec [16], which utilizes random graph
walks to generate input data for the NLP-based Word2Vec algorithm [17]. By doing
so, a mapping .γ : V ∪ R → Rd is trained and thus, alternative representations
of the graph nodes in V , but also for the relation types in R as well. Therefore,
Leveraging Semantic Representations via Knowledge Graph Embeddings 75
node embeddings can be derived via .γ (v) := γ (v). Besides transfer approaches
like RDF2Vec, various embedding algorithms exist, which are specifically tailored
toward KG structures. These are further discussed in the following.
Fig. 5 Sample KG with .n = 4 nodes and .k = 2 relations .r1 (blue) and .r2 (red), including their
respective adjacency matrices .A1 and .A2
Ah ≈ X · Rh · XT
.
of its h-th slice .Ah ∈ {0, 1}n×n by means of matrices .X ∈ Rn×d and .Rh ∈ Rd×d
with .d ⪡ n. Therefore, the i-th row of the matrix .X contains an alternative
representation .γ (vi ) := Xi,1 , . . . , Xi,d ∈ Rd of .vi ∈ V . The optimization of
the matrices .X and .(Rh )1≤h≤k is accordingly achieved by solving the minimization
problems
1 k
. minX,Rh f (X, Rh ) for f (X, Rh ) = ‖Ah − X · Rh · X ‖F ,
T 2
2 h=1
sition models for KG embeddings exist, including ANALOGY [22], SimplE [23],
and HolE [24].
(GCNs) were first introduced to account for directed edges [31]. Furthermore, to
accommodate different relation types, Relational Graph Convolutional Networks
(RGCNs) were elaborated as extensions of GCNs [32], which were subsequently
extended by means of attention mechanisms [33] in Relational Graph Attention
Networks (RGATs) [34].
In contrast to geometric KG embedding models that apply score functions
to individual triples and tensor decomposition models that intend to reduce the
dimensionality of the adjacency tensor .A, DL-based models perform predictions
for labeled nodes .v ∈ V , taking into account itself and its outgoing neighbors
These labels can be derived from the KG itself via node assertions or link
assignments, or they can be external, such as numerical or nominal node attributes.
Adjacent node representations are meant to be aggregated to receive a composite
node representation of v. By backpropagating a suitable loss, initial embeddings of
v and its neighbors are optimized. This process is repeated for each labeled training
node to generate latent feature representations for all .v ∈ V ∪ {N(v) : v ∈ V }. The
formalism proposed in [32] subdivides .N(v) into relation-specific neighborhoods
Nr (v) := {y ∈ V | ∃(s, p, o) ∈ E : (s = y ∧ o = v) ∨ (s = v ∧ o = y) ∧ p = r} ,
.
.Ah · X ∈ Rn×d ,
where .Ah denotes the h-th slice of .A. For instance, in the context of the KG from
Fig. 5, the composite representation of .v1 regarding the relation .r1 equals the sum
of the initial embeddings of .v2 and .v3 . To account for differing impacts of incoming
and outgoing edges, R is typically extended via inverse relations .r ' for each .r ∈ R.
Some works also consider a self-relation .r0 . Accordingly, by taking into account the
adjacency matrices .A0 = I d and .A2h = ATh for .1 ≤ h ≤ k, we extend the set R
via
. := R ∪ r ' | r ∈ R ∪ {r0 } with rh' = r2h .
R
By doing so, GNN models capture the semantics of directed and labeled graphs
by summing up weighted composite representations to receive a convoluted matrix
2k
. h · X · Wh ∈ Rn×d ' ,
A
h=0
Leveraging Semantic Representations via Knowledge Graph Embeddings 79
'
including relation-specific weight matrices .Wh ∈ Rd×d . Moreover, the extended
adjacency tensor .A ∈ R(2k+1)×n×n is not necessarily .{0, 1}-valued. Rather, it
is intended to contain normalization constants or attention scores to encode the
significance of individual nodes and relations to the forward pass of a GNN.
However,
. vi , rh , vj ∈ h,i,j = 0
/ E⇒A
2k
'
σ
. h · X · Wh
A =: X' ∈ Rn×d . (2)
h=0
The lack of use case scenarios poses a significant challenge to the application of
KGs and corresponding KG embeddings in the manufacturing domain. Without
specific applications, it becomes difficult to identify the relevant data sources,
design appropriate KG structures, and create meaningful embeddings that capture
the intricate relationships within manufacturing processes. Thus, the absence of
concrete use cases hinders the exploration of the full potential of KGs and KG
embeddings in improving efficiency, decision-making, and knowledge sharing
within this domain.
As a result of the research conducted within the Teaming.AI project, which
aims to enhance flexibility in Industry 4.0, while prioritizing human involvement
and collaboration in maintaining and advancing AI systems, we identified several
application scenarios within the manufacturing domain that can be leveraged by
introducing industrial KGs and KG embeddings. These are introduced in the
following.
Data Integration and Fusion Manufacturing involves diverse and complex data
from various sources, such as sensors, process logs, or maintenance records. While
KGs can integrate these heterogeneous data sources, KG embeddings map them into
a shared representation space. By representing KG nodes and their relationships in
this shared embedding space, it becomes easier to combine and analyze data from
different sources, leading to enhanced data fusion capabilities.
80 F. Krause et al.
Most of the existing works on dynamic graph embeddings do not account for
directed and labeled graphs. Rather, they are designed to be applicable to undirected
and/or unlabeled graphs [37, 38], or they aim to embed temporally enhanced
snapshots of non-dynamic graphs [39, 40]. Moreover, approaches like the one
proposed in [41] exist that intend to perform an online training of KG embeddings
by focusing on regions of the graph which were actually affected by KG updates.
However, the overall embedding structure is still affected, leading to a need for
continuous adjustments of embedding-based downstream tasks, such as graph-
based ML models. Thus, we require a dynamic KG embedding formalism that (i)
can produce real-time embeddings for dynamic KGs and (ii) is able to preserve
the original structure of KG embeddings to allow for consistent downstream
applications.
We propose to utilize the dynamic Navi approach [42], which is based on the
core idea of GNNs as per Eq. (2). Given an initial KG .Gt0 = (Vt0 , Et0 ) at timestamp
.t0 , we assume an embedding .γt0 : Vt0 → R
d based on some state-of-the-art
Since we leverage the idea of GNNs to reconstruct .γt0 (v) through local neighbor-
hoods, these reconstructions are basedon the unique adjacency tensors .(A(t))t∈T
with .A(t) ∈ Rk×nt ×nt . Here, .nt = τ ≤t Vτ denotes the number of nodes that
were known to exist since the graph’s initialization and thus .nt ≥ nt0 holds. Thus,
we assume an initial embedding matrix .Xt0 ∈ Rnt0 ×d that contains the initial
embeddings as per .γt0 . This matrix is then reconstructed based on itself via a single-
layer GNN
2k
. 0 )0 · Θt0 · W0 +
σ A(t 0 )h · Xt0 · Wh =: Xt0 ≈ Xt0
A(t
h=1
by taking into account the extended adjacency tensor .A(t 0 ) (cf. Sect. 3.1.3). During
the training process, a global embedding .γr0 ∈ R is implemented regarding the
d
self-relation .r0 so that .Θt0 ∈ Rnt0 ×d contains .nt0 copies of .γr0 . Moreover, instead
of zero-value dropouts, overfitting is prevented by randomly replacing node embed-
82 F. Krause et al.
dings with .γr0 in the input layer, simulating the semantic impact of nodes that are
not known at time .t0 . It is also used to represent self-loops, enabling reconstructions
that are independent of the (potentially unknown) initial representations. A detailed
overview, including training settings and benchmark evaluation results, can be found
in [42]. The evaluation implies that, given a timestamp .t > t0 , this approach allows
for high-qualitative and consistent embeddings .γt : Vt → Rd that are computed via
2k
0 · Θt · W0 +
σ A(t)
. h · Xt · Wh =: Xt ,
A(t)
h=1
i.e., the i-th row of .Xt represents the embedding .γt (vi ) of the node .vi ∈ Vt . In the
case of new nodes, .Xt and .Θt are the extensions of .Xt0 and .Θt0 by inserting copies
of .γr0 , respectively. Moreover, the update of the adjacency tensor can be performed
via
First, the matrix .I (t0 , t) ∈ {0, 1}nt0 ×nt accounts for newly inserted nodes, i.e.,
I (t0 , t)i,j = 1 ⇐⇒ i = j.
.
Second, the update matrices .B(t0 , t)h ∈ {−1, 0, 1}nt ×nt identify KG updates
1 ⇐⇒ the edge (vi , rh , vj ) was inserted between t0 and t
.B(t0 , t)i,j ==
−1 ⇐⇒ the edge (vi , rh , vj ) was deleted between t0 and t.
A(t ' )h = I (t, t ' )T · A(t)h · I (t, t ' ) + B(t, t ' )h for t0 < t < t ' .
.
6 Conclusions
Acknowledgments This work is part of the TEAMING.AI project which receives funding in the
European Commission’s Horizon 2020 Research Program under Grant Agreement Number 957402
(www.teamingai-project.eu).
References
1. Mohamed, S.K., Nováček, V., Nounu, A.: Discovering protein drug targets using knowledge
graph embeddings. Bioinformatics 36(2), 603–610 (2020)
2. Fu, X., Ren, X., et al.: Stochastic optimization for market return prediction using financial
knowledge graph. In: IEEE International Conference on Big Knowledge, ICBK, pp. 25–32.
IEEE Computer Society, New York (2018)
3. Buchgeher, G., Gabauer, D., et al.: Knowledge graphs in manufacturing and production: a
systematic literature review. IEEE Access 9, 55537–55554 (2021)
4. Hogan, A., Blomqvist, E., et al.: Knowledge graphs. ACM Comput. Surv. (CSUR) 54(4), 1–37
(2021)
5. Krause, F., Weller, T., Paulheim, H.: On a generalized framework for time-aware knowledge
graphs. In: Towards a Knowledge-Aware AI—Proceedings of the 18th International Confer-
ence on Semantic Systems, vol. 55, pp. 69–74. IOS Press, New York (2022)
6. Neo4j: Neo4j—the world’s leading graph database (2012)
7. Palumbo, E., Rizzo, G., et al.: Knowledge graph embeddings with node2vec for item
recommendation, In: The Semantic Web: ESWC Satellite Events, pp. 117–120 (2018)
84 F. Krause et al.
8. Diefenbach, D., Giménez-García, J., et al.: Qanswer KG: designing a portable question
answering system over RDF data. In: The Semantic Web: ESWC 2020, pp. 429–445 (2020)
9. Sun, Z., Hu, W., et al.: Bootstrapping entity alignment with knowledge graph embedding. In:
Proceedings of the 27th International Joint Conference on Artificial Intelligence, pp. 4396–
4402. AAAI Press, New York (2018)
10. Rossi, A., Barbosa, D., et al.: Knowledge graph embedding for link prediction: a comparative
analysis. ACM Trans. Knowl. Discov. Data 15(2), 1–49 (2021)
11. Nickel, M., Murphy, K., et al.: A review of relational machine learning for knowledge graphs.
Proc. IEEE 104(1), 11–33 (2016)
12. Bengio, Y., Courville, A., Vincent, P.: Representation learning: A review and new perspectives.
IEEE Trans. Pattern Anal. Mach. Intell. 35(8), 1798–1828 (2013)
13. Devlin, J., Chang, M.-W., et al.: BERT: Pre-training of deep bidirectional transformers for
language understanding. In: Proceedings of the Conference of the North American Chapter
of the Association for Computational Linguistics: Human Language Technologies, pp. 4171–
4186. Association for Computational Linguistics, Kerrville (2019)
14. Schuster, M., Nakajima, K.: Japanese and Korean voice search. In ICASSP, pp. 5149–5152.
IEEE, New York (2012)
15. Lu, Y., Xu, X., Wang, L.: Smart manufacturing process and system automation—a critical
review of the standards and envisioned scenarios. J. Manuf. Syst. 56, 312–325 (2020)
16. Ristoski, P., Rosati, J., et al.: Rdf2vec: RDF graph embeddings and their applications. Semantic
Web 10, 721–752 (2019)
17. Mikolov, T., Sutskever, I., et al.: Distributed representations of words and phrases and their
compositionality. In: Advances in Neural Information Processing Systems, vol. 26. Curran
Associates, Inc., New York (2013)
18. Kolda, T.G., Bader, B.W.: Tensor decompositions and applications. SIAM Rev. 51(3), 455–500
(2009)
19. Nickel, M., Tresp, V., Kriegel, H.-P.: A three-way model for collective learning on multi-
relational data. In: Proceedings of the 28th International Conference on International Con-
ference on Machine Learning, pp. 809–816. Omnipress, New York (2011)
20. Yang, B., Yih, W.-T., et al.: Embedding entities and relations for learning and inference in
knowledge bases. In: 3rd International Conference on Learning Representations, ICLR 2015,
San Diego, CA, USA, May 7–9, 2015, Conference Track Proceedings (2015)
21. Trouillon, T., Welbl, J., et al.: Complex embeddings for simple link prediction. In: Proceedings
of The 33rd International Conference on Machine Learning, vol. 48, pp. 2071–2080. PMLR,
New York (2016)
22. Liu, H., Wu, Y., Yang, Y.: Analogical inference for multi-relational embeddings. In: Proceed-
ings of the 34th International Conference on Machine Learning, vol. 70, pp. 2168–2178 (2017).
JMLR.org
23. Kazemi, S.M., Poole, D.: Simple embedding for link prediction in knowledge graphs. In:
NeurIPS, pp. 4289–4300. Curran Associates Inc., New York (2018)
24. Nickel, M., Rosasco, L., Poggio, T.: Holographic embeddings of knowledge graphs. In:
Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, pp. 1955–1961.
AAAI Press, New York (2016)
25. Bordes, A., Usunier, N., et al.: Translating embeddings for modeling multi-relational data. In:
Proceedings of the 26th International Conference on Neural Information Processing Systems,
vol. 2, pp. 2787–2795. Curran Associates Inc., New York (2013)
26. Wang, Z., Zhang, J., et al.: Knowledge graph embedding by translating on hyperplanes. In:
Proceedings of the AAAI Conference on Artificial Intelligence, vol. 28(1) (2014)
27. Lin, Y., Liu, Z., et al.: Learning entity and relation embeddings for knowledge graph
completion. In: Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence,
pp. 2181–2187. AAAI Press, New York (2015)
28. Ji, G., He, S., et al.: Knowledge graph embedding via dynamic mapping matrix. In: Proceedings
of the 53rd Annual Meeting of the Association for Computational Linguistics, pages 687–696.
Association for Computational Linguistics, New York (2015)
Leveraging Semantic Representations via Knowledge Graph Embeddings 85
29. Xiao, H., Huang, M., Zhu, X.: TransG: a generative model for knowledge graph embedding.
In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics,
pp. 2316–2325. Association for Computational Linguistics, New York (2016)
30. Wu, Z., Pan, S., et al.: A comprehensive survey on graph neural networks. IEEE Trans. Neural
Networks Learn. Syst. 32(1), 4–24 (2021)
31. Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. In:
ICLR (2017)
32. Schlichtkrull, M., Kipf, T.N., et al.: Modeling relational data with graph convolutional
networks. In: The Semantic Web ESWC, pp. 593–607. Springer, Berlin (2018)
33. Vaswani, A., Shazeer, N., et al.: Attention is all you need. In: Advances in Neural Information
Processing Systems, pp. 5998–6008 (2017)
34. Busbridge, D., Sherburn, D., et al.: Relational Graph Attention Networks (2019)
35. Schirner, G., Erdogmus, D., et al.: The future of human-in-the-loop cyber-physical systems.
Computer 46(1), 36–45 (2013)
36. Leng, J., Sha, W., et al.: Industry 5.0: prospect and retrospect. J. Manuf. Syst. 65, 279–295
(2022)
37. Pareja, A., Domeniconi, G., et al.: EvolveGCN: Evolving graph convolutional networks for
dynamic graphs. In: The Thirty-Fourth AAAI Conference on Artificial Intelligence AAAI, The
Thirty-Second Innovative Applications of Artificial Intelligence Conference IAAI, The Tenth
AAAI Symposium on Educational Advances in Artificial Intelligence EAAI, pp. 5363–5370.
AAAI Press, New York (2020)
38. Trivedi, R., Farajtabar, M., Biswal, P., Zha, H.: DyRep: learning representations over dynamic
graphs. In: International Conference on Learning Representations (2019)
39. Dasgupta, S.S., Ray, S.N., Talukdar, P.: HyTE: hyperplane-based temporally aware knowledge
graph embedding. In: Proceedings of the Conference on Empirical Methods in Natural
Language Processing, pp. 2001–2011. Association for Computational Linguistics, Belgium
(2018)
40. Liao, S., Liang, S., et al.: Learning dynamic embeddings for temporal knowledge graphs. In:
Proceedings of the 14th ACM International Conference on Web Search and Data Mining, pp.
535–543. Association for Computing Machinery, New York (2021)
41. Wewer, C., Lemmerich, F., Cochez, M.: Updating embeddings for dynamic knowledge graphs.
CoRR, abs/2109.10896 (2021)
42. Krause, F.: Dynamic knowledge graph embeddings via local embedding reconstructions. In:
The Semantic Web: ESWC Satellite Events, pp. 215–223. Springer, Berlin (2022)
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Architecture of a Software Platform for
Affordable Artificial Intelligence in
Manufacturing
1 Introduction
The huge transformation brought by the fourth industrial revolution into the
manufacturing world has forced any company to take on the digitalization journey,
regardless of its size, sector, or location. In this context, Artificial Intelligence (AI)
technologies are ready to take off as a new approach to solve business issues, and,
recently, AI tools are proliferating [1]. Forward-thinking results can be obtained by
analyzing huge amounts of data from a wide range of sources in the production
system and by identifying deviations and trends in real time for making decisions
[2]. The greater intelligence brought by AI embedded in production systems can
not only bring advantages for large companies but also support Small-Medium
Enterprises (SMEs) and mid-caps in achieving better operational performance. Yet
several challenges are still preventing them from embracing AI on a large scale.
To reduce the barriers, two conditions have to be satisfied: the technology has to
be affordable and accessible enough for mass use, and the level of awareness of
individuals and companies should be high enough to be able to understand how and
where to use it.
The first condition can be tackled by democratizing AI tools: by exploiting the
“as-a-service-model,” technologies can be made available to SMEs at an affordable
price and on-demand, thus reducing the financial gap with large companies and
avoiding SMEs getting lost in the hype around AI [3]. This is the best solution since,
on the one hand, the adoption of ad hoc solutions for specific company requirements
leads to integration problems, long implementation times, and flexibility limits. On
the other hand, adopting an all-in-one solution requires big investments for complex
systems, which exceed the effective needs and strictly depend on legacy providers.
The second condition is more difficult to be satisfied at the level of single
companies since, often, SMEs lack the skills and knowledge needed to turn AI into a
major boost for their business, thus lagging behind larger organizations in the uptake
level [4]. Successful implementation of AI requires identifying the right tools to
be chosen among a plethora of solutions and their harmonization with existing IT
systems and processes from both a technical and a strategic point of view so that
they can become real enablers for performance improvement. Upskilling workers
is essential to both empower a mutually supportive human–machine interaction and
lower adoption barriers, but building internal competences requires time. Support is
needed now to accompany European SMEs in their digitization journey so that they
can keep the pace with their larger counterparts and be key players in their value
chains. An innovation ecosystem should be created around SMEs so that they can
easily find locally the needed support to draw customized AI adoption plans and be
immersed in a vibrant and stimulating environment that makes them progress across
the digital innovation flow.
At the European level, initiatives have been launched to promote the development
of platforms that could support SMEs in the digital uptake, and the creation of local
Digital Innovation Hubs (DIHs) is promoted to create an innovation ecosystem
providing services to lower the entry barriers for SMEs. The AI uptake has to
pivot on digital platforms that act as one-stop shop for SMEs, showcasing advances
brought forward by synergistic efforts of DIHs, research centers, and technologies
providers and offering services to smooth the adoption. Being able to offer to SMEs
solutions tailored to their specific needs, built on modular kits, at a reasonable
cost, easy and fast to implement is a must to strengthen the European economy’s
competitiveness.
KITT4SME recognizes that SMEs are among the companies that could benefit
the most from the opportunities brought by AI solutions while, at the same
time, being the ones with the least capabilities and resources to embrace them.
KITT4SME specifically targets European SMEs and mid-caps to provide them
with scope-tailored and industry-ready hardware, software, and organizational kits,
delivered as a modularly customizable digital platform that seamlessly introduces
AI in their production systems. Uptake of the resulting packages and of the provided
Architecture of a Software Platform for Affordable AI in Manufacturing 89
1 https://www.fiware.org/.
90 V. Cutrona et al.
2 https://www.acumos.org/.
3 https://lfaidata.foundation/.
4 https://www.bonseyes.eu/.
5 https://www.nviso.ai.
Architecture of a Software Platform for Affordable AI in Manufacturing 91
platform. Companies evaluated and paid the winners after the call ended. Bonseyes
used a collaborative AI Marketplace to provide real-world solutions to the industry,
supporting scenarios where data must remain on the data provider’s premises and
online learning with distributed Cyber-Physical Systems (CPSs). The platform
allowed continuous feedback from human actors to evaluate model performance
and obtain metadata about context and users’ perspectives [8, 9].
GRACE AI Grace AI6 is an AI platform launched by 2021.AI in 2018, with the
mission to help customers in realizing their vision of AI by identifying innovative
business opportunities in key processes and functions. Grace AI Platform and the
AIaaS portfolio are the company’s main assets. The Grace platform is built for both
organizations at the beginning of their AI and ML journey and organizations that
have already established a data science team but are looking for ways to infuse
continuous intelligence into their business.
Grace AI aims to provide any organization access to AI implementation, includ-
ing automated documentation, validation, and certification through data exploration,
AI development, deployment, and operation.
PTC Inc. PTC Inc.7 is a software and services company founded in 1985, based
in Boston. It offers a range of products and services that support innovation and
Industry 4.0. It is a platform for developing IoT and Augmented Reality (AR)
solutions. PTC Marketplace is a digital space where customers and partners can
access IoT apps, market-ready solutions, and innovative technologies. PTC has
made recent enhancements to its marketplace, making it easier for solution builders
to find market-ready solutions and customized accelerators. It also provides a
platform for PTC partners to showcase their technologies, solutions, services, and
industry expertise to customers and prospects.
The platform offers a rich set of capabilities that enable solutions for design, man-
ufacturing, service, and industrial operations and incorporates modular functionality
that simplifies development. These include pre-built applications for the fast, easy
implementation of Industrial Internet of Things (IIoT) solutions for common use
cases in various industries.
6 https://2021.ai/offerings/grace-ai-platform/.
7 https://www.ptc.com/.
92 V. Cutrona et al.
development and the capabilities of SMEs’ IT systems. This often leads to situations
where models are not deployed or where the deployment and update process is time-
consuming.
To tackle these challenges, KITT4SME proposes a conceptual pipeline consist-
ing of six steps, which cover the process from data preparation to the practical use
of the model. The steps (presented in Fig. 1) are as follows:
1. Prepare data: This step involves collecting and preparing the data required for
training the AI model. It includes tasks such as data cleaning, transformation,
and feature engineering to ensure the data are suitable for model development.
2. Develop the model: In this step, AI researchers and developers focus on building
and training the AI model using the prepared data. This is where the core value
of the AI solution is generated.
3. Package the model: Once the model is developed, it needs to be packaged in
a way that it can be easily deployed and integrated into the existing systems
of the SME. Packaging involves encapsulating the model and its associated
dependencies into a deployable form.
4. Validate the model: Before deployment, it is crucial to validate the model to
ensure its accuracy, reliability, and suitability for the intended use. Validation
may involve testing the model’s performance on a separate dataset or using
techniques like cross-validation.
94 V. Cutrona et al.
5. Deploy the model: This step focuses on deploying the validated model into
the SME’s IT infrastructure. It involves integrating the model with the existing
systems, ensuring compatibility, and addressing any technical requirements or
constraints.
6. Use the model: The final step is when the SME can actively utilize the deployed
model in its operations. This includes making predictions, generating insights,
and incorporating the model’s outputs into decision-making processes.
The three intermediate steps, namely packaging, validating, and deploying the
model, are often complex and time-consuming. KITT4SME aims to simplify and
automate these steps, reducing the overall time and effort required to deploy and
update the AI model. By streamlining these processes, the platform enhances the
repeatability and efficiency of the entire pipeline, making it easier for SMEs to
leverage AI technologies effectively.
The software platform implementing the KITT4SME workflow is based on a
service mesh, multi-tenant cloud architecture. It provides a means to assemble
various AI components from a marketplace and facilitates their connection to the
shop floor while ensuring interoperability, security, and privacy-preserving data
exchange. The platform consists of loosely coupled web services running in a
cluster environment and relies on a dedicated cluster software infrastructure. Several
key concepts and guiding principles underpin the architecture of the KITT4SME
platform:
• Leveraging state-of-the-art technology and standards: The platform utilizes a
dedicated cluster software infrastructure, referred to as mesh infrastructure.
This infrastructure is built on industry-standard technologies such as Kuber-
netes8 and Istio9 . The platform reuses open communication and data standards
as much as possible to foster service interoperability (e.g., REST principles for
services interaction and NGSI standard for data exchange).
• Platform services: The platform comprises two types of services: application
services, which are integral to the KITT4SME workflow and provide the func-
tionality required for the platform’s core activities, and infrastructure services,
which consist of a network of intermediaries within the mesh infrastructure.
These intermediaries handle essential operational aspects such as routing,
security, and monitoring. By separating these concerns, AI developers can focus
on implementing service-specific features while relying on the platform for
operational support.
• Multi-tenancy: The platform is designed to support multiple SMEs sharing the
same instance. Each company is associated with a security protection domain,
referred to as a tenant, which isolates its data and users from other tenants.
The platform also allows for explicit sharing policies that enable companies to
selectively share data and resources if desired.
8 https://kubernetes.io/.
9 https://istio.io/.
Architecture of a Software Platform for Affordable AI in Manufacturing 95
10 https://www.fiware.org/.
Architecture of a Software Platform for Affordable AI in Manufacturing 97
Fig. 2 Powered by FIWARE KITT4SME architecture. Components represented as blue boxes are
from the FIWARE reference architecture [10, 11]
Data Gathering The data gathering aspect of the KITT4SME architecture encom-
passes the collection of data from diverse devices, situated at the lowest layer of
the architecture (as depicted in Fig. 2). These devices, deployed within the factory,
serve to enrich the system’s knowledge base with both raw and preprocessed data.
The following categories of devices contribute to the data gathering process:
• Wearable Sensors: These sensors are specifically designed to monitor the health
and well-being of workers within the factory setting. They provide valuable
insights into various physiological parameters and indicators.
• Environmental Sensors: Scattered throughout the factory, environmental sensors
play a vital role in monitoring and capturing data related to the prevailing
environmental conditions. This includes parameters such as air pollution levels,
temperature, and humidity.
• CPSs: The architecture also incorporates CPSs, with a particular emphasis on
those commonly involved in the manufacturing processes, such as machining
equipment and collaborative robots. These CPSs facilitate the capture of
relevant data pertaining to the operational aspects of the production line.
• Information Systems: Information systems represent a valuable source of raw
and value-added data, which contribute to update the contextual information of
the platform also with aggregated data.
• Cameras and IoT sensors: Together with environmental sensors, cameras and
IoT sensors are needed to monitor the production, usually requiring a real-time
processing to extract valuable knowledge from data streams.
Data Broker In the layer above, the FIWARE Orion Context Broker represents
the fundamental component of any solution powered by FIWARE. This context
broker facilitates the decentralized and scalable management of context information,
allowing data to be accessed through a RESTful API. Serving as the authoritative
source of information, the Context Broker stores the latest update status of all
devices, components, and processes that contribute data to the platform.
However, for the purpose of training and fine-tuning AI tools, it is often necessary
to access historical data. To address this requirement, FIWARE offers dedicated GEs
called QuantumLeap that automatically generate time series data from the evolving
context information, enabling AI tools to leverage the valuable insights gained from
historical data analysis.
Smart Industry Management Services The topmost layer of the architecture
encompasses analytical services and profilers that leverage the knowledge base
within the system. These services include Big Data applications and components
utilizing AI-based detection and optimization tools. It is in this layer that AI
developers and researchers can greatly benefit from the historical data and up-to-
date context information made available by the Powered by FIWARE platform.
Additionally, the KITT4SME architecture incorporates utility components in this
layer to extract additional knowledge from persistent information and provide
insights to human actors regarding the factory’s status. These components include:
• Human Description Database, which stores a comprehensive representation of
factory workers derived from physiological parameters, worker information,
machine parameters, and environmental data
• External IDS Connector, a component from the IDSA reference architecture12
that ensures a trustworthy interface between internal data sources and external
data consumers. This connector plays a critical role in enabling the integration
11 https://www.fiware.org/catalogue/.
12 https://docs.internationaldataspaces.org/ids-ram-4/.
Architecture of a Software Platform for Affordable AI in Manufacturing 99
The KITT4SME platform has been applied in 4 use cases within the KITT4SME
project and 18 external demonstrators made via Open Calls.13
In this section, we discuss how the KITT4SME platform has been exploited to
create an AI kit supporting one of the internal use cases. This use case is from
the injection molding industry, and it aims at facilitating an assembly task mainly
composed of screwdriving operations. The assembly task starts with a molding
13 https://kitt4sme.eu/open-call/.
100 V. Cutrona et al.
Fig. 3 The KITT4SME solution for the use case in the injection molding sector
press producing a molded piece every 90 seconds. Then, the task foresees a gantry
robot that automatically extracts the molded piece from the injection molding
machine and places it onto a conveyor belt. Subsequently, a human operator works
at a designated workstation to perform the assembly operations while also being
responsible for periodic quality checks on the molded pieces or quick maintenance
operations on the injection molding machine.
The KITT4SME platform has introduced an AI solution to mitigate workers’
physical stress caused by heavy workloads and the injection molding machine’s
demanding pace during operations. In particular, the use case relies on the concept of
human digital twin [13]. A dynamic task assignment between the collaborative robot
and the operator is performed by creating a digital representation of the operator and
the production system.
The Kit used for this use case, represented with its whole architecture in Fig. 3,
includes:
• Sensing Layer: This module supports the collection and use of IoT sensor data
to be used by data analysis and decision-making modules to take decisions or to
be visualized on dashboards. It provides a solution including the interoperability
elements (APIs and broker client) for bidirectional data exchange between
sensors and the KITT4SME’s Orion Context Broker. Data are also preprocessed
if needed.
• Fatigue Monitoring System: It is an AI model that estimates the perceived
fatigue of the workers based on physiological data (e.g., heart rate) from
wearable devices and on quasi-static characteristics (e.g., age). The estimation
is made using physiological data collected from wearable devices selected
by applying an Analytic Hierarchy Process (AHP)-based methodology [14]
and operator’s characteristics, including age, sex, handedness, exercise/healthy
habits, and routines, collected via interviews.
• Intervention Manager: It monitors the real-time status of the worker–factory
ecosystem, elaborating data from sensors, machines, workers monitoring sys-
Architecture of a Software Platform for Affordable AI in Manufacturing 101
tems, and ERP, and it knows what interventions can be applied and which are
the rules to decide which is the best one given a particular situation. It applies
AI models specifically developed to support decision-making.
The kit has been deployed to the platform to support the task assignment in a
screwdriving process in the following process:
1. The operator retrieves two molded parts from a conveyor belt and positions them
on the working bench.
2. The operator inserts six nuts into each part, flips one part, and places it on top of
the other.
3. The operator positions nine screws on one side of the assembled parts.
4. The Intervention Manager assigns each screw to the operator or the cobot.
The operator and a cobot simultaneously perform the screwdriving process.
Depending on the number of screws assigned to the operator, they may also
engage in other support activities, such as monitoring other machines, conducting
short maintenance operations, or removing the pallet.
5. The operator flips the assembled parts and repeats steps 3 and 4.
6. The assembled parts are stacked on a pallet.
The task assignment, performed by the Intervention Manager and confirmed by
the operator, consists of the allocation of the screwing operations (9x2 for each
assembled part), and it is made considering the following parameters:
• Current perceived fatigue of the operator as estimated by the Fatigue Monitoring
System.
• Work In Progress level.
• Cobot state (idle, current operation, and error).
Discussion The above use case exemplifies how the KITT4SME platform can actu-
ally ease AI adoption by SMEs, compared to other platforms in the AI ecosystems.
Indeed, compared to platforms to develop AI solutions, the SME from the use case
did not spend any effort on developing AI, given that they exploited the existing
application available on the platform. Also, the platform helped the company
compose the best kit to solve a real need, i.e., to facilitate an assembly task mainly
composed of screwdriving operations. The proposed kit already included all the
components needed to be implemented in the factory, i.e., data acquisition (Sensing
Layer), AI solution to derive data-driven knowledge (Fatigue Monitoring System),
and a reasoning engine (Intervention Manager), relieving the company from extra
development activities needed to connect the shop floor to the platform. Instead,
by considering platforms providing already developed applications, a similar use
case has been successfully tested in a laboratory environment [15], exploiting a
different IIoT platform [16]. This kind of platform enables the handling of third-
party applications, with no guarantees about the interoperability of components
in terms of application interfaces and data models, which are covered within the
KITT4SME platform by the FIWARE components. Also, while this kind of platform
comes with a ready-to-use solution, integrating and deploying such solutions is
often a burden solely on the developers. Again, the KITT4SME platform offers a
102 V. Cutrona et al.
5 Conclusion
Acknowledgments This work has been partly supported by EU H2020 research and innovation
program project KITT4SME (grant no. 952119).
References
1. Wu, S.-Y.: Key technology enablers of innovations in the ai and 5g era. In: 2019 IEEE
International Electron Devices Meeting (IEDM), pp. 36–3. IEEE (2019)
2. Alexopoulos, K., Nikolakis, N., Chryssolouris, G.: Digital twin-driven supervised machine
learning for the development of artificial intelligence applications in manufacturing. Int. J.
Comput. Integr. Manuf. 33(5), 429–439 (2020)
14 https://github.com/c0c0n3/kitt4sme.
Architecture of a Software Platform for Affordable AI in Manufacturing 103
3. Elger, P., Shanaghy, E.: AI as a Service: Serverless Machine Learning with AWS. Manning
Publications (2020)
4. Bettoni, A., Matteri, D., Montini, E., Gladysz, B., Carpanzano, E.: An ai adoption model
for SMEs: a conceptual framework. IFAC-PapersOnLine 54(1), 702–708 (2021). 17th IFAC
Symposium on Information Control Problems in Manufacturing INCOM 2021
5. IBM: What is platform-as-a-service (PaaS)? https://www.ibm.com/topics/paas (2021).
Accessed 08 Aug 2023
6. Lins, S., Pandl, K.D., Teigeler, H., Thiebes, S., Bayer, C., Sunyaev, A.: Artificial intelligence
as a service. Bus. Inf. Syst. Eng. 63(4), 441–456 (2021)
7. Zhao, S., Talasila, M., Jacobson, G., Borcea, C., Aftab, S.A., Murray, J.F.: Packaging and
sharing machine learning models via the Acumos AI open platform. In: 2018 17th IEEE
International Conference on Machine Learning and Applications (ICMLA), pp. 841–846. IEEE
(2018)
8. Llewellynn, T., Fernández-Carrobles, M.M., Deniz, O., Fricker, S., Storkey, A., Pazos, N.,
Velikic, G., Leufgen, K., Dahyot, R., Koller, S., et al.: Bonseyes: platform for open
development of systems of artificial intelligence. In: Proceedings of the Computing Frontiers
Conference, pp. 299–304, 2017
9. Prado, M.D., Su, J., Saeed, R., Keller, L., Vallez, N., Anderson, A., Gregg, D., Benini, L.,
Llewellynn, T., Ouerhani, N., et al.: Bonseyes ai pipeline—bringing ai to you: End-to-end
integration of data, algorithms, and deployment tools. ACM Trans. Internet Things 1(4), 1–25
(2020)
10. Ahle, U., Hierro, J.J.: Fiware for data spaces. In: Otto, B., ten Hompel, M., Wrobel, S. (eds.),
Designing Data Spaces: The Ecosystem Approach to Competitive Advantage, pp. 395–417.
Springer International Publishing (2022)
11. Fiware Smart Industry Reference Architecture: https://www.fiware.org/about-us/smart-
industry/ (2022). Access 08 Aug 2023
12. Macenski, S., Foote, T., Gerkey, B., Lalancette, C., Woodall, W.: Robot operating system 2:
Design, architecture, and uses in the wild. Sci. Robot. 7(66), eabm6074 (2022)
13. Montini, E., Bettoni, A., Ciavotta, M., Carpanzano, E., Pedrazzoli, P.: A meta-model for
modular composition of tailored human digital twins in production. Procedia CIRP 104, 689–
695 (2021)
14. Montini, E., Cutrona, V., Gladysz, B., Dell’Oca, S., Landolfi, G., Bettoni, A.: A methodology
to select wearable devices for industry 5.0 applications. In: 2022 IEEE 27th International
Conference on Emerging Technologies and Factory Automation (ETFA), pp. 1–4. IEEE (2022)
15. Montini, E., Cutrona, V., Dell’Oca, S., Landolfi, G., Bettoni, A., Rocco, P., Carpan-
zano, E.: A framework for human-aware collaborative robotics systems development.
Procedia CIRP 120, 1083–1088 (2023). 56th CIRP Conference on Manufacturing
Systems 2023. https://doi.org/10.1016/j.procir.2023.09.129. https://www.sciencedirect.com/
science/article/pii/S2212827123008612
16. Montini, E., Cutrona, V., Bonomi, N., Landolfi, G., Bettoni, A., Rocco, P., Carpanzano, E.: An
IIoT platform for human-aware factory digital twins. Procedia CIRP 107, 661–667 (2022)
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Multisided Business Model for Platform
Offering AI Services
1 Introduction
Platform businesses have become one of the latest research topics in various
management disciplines [10]. A platform is an interface that facilitates interac-
tions between different parties, usually complementors and customers [9]. In the
platform business, the platforms and their complementors have a strong one-way
complementarity, where the total value of the platform and its complementors is
more than the sum of the two combined [18], and this complementarity requires the
interdependencies between the platforms and the complementarities to be managed
in an ecosystem level.
There are two basic types of platforms: innovation platforms (as an intermediary
for direct exchange or transactions) and transaction platforms (as a technological
foundation upon which other firms develop complementary innovations). Some
companies combine the features of the two and create “hybrid platforms” [13]. Mul-
tisided platforms (MSPs) allow direct interactions between two or more different
entities, where each entity is associated with the platform [17]. Examples of well-
known MSPs include Facebook, Uber, PayPal, Airbnb, Alibaba, eBay. The growing
interest in MSPs is due to two key factors: their essential role in minimizing the
transaction costs between sides [15] and the power of the business models (BM)
in the digital economy because of their ability to adapt and cope with complexity,
rapid scaling, and value capture [1]. Although many companies are opting for MSP
BMs, only a few have been successful. MSPs should strive to attract users and must
achieve direct and indirect network effects to be successful. More importantly, they
ought to solve the chicken-or-egg problem, which refers to a network effect meaning
“one side of the market realizes the value only if another side is fully engaged” [13].
1 https://www.boardofinnovation.com/tools/business-model-kit/
2 https://leanstack.com/lean-canvas
3 https://www.boundaryless.io/pdt-toolkit/
Multisided Business Model for Platform Offering AI Services 107
to provide companies with support in describing the platform’s vision, the core
and ancillary value propositions, the platform’s infrastructure and core components,
and the characteristics of the platform ecosystem expressed through transaction
dynamics [5, 11]. It was optimized to support the development of multisided,
transformative platform strategies to empower ecosystems to create shared value.
It is an open-source method adopted worldwide by global Fortune 500 leaders,
leading institutions, start-ups, and scale-ups. The PDT covers all stages, from
exploration to design, validation, and growth. The core of the PDT methodology
in developing a business model is the design stage: an extensive and proven step-
by-step process that helps move from contextualizing entities in the ecosystem, their
role and relationships, detailing possible transactions between entities, to designing
the platform experience.
PDT, in the design stage, contains eight templates (canvases) to be completed,
considering as many aspects of the business. The steps are as follows:
1. Mapping the ecosystem: entities present in the ecosystem are mapped onto the
canvas, allowing us to understand the role they may play and identify possible
clusters.
2. Portraying ecosystem’s entities roles: a coherent and deep picture of the role of
each of the entities identified in step 1 is created by defining what their context
is, what they want to achieve, with whom and how they want to integrate, what
potential they can represent and what kind of experience gains they are looking
for, and what the platform shaper can provide them with.
3. Analyzing the potential to exchange value: using the so-called “ecosystem’s
motivation matrix,” entities’ potential to exchange value flows is analyzed. This
is a mapping of what type of value exchange is already being performed (or
attempted to be performed) by the entities and what additional value they could
exchange if adequately enabled.
4. Choosing the core relationships you want to focus on: the platform shaper needs
to identify which entities in the ecosystem they want to focus on and which
relationships will form the core of the platform design.
5. Identifying the elementary transactions: the “transaction board” tool is used to
map how the ecosystem currently exchanges value (focusing on the entities and
relationships prioritized in step 4) and how the platform’s strategy is to help them
make value transactions more manageable, faster, and cheaper by providing and
curating channels and contexts that increase the likelihood of interactions and
transactions.
6. Designing the learning engine: through the “learning engine canvas,” a step-
by-step process has been designed to support/enable services that will support
entities to adopt the platform strategy. These services will not only help them
evolve and become better producers and consumers but also radically evolve and
discover new opportunities and behaviors that were not initially intended.
7. Assembling the platform experiences: with the “platform experience canvas,”
the elements emerged from the transaction board (step 5) and those from the
learning engine canvas (step 6) are combined to create an experience persistence
108 K. Ejsmont et al.
model that summarizes the key value propositions arising from the strategy being
developed. This allows consideration of what resources and components need to
be put in place and managed to deliver these experiences and derive value from
them.
8. Setting up the minimum viable platform (MVP): this allows us to test in the
natural environment (market) whether the design assumptions are suitable for
the future. By analyzing design products, in particular the compiled “platform
experience canvases” (step 7), the riskiest assumptions of the strategy are iso-
lated, as well as experiments and indicators to validate them with the ecosystem
are identified.
The resulting business model is then summarized in the platform design canvas,
which is the final output of this reference methodology. According to the author’s
knowledge and experience, by far, the most essential element of business models for
MSPs is to identify the value that can be transferred to the different entities through
the platform [12]. Taking this into account, it was decided to focus on the first five
steps of the PDT methodology.
4 https://kitt4sme.eu/
Multisided Business Model for Platform Offering AI Services 109
in the ecosystem; supply entities (partners, peer producers) – they are interested in
“producing” the value consumed in the ecosystem.
Considering a single entity, its position in this framework may vary. For example,
an AI developer (peer producer) may become a partner after a certain period of
time if it provides many AI solutions and takes an active part in the development
of the platform. An entity may also have a dual role, as access to the platform may
create new opportunities: a company initially interested in offering its products (peer
producer) may later be interested in using its belonging to the ecosystem to seek
ideas for improving manufacturing processes in SMEs (peer consumer).
In the second step, the aim is to develop a portrait of the leading entities accessing
the platform from both the demand and supply sides. It should be noted that this
second step aims to map what the entities are currently looking for rather than what
the idea behind the platform service is. Thus, it is possibly better to characterize the
value from their point of view. In the KITT4SME ecosystem, six different entities
have been identified (Fig. 1). Figure 2 shows a portrait of AI developers, as they
appear to be the most important in the initial lifecycle of the platform – they will
be responsible for delivering AI solutions/services that can be transacted. Similarly,
portraits should be taken for all other identified entities.
The ecosystem motivation matrix (step 3) maps the values exchanged between
pairs of entities through the KITT4SME platform. Money is undoubtedly exchanged
as a consequence of interactions through the platform, but even more important for
shaping the KITT4SME BM is the identification of intangible values resulting from
the opportunities the platform brings. The matrix shown in Fig. 3 details the central
Multisided Business Model for Platform Offering AI Services 111
values exchanged between peer consumers (PC), peer producers (PP), and partners
(Pa) – previously mapped in the ecosystem canvas (Fig. 1). The cells report what the
entity in the first column from the left can “give to” the entities on the upper axis.
The goal of the fourth step is to decide which subset of relationships to focus
on to ensure that enough attention is paid to defining and implementing the core
112 K. Ejsmont et al.
experience. The value flows identified in the ecosystem motivation matrix (Fig.
3) were transferred to the ecosystem map (Fig. 1). Figure 1 shows the division
of relationships into those relating to resource sharing (brown lines) and those
supporting AI solution implementation (blue lines). In the first case, entities contact
each other to share resources. Manufacturing SMEs in this context seek dedicated
AI solutions to develop and improve their production capabilities. The remaining
entities, i.e., AI developers, cloud providers and platform components providers,
are identified as suppliers and partners, offering their knowledge, expertise, and
AI solutions through the platform. Supporting AI solution implementation is a
relationship that involves entities seeking to collaborate on creating and improving
AI solutions.
The identification of the underlying transactions and channels serves to illustrate
how the ecosystem exchanges value (step 5) and highlights the role of the
KITT4SME platform as an intermediary in this process. Most of the interactions
take place through the platform itself, which creates value from the exchange
of information, while the three interactions involving the exchange of software
(AI solution/module), AI service (e.g., support to solving problems using AI,
implementation AI solution, consultation), and payment are realized outside the
platform.
The transaction matrix helps analyze the relationship between the demand side
(entity 1) and supply side (entity 2). It helps identify all transactions/interactions and
their channels that are already taking place or may take place. In addition, for each
transaction/interaction is assigned what is the unit of value. One of the key roles of
the platform (owner) is to create channels that can reduce coordination/transaction
costs.
The transaction matrix (Table 1) confirms that the KITT4SME platform is the
main channel of interaction and, to be successful during each interaction, the
exchange of information must add value for the stakeholders. A crucial role of the
platform is to participate in the facilitation of the communication process actively
and the interaction between stakeholders, thereby reducing transaction costs and
facilitating transactions.
The analyses conducted in the previous chapters were finally aggregated into the
platform design canvas and structured as follows:
• Enabling services (platform to partners): focused on helping partners gener-
ate/create value from their assets and capabilities, access to potential consumers,
increasing competitiveness and visibility, and decisively improving as profes-
sional entities (reputation). For KITT4SME, these are designed services to
facilitate the implementation of technical specifications and core service stan-
Table 1 KITT4SME transaction board for core relationships
Currency/value
Entity 1 Transaction/ interaction Entity 2 unit Channel or context
Mfg. SMEs Request for support to solve a AI developers Information KITT4SME platform
problem/implement a solution Cloud providers
High level of request Platform components providers
Request or complete NDA Multipliers
Sending enquiry Know-how providers
Communication/interaction
Sending offer
Reply to the offer
Mfg. SMEs Development/tailored of an AI solution AI developers AI solution Companies developing AI solutions
Problem-solving/implementation of AI Cloud providers AI module Software houses
solution Platform components providers AI service Companies implementing AI
Know-how providers solutions
AI consulting firms
Mfg. SMEs Testing/validation of the AI solution AI developers Information KITT4SME platform
Cloud providers
Platform components providers
Mfg. SMEs Corrections/elimination of errors AI developers AI solution Companies developing AI solutions
Cloud providers AI module Software houses
Multisided Business Model for Platform Offering AI Services
Multipliers
Know-how providers
114 K. Ejsmont et al.
• Infrastructure and core components: these are assets owned and controlled by
the platform owner. They are managed according to the platform’s governance
rules. Assets can be tangible (e.g., server or venue) or intangible (e.g., common
standard – FIWARE). They guarantee the platform’s operation and use by the
ecosystem. KITT4SME identifies the critical elements of the platform’s IT
environment as the core components of the platform BM, namely the AI module
standards, protocols, the standard enablers (CPS-izers, runtime), codes, and the
functionalities and channels that enable its dissemination (such as RAMP).
• Transactions: are part of a more complex “experience.” They should be under-
stood as a sub-activity during which value is created, delivered, exchanged, or
transferred between typically two (or more) platform users. KITT4SME assumes
two main types of transactions: the first is intangibles (information), which the
platform completes by providing it through the systems typically used in such
kinds of platforms; the second is monetary and related to AI services that are
exchanged through the platform (AI solutions, applications, modules, services,
runtime).
• Channels and contexts: enable exchanges within the platform and are the
platform’s interface with users. Channels are user touch points that play an
essential role in the user experience. They are crucial in creating added value:
they should be actively created and continuously improved by the platform
owner. The marketplace should be considered the principal channel provided
by the KITT4SME ecosystem, where AI solutions, applications, modules, and
services are purchased, exchanged, transferred, and downloaded, respectively.
Channels for exchanging/obtaining information and processing payments are
also important.
Fig. 4 Revenue streams in the platform within the KITT4SME ecosystem entities defined in PDT
need to work together to generate value using the KITT4SME platform (enhancing
the quality of the match). Interactions can also occur between peer consumers
and advertisers (very often advertisers will be peer producers, but not only,
e.g., consultants). In this case, no transaction is taking place. Furthermore, the
KITT4SME platform can enable advertising services or matching offers and charge
an advertising fee for this and charge a premium fee for continued access to all
KITT4SME services (i.e., a membership fee).
118 K. Ejsmont et al.
In addition to most traditional strategies for defining business models, this study
allowed us to understand better the users’ needs of the platform offering AI services,
to identify the values that can be exchanged through the platform, and to formalize
the relationships and partnership mechanisms between entities accessing the MSP.
This was done using the platform business model developed for the KITT4SME
ecosystem as a case study.
The adoption of the PDT method has shown that this tool provides a relevant
methodological approach to define business model scenarios dedicated to MSPs
qualitatively. Dividing the development of a business model into a few canvases
allows one to focus on the different steps and to go deeper into the details of their
design. The first five stages of the PDT have made it possible to define which entities
can exchange values through which transaction channels. Although the completion
of the canvases still does not allow a quantitative approach to assess the extent to
which the elaborated BM can remain sustainable under the dynamic evolution of the
boundary conditions.
The following steps should be setting up the MVP and determining the value of
the different fees charged for using the platform. The KITT4SME project will be
used as a case study for these steps. In this way, the canvases proposed by Cicero
[11] will be expanded by developing a methodology that guides the user to quantify
the BM elements required for economic feasibility.
Acknowledgments The work presented in this chapter was developed within the project
“KITT4SME – platform-enabled KITs of arTificial intelligence FOR an easy uptake by SMEs”
and received funding from the European Union’s Horizon 2020 research and innovation program
under grant agreements No. 952119.
References
1. Abdelkafi, N., Raasch, C., Roth, A., Srinivasan, R.: Multi-sided platforms. Electr. Mark. 29,
553–559 (2019). https://doi.org/10.1007/s12525-019-00385-4
2. Allweins, M.M., Proesch, M., Ladd, T.: The platform canvas—conceptualization of a design
framework for multi-sided platform businesses. Entrepren. Educ. Pedagogy. 4(3), 455–477
(2021). https://doi.org/10.1177/2515127420959051
3. Armstrong, M.: Competition in two-sided markets. Rand J. Econ. 37, 668–691 (2006)
4. Azadegan, A., Papamichail, K.N., Sampaio, P.: Applying collaborative process design to user
requirements elicitation: a case study. Comput. Ind. 64(7), 798–812 (2013). https://doi.org/
10.1016/j.compind.2013.05.001
5. Barni, A., Montini, E., Menato, S., Sorlini, M., Anaya, V., Poler, R.: Integrating agent
based simulation in the design of multi-sided platform business model: a methodological
approach. In: 2018 IEEE International Conference on Engineering, Technology and Inno-
vation (ICE/ITMC), Stuttgart, Germany, 17–20 June 2018 (2018) https://doi.org/10.1109/
ICE.2018.8436360
Multisided Business Model for Platform Offering AI Services 119
6. Belleflamme, P., Peitz, M.: Platform competition and seller investment incentives. Eur. Econ.
Rev. 54, 1059–1076 (2010). https://doi.org/10.1016/j.euroecorev.2010.03.001
7. Bettoni, A., Corti, D., Matteri, D., Montini, E., Fiorello, M., Masiero, S., Barut, Z.M.,
Gretenkord, S., Gladysz, B., Ejsmont, K.: KITT4SME report 2021: Artificial Intelligence
adoption in European Small Medium Enterprises (2021). https://kitt4sme.eu/wp-content/
uploads/2021/09/AI-readinessSurvey_resultstorespondentsV8.1.pdf. Accessed 12 July 2023
8. Bettoni, A., Matteri, D., Montini, E., Gładysz, B., Carpanzano, E.: An AI adoption model for
SMEs: a conceptual framework. IFAC-Papers OnLine. 54(1), 702–708 (2021). https://doi.org/
10.1016/j.ifacol.2021.08.082
9. Boudreau, K., Hagiu, A.: Platform rules: multi-sided platforms as regulators. SSRN Electron.
J. (2008). https://doi.org/10.2139/ssrn.1269966
10. Chen, L., Tong, T.W., Tang, S., Han, N.: Governance and design of digital platforms: a review
and future research directions on a meta-organization. J. Manag. 48(1), 147–184 (2022). https:/
/doi.org/10.1177/01492063211045023
11. Cicero, S.: That’s cognitive capitalism (2015). https://medium.com/@meedabyte/that-s-
cognitivecapitalism-baby-ee82d1966c72. Accessed 12 July 2023
12. Corti, D., Bettoni, A., Montini, E., Barni, A., Arica, E.: Empirical evidence from the design
of a MaaS platform. IFAC Papers OnLine. 54(1), 73–79 (2021). https://doi.org/10.1016/
j.ifacol.2021.08.008
13. Cusumano, M.A., Gawer, A., Yoffie, D.B.: The Business of Platforms: Strategy in the Age of
Digital Competition, Innovation, and Power. Harper Business, New York (2019)
14. De Matta, R., Lowe, T.J., Zhang, D.: Competition in the multi-sided platform market channel.
Int. J. Prod. Econ. 189, 40–51 (2017). https://doi.org/10.1016/j.ijpe.2017.03.022
15. Hagiu, A.: Pricing and commitment by two-sided platforms. Rand J. Econ. 37(3), 720–737
(2006) http://www.jstor.org/stable/25046268
16. Hagiu, A., Wright, J.: Marketplace or reseller? Manag. Sci. 61, 184–203 (2015). https://doi.org/
10.2139/ssrn.2794585
17. Hagiu, A., Wright, J.: Multi-sided platforms. Int. J. Ind. Organ. 43 (2015). https://doi.org/
10.2139/ssrn.2794582
18. Holgersson, M., Baldwin, C.Y., Chesbrough, H., Bogers, M.L.A.M.: The forces of
ecosystem evolution. Calif. Manag. Rev. 64(3), 5–23 (2022). https://doi.org/10.1177/
00081256221086038
19. KITT4SME D1.1 Stakeholder analysis: (2021) https://kitt4sme.eu/wp-content/uploads/2021/
09/kitt4sme-d1.1-stakeholder-analysis.pdf. Accessed 12 July 2023
20. KITT4SME Trifold-brochure: (2021) https://kitt4sme.eu/wp-content/uploads/2021/03/Trifold-
brochure-KITT4SME.pdf.pdf. Accessed 12 July 2023
21. Osterwalder, A., Pigneur, Y.: Business Model Generation: a Handbook for Visionaries, Game
Changers, and Challengers. Wiley, New Jersey (2010)
22. Wang, Y., Tang, J., Jin, Q., Ma, J.: On studying business models in mobile social networks
based on two-sided market (TSM). J. Supercomput. 70, 1297–1317 (2014). https://doi.org/
10.1007/s11227-014-1228-4
23. Zhao, Y., von Delft, S., Morgan-Thomas, A., Buck, T.: The evolution of platform business
models: exploring competitive battles in the world of platforms. Long Range Plan. 53(4),
101892 (2020). https://doi.org/10.1016/j.lrp.2019.101892
120 K. Ejsmont et al.
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Self-Reconfiguration for Smart
Manufacturing Based on Artificial
Intelligence: A Review and Case Study
1 Introduction
2 Reconfiguration in Manufacturing
Although FMSs can deal with the market exigence for new products or modifica-
tions of existing products, they cannot efficiently adjust their production capacity.
This means that if a manufacturing system was designed to produce a maximum
number of products annually and, after 2 years, the market demand for the product is
reduced to half, the factory will be idle 50% of the time, creating a big financial loss.
124 Y. J. Cruz et al.
On the other hand, if the market demand for the product surpasses design capability
and the system is unable to handle it, the financial loss can be even greater [8]. To
handle such scenarios, during the 1990s, a new type of manufacturing system known
as reconfigurable manufacturing system (RMS) was introduced. RMSs adhere to
the typical goals of production systems: to produce with high quality and low cost.
However, additionally, they also aim to respond quickly to market demand, allowing
for changes in production capacity. In other words, they strive to provide the
capability and functionality required at the right time [3]. This goal is achieved by
enabling the addition or removal of components from production lines on demand.
Design principles such as modularity, integrability, and open architecture control
systems started to take more significance with the emergence of RMSs, given the
relevance of dynamic equipment interconnection in these systems [9]. Consider-
ing their advantages, RMSs have been applied to the manufacturing of medical
equipment [10], automobiles [11], food and beverage [12], and so on. Because they
require less investments in equipment and infrastructure, these systems often offer
a more cost-effective alternative to FMSs.
Although these systems can adapt to changing production requirements, the
reconfiguration decisions are usually made or supervised by a human, which means
the systems cannot autonomously reconfigure themselves. This gives more control
to the plant supervisor or operator, but the downside is that it limits the response
speed.
3 Current Approaches
Computer simulation is a particularly valuable tool for the design and optimization
of self-reconfigurable manufacturing systems. In this context, these tools aim to
enhance the system’s responsiveness to changes in production requirements. The
recent increase in computational capacities has enabled the testing of various
configurations and scenarios before their actual implementation [19]. Currently,
commercial applications such as AutoMod, FlexSim, Arena, Simio, and AnyLogic,
among others, allow to create high-fidelity simulations of industrial processes [20],
that even include three-dimensional recreations of factories for use in augment-
ed/virtual reality applications. Computer simulation becomes a powerful tool when
integrated with the production process it represents. Based on this idea, digital twins
have gained significant attention in both industry and academia [21]. Digital twins
enable real-time data integration from the production process into the simulation,
replicating the actual production environment. By evaluating different options and
identifying the optimal configuration for the new scenarios, digital twins provide
feedback to the production process, facilitating real-time modifications.
126 Y. J. Cruz et al.
Fuzzy logic is a mathematical framework that can be used to model and reason with
imprecise or uncertain data. This capability makes fuzzy logic particularly useful
in situations where the system may not have access to precise data or where the
data may be subject to noise or other sources of uncertainty. In the context of self-
reconfiguration, fuzzy systems can be used to model the behavior of the physical
processes and make decisions about how to reconfigure them based on imprecise
data. For instance, it is often very complex to assign a precise value to indicators
such as expected market demand, product quality, or energy consumption [22].
These variables can be assigned to fuzzy membership functions and then, following
predefined rules, combined using fuzzy operators to determine how the production
system should be optimally reconfigured depending on the available data.
Data-driven methods deal with the collection and analysis of data, the creation of
models, and their use for decision-making. This approach is extensively applied
when historical data of the production process is available. By using data ana-
lytics, it is possible to identify bottlenecks or the inefficient use of assets in the
production process. Also, data-driven methods make extensive use of machine
learning algorithms for modeling the production process behavior [23]. Machine
learning methods can be trained with datasets containing a large number of features
and samples, learning to identify correlations, patterns, and anomalies that are
beyond human perception [24]. Moreover, by collecting new data of the production
process, machine learning models can be retrained or fine-tuned to improve their
performance over time. Once the machine learning model has been trained with
production data, it can be used as an objective function of an optimization algorithm
to make decisions about how to reconfigure the manufacturing process to optimize
desired indicators.
situation. Once the slotting irregularities are corrected, the system continues with
the normal workflow.
In some cases, the slotting process may cause damage to the panels. This
can happen when working with new materials or previously unverified machining
configurations. In those cases, the visual inspection system should detect that the
panel is damaged and it should be sent directly to a stack of damaged parts. Figure
3 shows this situation.
Making accurate decisions about the process workflow depending on the quality
of products, specifically on the result of the slotting process, has a direct impact
on the productivity of the pilot line. For instance, in cases where a panel is
damaged during slotting, it is crucial to remove it from the production line to
prevent unnecessary time and resources from being spent on machining it during
the engraving stage. To achieve this, the presence of a reliable visual inspection
system becomes essential. Although a deep learning classifier could be used for
this task, one drawback is that it is very hard to understand how the decision is
made. For this reason, it is proposed a deep learning segmentation model, whose
function is to separate the desired areas of the images from the unwanted regions.
Self-Reconfiguration for Smart Manufacturing Based on Artificial Intelligence:. . . 129
The segmentation model developed for application in the pilot line intends to
separate the side surface of the panel from other elements within an image, allowing
for later decisions on the panel quality and modifications of the process workflow.
This model is based on a U-net architecture, which consists of an encoder path
that gradually downsamples the input image and a corresponding decoder path that
upsamples the feature maps to produce a segmentation map of the same size as the
input image. This network also includes skip connections between the encoder and
decoder paths that allow to retain and fuse both high-level and low-level features,
facilitating accurate segmentation and object localization [26].
A dataset containing 490 images with their corresponding masks was prepared
for training and evaluating the model. The image dataset was split into three subsets:
training (70% of the data), validation (15% of the data), and testing (15% of the
data). In this case, the validation subset serves the objective of facilitating early
stopping during training. This means that if the model’s performance evaluated on
the validation subset fails to improve after a predetermined number of epochs, the
training process is halted. By employing this technique, overfitting can be effectively
mitigated and the training time can be significantly reduced.
A second version of the dataset was prepared by applying data augmentation
to the training set while keeping the validation and test sets unchanged. The
dataset was augmented using four transformations: horizontal flip, coarse dropout,
random brightness, and random contrast. This helps increase the number of training
examples, improving the model’s prediction capability and making it more robust to
noise. Using the two versions of the dataset, two models with the same architecture
were trained. Table 1 presents the output of the two models for three examples taken
from the test set. As can be observed, the predictions obtained with the model trained
on the augmented dataset are significantly better than those obtained with the model
130
3
Y. J. Cruz et al.
Self-Reconfiguration for Smart Manufacturing Based on Artificial Intelligence:. . . 131
Fig. 4 Squared contours detection for a compliant panel, a panel without slot, and a damaged
panel, respectively
trained on the original dataset. This is also confirmed by the values obtained in
several metrics, which are shown in Table 2.
After the image is segmented by the deep learning model, a second algorithm
is used. Here, a convex hull is adjusted to each separate contour in the segmented
image. Then, a polygonal curve is generated for each convex hull with a precision
smaller than 1.5% of the perimeter of the segmented contour. Finally, if the
polygonal curve has four sides, it is drawn over the original image. After this
procedure, if two rectangles were drawn over the image it is assumed that the
slotting was correct and the panel did not suffer any significative damage; thus, it
can be sent to the next stage of the line. On the other hand, if only one rectangle was
drawn, it is assumed that the slotting was not carried out or the panel was damaged
during this process. Figure 4 shows the results obtained for illustrative cases of a
compliant panel, a panel with missing slots, and a damaged panel, respectively. If
only one rectangle was drawn, depending on its size and location, the panel will be
sent to the slotting stage again or removed from the line. This method was applied
to the test set images and in all the cases the output produced matched the expected
output.
As outlined in the previous sections, the working conditions of the pilot line are
subject to rapid variations. To effectively address these variations and generate
132 Y. J. Cruz et al.
Machine learning aims to create accurate and reliable models capable of identifying
complex patterns in data. The creation and exploitation of these models is typically
achieved through a series of steps that involve preparing the dataset, transforming
the data to enhance its quality and relevance, selecting and training an appropriate
machine learning model, evaluating the model’s performance, and deploying the
model in a real-world setting. Figure 5 depicts these steps. By following this
workflow, machine learning practitioners can build models that harness the power
of data-driven learning, enabling them to effectively derive meaningful insights and
make accurate predictions in practical applications.
Data Preprocessing
Data preprocessing is the initial step in the creation of a machine learning system.
The data to be used may have a variety of sources and formats, thus it should be
prepared before being used by any algorithm. If data are originated from different
sources, it must be merged into a single dataset. Furthermore, most methods are
Self-Reconfiguration for Smart Manufacturing Based on Artificial Intelligence:. . . 133
not designed to work with missing data, so it is very common to remove samples
with missing information. Preprocessing may also include filtering data to remove
noise, which can result later in more robust models. In this stage, the data may be
transformed to a format that is suitable for analysis, which can include operations
such as normalization, bucketizing, and encoding. Finally, one common operation
carried out in this stage is splitting. This refers to the partition of the dataset into
two subsets, which will be used for training and evaluation purposes. Additionally, a
third subset can be created if it is planned to carry out a hyperparameter optimization
or neural architecture search over the model.
Feature Engineering
The goal of the feature engineering stage is to convert raw data into relevant features
that contain the necessary information to create high-quality models. One of the
most interesting techniques that can be used in this stage is feature selection.
Feature selection aims to determine which features are the best predictors for a
certain output variable. Then, when these features are selected, they can be extracted
from the original dataset to build a lower dimensional dataset, allowing to build
more compact models with better generalization ability and reduced computational
time [27, 28]. Typically, for problems with numerical input and output variables,
Pearson’s [29] or Spearman’s correlation coefficients [30] are used. If the input is
numerical but the output is categorical, then the analysis of variance (ANOVA) [31]
or Kendall’s rank coefficient [32] are employed. Other situations may require the
use Chi-squared test or mutual information measure [33].
Other techniques that can be applied in the feature engineering stage include
feature creation and dimensionality reduction. Feature creation implies creating new
features either by combining the existing ones or by using domain knowledge [34].
On the other hand, dimensionality reduction techniques such as principal component
analysis (PCA) or t-distributed stochastic neighbor embedding (t-SNE) algorithms
are used to map the current data in lower dimensional space while retaining as much
information as possible [35].
Model Selection
The model selection step implies the creation, training, and evaluation of different
types of models to, in the end, select the most suitable for the current situation.
This practice is carried out since it does not exist a methodology for determining a
priori which algorithm is better for solving a problem [36]. Therefore, the most
adequate model may vary from one application to another as in the following
cases: long short-term memory network (LSTM) [37], multilayer perceptron (MLP)
[38], support vector regression (SVR) [39], Gaussian process regression (GPR)
[40], convolutional neural network (CNN) [41], gradient boosted trees (GBT)
[42]. The number and types of models to explore in this stage will depend on
134 Y. J. Cruz et al.
the characteristics of the problem and the available computational resources. The
selection of the model is carried out taking into consideration one or more metrics.
For regression problems is common to rely on the coefficient of determination (R2 ),
mean squared error (MSE), and mean absolute percentage error (MAPE), among
other metrics [43]. On the other hand, for classification problems, typical metrics
are accuracy, recall, precision, F1-score,and so on.
Optionally, this stage can also include hyperparameter optimization. Hyperpa-
rameters determine a model’s behavior during training and, in some cases, also how
its internal structure is built. They are set before a model is trained and cannot
be modified during training. The selection of these values can greatly affect a
model’s performance. However, finding an optimal or near-optimal combination of
hyperparameters is not a trivial task and, usually, it is computationally intensive. The
most commonly used techniques for this task include grid search, random search,
Bayesian optimization, and so on.
Once a model has been created for representing a process, it can be used for
optimizing it. Assuming the model exhibits robust predictive capabilities and the
constraints are accurately defined, various input values can be evaluated in the model
to determine how the system would respond, eliminating the need for conducting
exhaustive tests on the actual system. In other words, the model created can be
embedded as the objective function of an optimization algorithm for finding the
input values that would make the production process work in a desired regime. In
this context, popular strategies such as particle swarm optimization [44], simulated
annealing [45], evolutionary computation [46], and Nelder-Mead [47], among
others, are commonly employed.
First, in the data preprocessing step, the dataset is inspected searching for missing
values. If any are found, the corresponding sample is eliminated. Next, the features’
values are standardized and the dataset is divided into training and validation sets.
In this case, hyperparameter optimization was not implemented for making the
methodology applicable in scenarios with low computational resources. For this
reason, a test set is not required. Following that, feature selection is carried out
by computing the Pearson’s correlation coefficient (r) individually between each
feature and the output variable on the training set, using the following equation:
n
− x) (yi − y)
i=1 (xi
r =
2 n
.
n
i=1 (xi − x) i=1 (yi − y)
2
where n is the number of samples, xi represents the value of the i-th sample of
feature x, yi represents the value of the i-th sample of the output variable, and .x and
.y represent the mean of the respective variables.
they are evaluated on the validation set. The metric used for comparison was the
coefficient of determination (R2 ). After this process is finished, the model that
produced the best result is selected. The model selected during the application of
the methodology to the pilot line was MLP with R2 = 0.963 during validation. The
R2 value for the remaining candidate models was 0.958 for GPR, 0.955 for CNN,
and 0.947 for SVR. One of the enablers of these results was the feature selection
process, which allowed to retain the relevant predictors.
Finally, an optimization method is applied for determining the most favorable
parametrization of the production process to minimize or maximize the desired
KPI using the selected model as the objective function. In this case, the goal is to
maximize throughput. The optimization is carried out using random search, which
is a simple, low-complexity, and straightforward optimization method [53]. This
method can be applied to optimizing diverse types of functions, even those that are
not continuous or differentiable. It has been proven that random search is asymptot-
ically complete, meaning that it converges to the global minimum/maximum with
probability one after indefinitely run-time computation and, for this reason, it has
been applied for solving many complex problems [54]. One aspect to consider
before executing the optimization is that the feasible range of the parameters must
be carefully decided to prevent the result of the optimization from being invalid. In
the case analyzed, where the objective is to maximize the throughput of the pilot
line, the obvious choice is to make the assets work at the maximum speed within
the recommended ranges. To evaluate if the proposed methodology was capable
of inferring this parametrization, during the preparation of the dataset the samples
where all the assets were parametrized with the maximum speed were intentionally
eliminated. As desired, the result of the methodology was a parametrization where
all the assets were set to the maximum speed, yielding an expected throughput value
of 163.37 panels per hour, which represents an expected improvement of 55.1% with
respect to the higher throughput value present in the dataset. It is noticeable that the
higher throughput value of the samples that were intentionally eliminated from the
dataset is 158.52. The reason why the proposed methodology slightly overestimates
this value is that the model is not perfectly accurate.
Self-Reconfiguration for Smart Manufacturing Based on Artificial Intelligence:. . . 137
Once the parametrization of the assets has been determined by the AutoML
methodology to meet a desired KPI performance, it is important to ensure that the
system will continue to work as desired. Unfortunately, some situations may prevent
the system from functioning as intended. For instance, a degradation in one of the
assets may result in a slower operation, reducing the productivity of the entire line.
For such cases, a fuzzy logic-based reconfigurator is developed. The intuition behind
this component is that if the behavior of some assets varies from their expected
performance, the reconfigurator can modify the parameters of the assets to make
them work in the desired regime again, as long as the modification of the parameters
is within a predefined safety range. Additionally, if the deviation from the expected
performance is significant, the component should be able to detect it and inform the
specialists that a problem needs to be addressed.
The proposed reconfigurator has two inputs and generates three outputs using the
Mamdani inference method [55]. These variables are generic, so the reconfigurator
can use them without any modification to try to keep each asset’s throughput level
constant. The first input is the deviation from nominal production time (ΔT) and
its safety range was defined as ±50% of the nominal production time. The second
input is the change in the trend of the deviation from nominal production time (ΔT2 )
and its safety range was defined as ±20% of the nominal production time. There is
an instance of these two variables for each asset in the line and they are updated
whenever a panel is processed. These values are normalized in the interval [−1, 1]
before being used by the reconfigurator. Figure 7 presents the membership functions
defined for the two inputs.
On the other hand, the first output is the operation that the reconfigurator must
apply to the current asset’s working speed (Reco1). If the operation is Increase or
Decrease, the values in the interval [−1, 1] are denormalized to a range comprising
±50% of the nominal asset speed. The second output represents the timing when
the modifications should be applied (Reco2), and the third output represents the
operation mode (Reco3), which specifies if the previous reconfigurator outputs
5 Conclusions
Acknowledgments This work was partially supported by the H2020 project “platform-enabled
KITs of arTificial intelligence FOR an easy uptake by SMEs (KITT4SME),” grant ID 952119.
The work is also funded by the project “Self-reconfiguration for Industrial Cyber-Physical
Systems based on digital twins and Artificial Intelligence. Methods and application in Indus-
try 4.0 pilot line,” Spain, grant ID PID2021-127763OB-100, and supported by MICINN and
NextGenerationEU/PRTR. This result is also part of the TED2021-131921A-I00 project, funded
by MCIN/AEI/10.13039/501100011033 and by the European Union “NextGenerationEU/PRTR.”
References
1. Hees, A., Reinhart, G.: Approach for production planning in reconfigurable manufacturing
systems. Proc. CIRP. 33, 70–75 (2015). https://doi.org/10.1016/j.procir.2015.06.014
2. Yin, Y., Stecke, K.E., Li, D.: The evolution of production systems from Industry 2.0
through Industry 4.0. Int. J. Prod. Res. 56, 848–861 (2018). https://doi.org/10.1080/
00207543.2017.1403664
3. Mehrabi, M.G., Ulsoy, A.G., Koren, Y., Heytler, P.: Trends and perspectives in flexible and
reconfigurable manufacturing systems. J. Intell. Manuf. 13, 135–146 (2002). https://doi.org/
10.1023/A:1014536330551
4. Cronin, C., Conway, A., Walsh, J.: Flexible manufacturing systems using IIoT in the automo-
tive sector. Proc. Manuf. 38, 1652–1659 (2019). https://doi.org/10.1016/j.promfg.2020.01.119
5. Parhi, S., Srivastava, S.C.: Responsiveness of decision-making approaches towards the per-
formance of FMS. In: 2017 International Conference of Electronics, Communication and
Aerospace Technology (ICECA), pp. 276–281 (2017)
6. Bennett, D., Forrester, P., Hassard, J.: Market-driven strategies and the design of flexible
production systems: evidence from the electronics industry. Int. J. Oper. Prod. Manag. 12,
25–37 (1992). https://doi.org/10.1108/01443579210009032
7. Singh, R.K., Khilwani, N., Tiwari, M.K.: Justification for the selection of a reconfigurable
manufacturing system: a fuzzy analytical hierarchy based approach. Int. J. Prod. Res. 45, 3165–
3190 (2007). https://doi.org/10.1080/00207540600844043
8. Koren, Y.: The emergence of reconfigurable manufacturing systems (RMSs) BT - reconfig-
urable manufacturing systems: from design to implementation. In: Benyoucef, L. (ed.) , pp.
1–9. Springer International Publishing, Cham (2020)
9. Koren, Y., Shpitalni, M.: Design of reconfigurable manufacturing systems. J. Manuf. Syst. 29,
130–141 (2010). https://doi.org/10.1016/j.jmsy.2011.01.001
10. Epureanu, B.I., Li, X., Nassehi, A., Koren, Y.: An agile production network enabled by recon-
figurable manufacturing systems. CIRP Ann. 70, 403–406 (2021). https://doi.org/10.1016/
j.cirp.2021.04.085
11. Koren, Y., Gu, X., Guo, W.: Reconfigurable manufacturing systems: principles, design, and
future trends. Front. Mech. Eng. 13, 121–136 (2018). https://doi.org/10.1007/s11465-018-
0483-0
12. Gould, O., Colwill, J.: A framework for material flow assessment in manufacturing systems. J.
Ind. Prod. Eng. 32, 55–66 (2015). https://doi.org/10.1080/21681015.2014.1000403
13. Shin, K.-Y., Park, H.-C.: Smart manufacturing systems engineering for designing smart
product-quality monitoring system in the Industry 4.0. In: 2019 19th International Conference
on Control, Automation and Systems (ICCAS), pp. 1693–1698 (2019)
14. Arıcı, M., Kara, T.: Robust adaptive fault tolerant control for a process with actuator faults. J.
Process Control. 92, 169–184 (2020). https://doi.org/10.1016/j.jprocont.2020.05.005
15. Ghofrani, J., Deutschmann, B., Soorati, M.D., et al.: Cognitive production systems: a mapping
study. In: 2020 IEEE 18th International Conference on Industrial Informatics (INDIN), pp.
15–22 (2020)
142 Y. J. Cruz et al.
16. Scholz, S., Mueller, T., Plasch, M., et al.: A modular flexible scalable and reconfigurable system
for manufacturing of microsystems based on additive manufacturing and e-printing. Robot.
Comput. Integr. Manuf. 40, 14–23 (2016). https://doi.org/10.1016/j.rcim.2015.12.006
17. Cedeno-Campos, V.M., Trodden, P.A., Dodd, T.J., Heley, J.: Highly flexible self-reconfigurable
systems for rapid layout formation to offer manufacturing services. In: 2013 IEEE International
Conference on Systems, Man, and Cybernetics, pp. 4819–4824 (2013)
18. Lee, S., Ryu, K.: Development of the architecture and reconfiguration methods for the
smart, self-reconfigurable manufacturing system. Appl. Sci. 12 (2022). https://doi.org/10.3390/
app12105172
19. Mourtzis, D.: Simulation in the design and operation of manufacturing systems: state of
the art and new trends. Int. J. Prod. Res. 58, 1927–1949 (2020). https://doi.org/10.1080/
00207543.2019.1636321
20. dos Santos, C.H., Montevechi, J.A.B., de Queiroz, J.A., et al.: Decision support in productive
processes through DES and ABS in the Digital Twin era: a systematic literature review. Int. J.
Prod. Res. 60, 2662–2681 (2022). https://doi.org/10.1080/00207543.2021.1898691
21. Guo, H., Zhu, Y., Zhang, Y., et al.: A digital twin-based layout optimization method for discrete
manufacturing workshop. Int. J. Adv. Manuf. Technol. 112, 1307–1318 (2021). https://doi.org/
10.1007/s00170-020-06568-0
22. Abdi, M.R., Labib, A.W.: Feasibility study of the tactical design justification for reconfigurable
manufacturing systems using the fuzzy analytical hierarchical process. Int. J. Prod. Res. 42,
3055–3076 (2004). https://doi.org/10.1080/00207540410001696041
23. Lee, S., Kurniadi, K.A., Shin, M., Ryu, K.: Development of goal model mechanism for self-
reconfigurable manufacturing systems in the mold industry. Proc. Manuf. 51, 1275–1282
(2020). https://doi.org/10.1016/j.promfg.2020.10.178
24. Panetto, H., Iung, B., Ivanov, D., et al.: Challenges for the cyber-physical manufacturing
enterprises of the future. Annu. Rev. Control. 47, 200–213 (2019). https://doi.org/10.1016/
j.arcontrol.2019.02.002
25. Schwung, D., Modali, M., Schwung, A.: Self -optimization in smart production systems using
distributed reinforcement learning. In: 2019 IEEE International Conference on Systems, Man
and Cybernetics (SMC), pp. 4063–4068 (2019)
26. Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image
segmentation BT - medical image computing and computer-assisted intervention – MICCAI
2015. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) , pp. 234–241. Springer
International Publishing, Cham (2015)
27. Al-Tashi, Q., Abdulkadir, S.J., Rais, H.M., et al.: Approaches to multi-objective feature
selection: a systematic literature review. IEEE Access. 8, 125076–125096 (2020). https://
doi.org/10.1109/ACCESS.2020.3007291
28. Solorio-Fernández, S., Carrasco-Ochoa, J.A., Martínez-Trinidad, J.F.: A review of unsuper-
vised feature selection methods. Artif. Intell. Rev. 53, 907–948 (2020). https://doi.org/10.1007/
s10462-019-09682-y
29. Jebli, I., Belouadha, F.-Z., Kabbaj, M.I., Tilioua, A.: Prediction of solar energy guided by
Pearson correlation using machine learning. Energy. 224, 120109 (2021). https://doi.org/
10.1016/j.energy.2021.120109
30. González, J., Ortega, J., Damas, M., et al.: A new multi-objective wrapper method for feature
selection – accuracy and stability analysis for BCI. Neurocomputing. 333, 407–418 (2019).
https://doi.org/10.1016/j.neucom.2019.01.017
31. Alassaf, M., Qamar, A.M.: Improving sentiment analysis of Arabic Tweets by One-way
ANOVA. J. King Saud Univ. Comput. Inf. Sci. 34, 2849–2859 (2022). https://doi.org/10.1016/
j.jksuci.2020.10.023
32. Urkullu, A., Pérez, A., Calvo, B.: Statistical model for reproducibility in ranking-based feature
selection. Knowl. Inf. Syst. 63, 379–410 (2021). https://doi.org/10.1007/s10115-020-01519-3
33. Bahassine, S., Madani, A., Al-Sarem, M., Kissi, M.: Feature selection using an improved Chi-
square for Arabic text classification. J. King Saud Univ. Comput. Inf. Sci. 32, 225–231 (2020).
https://doi.org/10.1016/j.jksuci.2018.05.010
Self-Reconfiguration for Smart Manufacturing Based on Artificial Intelligence:. . . 143
34. Lu, Z., Si, S., He, K., et al.: Prediction of Mg alloy corrosion based on machine learning
models. Adv. Mater. Sci. Eng. 2022, 9597155 (2022). https://doi.org/10.1155/2022/9597155
35. Anowar, F., Sadaoui, S., Selim, B.: Conceptual and empirical comparison of dimensionality
reduction algorithms (PCA, KPCA, LDA, MDS, SVD, LLE, ISOMAP, LE, ICA, t-SNE).
Comput. Sci. Rev. 40, 100378 (2021). https://doi.org/10.1016/j.cosrev.2021.100378
36. Cruz, Y.J., Rivas, M., Quiza, R., et al.: A two-step machine learning approach for dynamic
model selection: a case study on a micro milling process. Comput. Ind. 143, 103764 (2022).
https://doi.org/10.1016/j.compind.2022.103764
37. Castano, F., Cruz, Y.J., Villalonga, A., Haber, R.E.: Data-driven insights on time-to-failure
of electromechanical manufacturing devices: a procedure and case study. IEEE Trans. Ind.
Inform, 1–11 (2022). https://doi.org/10.1109/TII.2022.3216629
38. Mezzogori, D., Romagnoli, G., Zammori, F.: Defining accurate delivery dates in make to order
job-shops managed by workload control. Flex. Serv. Manuf. J. 33, 956–991 (2021). https://
doi.org/10.1007/s10696-020-09396-2
39. Luo, J., Hong, T., Gao, Z., Fang, S.-C.: A robust support vector regression model for
electric load forecasting. Int. J. Forecast. 39, 1005–1020 (2023). https://doi.org/10.1016/
j.ijforecast.2022.04.001
40. Pai, K.N., Prasad, V., Rajendran, A.: Experimentally validated machine learning frame-
works for accelerated prediction of cyclic steady state and optimization of pressure swing
adsorption processes. Sep. Purif. Technol. 241, 116651 (2020). https://doi.org/10.1016/
j.seppur.2020.116651
41. Cruz, Y.J., Rivas, M., Quiza, R., et al.: Computer vision system for welding inspection of
liquefied petroleum gas pressure vessels based on combined digital image processing and deep
learning techniques. Sensors. 20 (2020). https://doi.org/10.3390/s20164505
42. Pan, Y., Chen, S., Qiao, F., et al.: Estimation of real-driving emissions for buses fueled with
liquefied natural gas based on gradient boosted regression trees. Sci. Total Environ. 660, 741–
750 (2019). https://doi.org/10.1016/j.scitotenv.2019.01.054
43. Chicco, D., Warrens, M.J., Jurman, G.: The coefficient of determination R-squared is more
informative than SMAPE, MAE, MAPE, MSE and RMSE in regression analysis evaluation.
PeerJ. Comput. Sci. 7, e623 (2021). https://doi.org/10.7717/peerj-cs.623
44. Eltamaly, A.M.: A novel strategy for optimal PSO control parameters determination for PV
energy systems. Sustainability. 13 (2021). https://doi.org/10.3390/su13021008
45. Karagul, K., Sahin, Y., Aydemir, E., Oral, A.: A simulated annealing algorithm based solution
method for a green vehicle routing problem with fuel consumption BT - lean and green supply
chain management: optimization models and algorithms. In: Weber, G.-W., Huber, S. (eds.)
Paksoy T, pp. 161–187. Springer International Publishing, Cham (2019)
46. Cruz, Y.J., Rivas, M., Quiza, R., et al.: Ensemble of convolutional neural networks based on
an evolutionary algorithm applied to an industrial welding process. Comput. Ind. 133, 103530
(2021). https://doi.org/10.1016/j.compind.2021.103530
47. Yildiz, A.R.: A novel hybrid whale–Nelder–Mead algorithm for optimization of design and
manufacturing problems. Int. J. Adv. Manuf. Technol. 105, 5091–5104 (2019). https://doi.org/
10.1007/s00170-019-04532-1
48. Gao, X., Li, X., Zhao, B., et al.: Short-term electricity load forecasting model based on EMD-
GRU with feature selection. Energies. 12 (2019). https://doi.org/10.3390/en12061140
49. Mu, C., Xing, Q., Zhai, Y.: Psychometric properties of the Chinese version of the Hypo-
glycemia Fear SurveyII for patients with type 2 diabetes mellitus in a Chinese metropolis.
PLoS One. 15, e0229562 (2020). https://doi.org/10.1371/journal.pone.0229562
50. Schaefer, J.L., Nara, E.O.B., Siluk, J.C.M., et al.: Competitiveness metrics for small
and medium-sized enterprises through multi-criteria decision making methods and neural
networks. Int. J. Proc. Manag. Benchmark. 12, 184–207 (2022). https://doi.org/10.1504/
IJPMB.2022.121599
51. Manimuthu, A., Venkatesh, V.G., Shi, Y., et al.: Design and development of automobile
assembly model using federated artificial intelligence with smart contract. Int. J. Prod. Res.
60, 111–135 (2022). https://doi.org/10.1080/00207543.2021.1988750
144 Y. J. Cruz et al.
52. Zagumennov, F., Bystrov, A., Radaykin, A.: In-firm planning and business processes manage-
ment using deep neural network. GATR J. Bus. Econ. Rev. 6, 203–211 (2021). https://doi.org/
10.35609/jber.2021.6.3(4)
53. Ozbey, N., Yeroglu, C., Alagoz, B.B., et al.: 2DOF multi-objective optimal tuning of
disturbance reject fractional order PIDA controllers according to improved consensus ori-
ented random search method. J. Adv. Res. 25, 159–170 (2020). https://doi.org/10.1016/
j.jare.2020.03.008
54. Do, B., Ohsaki, M.: A random search for discrete robust design optimization of linear-elastic
steel frames under interval parametric uncertainty. Comput. Struct. 249, 106506 (2021). https:/
/doi.org/10.1016/j.compstruc.2021.106506
55. Mamdani, E.H.: Application of fuzzy algorithms for control of simple dynamic plant. Proc.
Inst. Electr. Eng. 121, 1585–1588 (1974). https://doi.org/10.1049/piee.1974.0328
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Part II
Multi-agent Systems and AI-Based Digital
Twins for Manufacturing Applications
Digital-Twin-Enabled Framework for
Training and Deploying AI Agents for
Production Scheduling
1 Introduction
physical counterpart [6]. DT have been used in many different domains, including
healthcare, urban planning and manufacturing [7]. In a variety of industries, the
combination of real-time data and the digital replica of the physical environment
promotes the decision-making and allow the industries to continuously improve
their performance [8, 9].
Asset Administration Shell (AAS) technology was introduced within the Ref-
erence Architecture Model Industry 4.0 (RAMI4. 0) [10] and has become a
ground-breaking idea in manufacturing of how assets are managed and used [11].
AAS are standardized models that allow industries to combine the physical assets
with their digital counterparts (i.e., machines, production systems, or tools), where
AASs provide a framework to control and monitor the physical assets. AI scheduling
agents have been used coupled with AAS concept in the literature [9, 12, 13].
Additionally, AI scheduling agents are intelligent autonomous systems that take as
an input production system information to plan resource allocation tasks [14]. The
AI scheduling agents play an important role due to the fact that they can generate
a real-time efficient schedule. Coupled with the AAS and DT technologies, AI
scheduling agents can be used for the real-time decision-making or for predictions
[14, 15].
Multi-agent system (MAS) are systems that are used to compose many
autonomous agents, where these agents interact to each other [15]. MAS provides
decentralized and collaborative decision-making, where it allows the collaboration
of different agents. Each agent in the MAS has some capabilities, decision-making
abilities and takes decisions. MAS is used in order to solve complex problems where
one agent is impossible to solve. The idea of dived and conquer is used to divide
the problem into subproblems, where each agent solves a subproblem, and provides
solutions that are adaptable, robust, and able to handle real-time uncertainties. MAS
is also combined with DT, AAS, and AI scheduling agents. In a further analysis,
AASs can enable interaction between different agents, where the agents can be
models as different assets.
The contributions of this work are the use of the DT in order to accurately
simulate and validate the AI agents that have been developed, as well as training
some of the agents. Moreover, the use of the AAS technology to exchange data
between the DT and AI agents within the MAS and finally the developed AI
scheduling agents that were developed and modeled based on the bicycle industry’s
requirements and challenges.
This chapter is organized in four sections, where the first section introduces
the concepts of digital twins (DT), Asset Administration Shells (AAS), scheduling
problem, and artificial intelligence (AI) applications. In second section are discussed
related works. In the third section is explained the proposed MAS framework and
explain the optimization tools that have been developed. In the fourth section is
described the case study that the proposed framework is implemented. Finally, the
last section is the conclusion of this work, where some future works are discussed.
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 149
2 Related Works
real processes and assets to their digital representative ones can be characterized as
twinning.
One of the main technologies used, in order to realize most of DT implementation
approaches, is simulation [24]. As already mentioned, the idea of DT is to build a
virtual version of a real system. This replica can be used to simulate and forecast
how the physical system will respond to certain situations. Thus, one of the best
methods to construct a virtual representation of the physical system seems to be
simulation, which enables engineers to test and improve the system before it is built,
lowering costs and increasing efficiency. Digital twin and simulation technology
are being used more and more frequently in sectors such as manufacturing and
aerospace, exhibiting their ability to completely change how complex systems are
created and optimized [25].
Furthermore, digital twin implementation methods can support decision-making
related to the scheduling task for a production system with potential uncertainties
[26]. A crucial aspect for the development of a digital twin is the achievement of a
high level of standardization and interoperability with systems outside the digital
environment. The digital twin simulates some of the behaviors of the physical
environment, and thus requires some kind of seamless information exchange with
the physical entities and the information they provide. OPC UA is a standard that
can provide standardization in the data exchange between the digital twin and
production hardware, achieving real-time monitoring and control, interconnectivity,
security, access control, while also data modelling and semantics [27].
The Asset Administration Shell (AAS) could be also used in order to standardize
the description and management of assets. The digital twin technology can exchange
information with the asset via the establishment of a common information language
[28]. In addition, the AAS and OPC UA are complementary standards that can be
both used to define the framework and protocol for that communication [29]; it
is worth noticing that AAS is a collection of standards, namely IEC 62875, DIN
SPEC 91345, IEC 62541 (OPC UA), and RAMI 4.0. In cases where the digital twin
composes a higher level system such as a production line, a station, or a production
system, it is usually composed of multiple assets and thus AAS models. From the
digital twin side, the AAS can be the middleware for exchanging information with
the assets or managing their behavior. It is important, however, to highlight that there
is no standard way for describing an asset using the AAS; although the metamodel
will always be the same, there is a freedom to the different submodels and submodel
elements that will be selected in order to describe any different asset. It is thus usual
to exploit additional information modelling standards or framework to define the
specific components and information structures within the AAS metamodel – e.g.,
ISA-88, ISO 22400, and ISA-95.
Digital twin is not only about simulating the environment but also taking
decisions over the next actions, which can then be used on the physical environment.
Simulation on its own cannot address this issue, and AI agents is a way that this
challenge can be solved. Multi-agent systems are preferred over centralized software
components in cases where the problem is hard enough to be solved from a mono-
lithic software component. It is a decentralized approach that breaks the problem
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 151
into subproblems and each agent has access only to the subproblems compatible
with its skills. In the case of production scheduling, this is a useful approach as it
enables different types of scheduling problems being solved by different AI methods
based on which method best satisfies the requirements. AI is a broad term and in
scheduling in particular the most common methods are heuristic, metaheuristic,
mathematical optimization, machine learning, reinforcement learning, and policy-
making.
Mathematical optimization, also referred as mathematical programming, is an
optimization model consisted of input sets and parameters, decision variables,
constraints/expressions, and the objective function. Based on the constraints and
the objectives, the model may be classified as linear, nonlinear, convex, integer, and
mixed-integer problem, with different type of algorithms to optimize the objectives.
As such, as important as the model, the algorithm that is used in order to find a
both feasible and accurate solution is also crucial for the quality of the solution. The
algorithms may be exact or heuristic-based, while metaheuristic methods are also
popular for various optimization problems.
Heuristics have been deployed to solve various production scheduling opti-
mization problems. A combination of constructive heuristics and iterated greedy
algorithm was used to solve the distributed blocking flowshop scheduling problem
(DBFSP) and lead to makespan minimization [30]. Montiel et al. (2017) proposed
an approach for the stochastic optimization of mine production schedules with the
use of heuristics, implementing iterative improvement by swapping periods and
destinations of the mining blocks to create the final solution [31]. Heuristics can also
be successfully deployed to optimize the scheduling task, aiming at reducing total
energy consumption [32]. Jélvez et al. (2020) worked for a new hybrid heuristic
algorithm to solve the Precedence Constrained Production Scheduling Problem
(PCPSP) for an open-pit mining industry [33].
Heuristic and metaheuristic algorithms focus on an intelligent search along the
solution space, which does not ensure the quality of the solution, and in complex
optimization problems require flexible time delays. Deep learning methods, on the
other hand, do not depend on searching the solution space, but rather predicting
the solution based on patterns from historical information. Although in most cases
the results are guaranteed to be fast, it is not necessarily of high quality. In
reality it depends on the deep learning model that was used, the dataset quality
and quantity. In some cases, there is also a dataset shortage which makes the
problem even more difficult to solve. In practice, researchers may address this
problem via the utilization of a system digital replica which is able to simulate
the behaviors of the actual system in a realistic manner. This can support the
development of either reinforcement learning methods that use the simulation as
a reward retrieval plugin or for extracting artificial dataset that can then be used in
supervised learning models to learn and adapt to the actual system implementation.
Especially, deep reinforcement learning has showed great potential in recent years in
dealing with complex scheduling optimization problems. Researchers have focused
on the implementation of deep reinforcement learning techniques for production
scheduling-related problems where there is lack of data, and the problem appears
152 E. Bakopoulos et al.
high complexity. The Job-Shop Scheduling Problem (JSSP) is one of the most
common optimization problems related to production scheduling that the scientific
community has tried to solve with the application of deep reinforcement learning.
Zhang et al. (2020) developed a deep reinforcement learning agent, able to select
priority dispatch rules to solve the JSSP [19]. Liu et al. (2020) followed a similar
deep reinforcement learning approach to solve both the static and dynamic JSSP
[34]. Rather than only solving the JSSP, there have been also solutions for the
optimization of the whole production system with the use of deep Q-learning, a
very popular deep reinforcement learning technique in the last decade [35].
While all the technologically innovative techniques have helped to develop
smarter and more efficient systems and tools, these solutions could also be integrated
in an efficient way in the actual production system through a digital twins (DT)
and can help in integrating such solution to increase productivity. Villalonga et al.
(2021) proposed a framework for dynamic scheduling with the use of digital twins to
represent actual production assets in order to enhance decision-making [36]. Zhang
et al. (2021) use the digital twin concept to gather real-time data from the shop
floor and realize an effective dynamic production scheduling [37]. To achieve real-
time decision-making, the implementation of a digital twin appears a great potential,
since uncertain and dynamic events are addressed effectively. Dynamic interactive
scheduling method can be enhanced and strengthened by the use of DT [26, 38].
However, digital twin concept can also be implemented to support production
scheduling in an offline mode, such as the offline simulation of a production system.
This gives the ability to train scheduling agents in more dynamic environments and
respond to uncertainties even when they have not yet been identified. Nevertheless,
a main challenge in implementing production scheduling solutions and digital twins
is the lack of a well-defined data model. A solution to this issue can be offered by the
Asset Administration Shell (AAS) concept. AAS is basically a method to represent
data in a defined architecture [13, 39]. While in other problems there is some effort
made by the literature to implement AAS concept, in production scheduling it is not
explored.
The need to explore and address well-defined standards for production opti-
mization agents is clearly revealed when there is a need for cooperation between
different production agents, in order to formulate a multi-agent system. Researchers
from a variety of fields have given multi-agent systems (MASs) a great deal of
attention as a way to break down complicated problems into smaller jobs. Individual
tasks are assigned to agents, which are autonomous entities. Using a variety but
well-defined inputs, each agent chooses the most appropriate plan of action to
complete the task [40]. Agents make decisions based on the information provided
in the environment they are integrated and choose their actions proactively or
reactively [41]. In manufacturing, multi-agent systems have gathered attention of
many researchers during recent years. MAS can limit the complexity of order
scheduling in production systems through a cooperative multi-agent system for
production control optimization [42]. A similar approach was followed for the
implementation decentralized scheduling algorithms in a test-bed environment [43].
A scheduling strategy to assist a manufacturing system experiencing learning and
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 153
forgetting was supported by a multi-agent system to carry out the scheduling tasks in
conventional production systems in close to real-time, and a simulation was utilized
for validation [44].
While the multi-agent systems implementation methods have been explored in
recent years, further investigation to address challenges is required. For example,
the use of standards in a scheduling multi-agent system is something crucial, in
order to develop systems that could be easily transformed to a “plug & play”
application. In addition, agents that control or implement different applications
and software should follow a hierarchical implementation to achieve better multi-
agent system utilization and agents’ distribution. Lastly, if external applications are
controlled through a multi-agent system functionality, Application Programming
Interface (API) and standards are almost inevitable for the proper scheduling MAS
integration for the actual production system. The implementation of the scheduling
multi-agent system proposed in this work addresses the aforementioned issues and
gives the opportunity for a more flexible implementation of scheduling algorithms,
with different functionalities and heterogenous optimization techniques.
scope of this framework, and although the AAS is used for exchanging information
between ERP software and the agents, the underlying model is not standardized.
As displayed in Fig. 1, information from the enterprise resource planning (ERP)
are described within AASs for the corresponding work orders that the manager is
called to satisfy within the following production period. This type of information is
restored from the user interface (UI), allowing the user (in this case the production
manager) review the workload of the upcoming days. The connection between the
AAS and the ERP is performed via an ERP-to-AAS connector so that the proposed
UI platform depends on the AAS model rather than the specific ERP information
model structure. The UI rather than visualization of production information, it is
also an enabler for interaction of the user with the MAS as well as the production
digital twin. It is important to highlight that, unlike other systems, the integration of
decision-making results to the actual system is not a trivial task. In practice human
interferences is required to review and apply the production plan.
The exchange of information between the UI and the MAS is achieved via a
MAS API, which is in practice a way of passing and receiving data regarding the
production workload and status. The MAS is responsible for handling the data and
provide scheduling decisions for the user given the current production scenario.
There are multiple AI agents that were developed to address this problem each one
giving its own benefits for the user. The reason for using more than one agent for a
scheduling problem arises due to complexity of the problem, the user requirements,
as well as the problem itself. Scheduling problems are widely diverse with respect to
the environment, constraints, objectives, and equivalently the optimization methods
are usually compatible with a small portion of the overall set of scheduling
problems available. To this end, there cannot be a monolithic approach capable
of addressing all production scheduling problems without lacking on satisfying the
user requirements. In order to address this issue, there was proposed the concept of a
meta-scheduling agent, which in practice was a compound of multiple AI scheduling
agents each one providing different optimization attributes.
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 155
Fig. 2 MAS implementation based on the AAS description for the agent, showing interaction
between agents and services/schedulers as well as the agent spawn procedure
The AAS was developed for the description of agent, which was retrieved by
the MAS framework in order to deploy the corresponding entities and bring the
algorithms to life. The AAS model for the meta-agent was consisted of a toolbox
of optimization methods, with the description of connection dependencies as well
as capabilities and skills provided by the specific method. During the initialization,
there were spawned individual entities within the MAS, each one carrying a specific
set of skills (operations) corresponding to the AAS operations. It is important to
highlight though that the deployment of the agent within the MAS with the actual
algorithm runtime may differ. Specifically, the MAS is operating within a single
framework, which is usually a local installation of all the partial components, and in
this case the deployment of the algorithms is better to be remote. Figure 2 illustrates
this aspect for an example case of a scheduling agent. It can be displayed that a
scheduling agent AAS may contain more than one scheduling methods, which are
spawned as individual agents within the MAS framework. On top of that a meta-
scheduling agent is spawned within this framework in order to support the scheduler
selection and orchestration process within the MAS. The scheduling algorithms,
however, may be deployed in different remote servers depending on the case. While
a scheduling operation is requested from one of the schedulers, the AAS interfaces
support the communication between the agent installation and the actual algorithm.
In the previous architecture, it is important to clarify the need for the meta-
agent as well as the requirement for generating multiple agent entities within
the MAS framework. In essence, the notion of an agent, as an independent
156 E. Bakopoulos et al.
Fig. 3 Simulation model of the wheels assembly department utilized as a digital twin to apply the
schedule outcome from the MAS and observe performance
more than one scheduler would comply with the rules and more than one responses
may be produced.
It is also important to highlight that within a request there was used a specific
information structure for providing the production data and similarly the scheduling
outputs were contained within a specific scheduling response. The structure of the
information may vary based on the implementation and thus it was not specified
within this section. There are different alternative standards also to be used, while
in some cases a specific ERP data model could be also utilized. In any case, this
is another important aspect that is not specified within the chapter. However, the
methodology remains the same, with the exception that the problem classification
should be applied to a different model.
The digital twin was the final component of the architecture and ensured that
information is validated in a close-to-reality scenario and the system performance
is approved by the user. The production schedule was received by the MAS and
then sent (on-demand) to the digital twin in order to calculate its performance (see
Figs. 3 and 4). This step was executed before the schedule was displayed in detail
to the user as there could be one of multiple competitive schedules available for a
single case from different schedulers. The reason for using a digital representation
of the production system was to give the ability to the user to evaluate the resulted
schedule.
158 E. Bakopoulos et al.
Fig. 4 Simulation model of the painting department utilized as a digital twin to apply the schedule
outcome from the MAS and observe performance
The paint shop scheduling agents were designed in order to be able to give
solution to the Paint Shop Scheduling Problem (PSSP) as it can be found in the
literature. This problem addresses the sequence of the items entering the painting
line of the factory in order to optimize the performance indicators. This problem is
different from other scheduling problems as it usually encounters higher detail in the
combination of items and sorting before entering the line. The line itself is usually a
moving conveyor of carriers with some specific spatial constraints, setup delays due
to color, and a constant speed. The objective is to find the optimal combination of
the items within the “bill of material” of the products and sequence them in order to
comply with the desired performance.
Figure 5 illustrates the PSSP in a simplistic way. As it can be seen the goal
is to create a schedule – sequence and combination of items – for entering the
painting line so as to create the maximum utilization of the line, which will reflect
in reducing the makespan for the system. There are some requirements, however,
that the decision-making system needs to comply with in order to be in line with
the physical characteristics of the system. The following aspects were taken into
consideration:
• The conveyor speed is constant and the carriers are equally spaced along the line.
This ensures that the input and output rate of the line is also constant.
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 159
• Each carrier has a unique capacity (100%) similar of all the carriers that cannot
be exceeded.
• Items with different color cannot be placed within the same carrier. This is
because in most cases the items are painting all together within the painting
cabins.
• In cases where two consecutive colors are placed within the line there should be
a setup delay, expressed in empty carriers so that the operators have the time to
setup the new color.
• Each item type occupies a specific percentage of the carrier and can be mixed
with others as long as the max capacity is not violated.
• In cases where an item cannot fit into one container/carrier, then it will be used
the next consecutive carrier in order to hold the remaining capacity of the item.
It is made the exception, however, that no item needs more capacity than two
carriers.
Based on the proposed conversions, the following mathematical formulation can
be created:
Sets:
P Set of production orders that need to be painted
I Set of different items (types) that need to be painted
C Set of color codes
Parameters:
qp, i Quantity of items i ∈ I included in production order p ∈ P, qp, i ∈ Z≥0
ci Capacity a carrier required in order to carry item i ∈ I; ci ∈ R>0
dp, p’ Setup delay required between the items of two different production orders
(p, p’ ) ∈ P, dp, p’ ∈ Z≥0
Auxiliary variables:
si Defines whether item i ∈ I has a size higher than one carrier, si ∈ {0, 1}
ac, t Defines whether this color c ∈ C has entered the line on time t ∈ Z≥0 ,
ac, t ∈ {0, 1}
ep, t Defines whether this item i ∈ I has entered the line on time t ∈ Z≥0 , ep, t ∈ {0, 1}
Decision variable:
160 E. Bakopoulos et al.
xp, i, t The number of items i ∈ I from product p ∈ P that will enter the line on time
t, xp, i, t ∈ Z≥0,
Counters:
nt Number of timesteps available in the schedule
ni Number of item types in I-set
np Number of products in P-set
nc Number of colors in C-set
Constrains:
First, in a feasible schedule, we need to ensure that all items enter the resource
exactly once at some point during production time. This can be covered from the
following linear equality:
∞
. xp,i,t = qp,i • (1 + si ) , ∀p ∈ P , ∀i ∈ I
t=0
Number of constrains : np ∗ ni
.
Limitation for not allocating into the same carrier more than the items that can
hold based on its capacity can be achieved via the following inequality:
1
. ci • 1 − si • xp,i,t ≤ 1, ∀t ∈ Z≥0
2
∀p∈P ∀i∈I
.Number of constrains : nt
In addition, for cases when items require more than one carrier, it needs to be
placed in two consecutive carriers. This can be ensured by the following nonlinear
expression:
N
. si xp,i,t • xp,i,t+1 ≥ 1
t=0 ∀p∈P ∀i∈I
Number of constrains : 1
.
In order to transition to the linear version of the above expression, we start from
the logical expression:
. xp,i,t > 0 ∧ xp,i,t+1 > 0 ∨ xp,i,t−1 > 0
Then, we define two auxiliary variables to carry the outputs of the above logical
expressions:
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 161
zp,i,t = xp,i,t+1 > 0 ∨ xp,i,t−1 > 0
.
.yp,i,t = xp,i,t > 0 ∧ zp,i,t
In addition, when changing color between subsequence items the setup delay
must be applied. This can be achieved by the following linear inequalities:
∀p ∀i xp,i,t • fp,c
ac,t ≥
. , ∀t ∈ 0, Z≥0 , ∀c ∈ C
∀p ∀i qp,i
t − t '
ac,t
. + ac’,t ’ ≤ 1 + ∀ c, c' ∈ C, ∀t ∈ Z≥0 , ∀t ' ∈ t, t + dc,c' | c /= c'
dc,c' + 1
The same constraint can be achieved via the following nonlinear equation:
⎛ ⎞
t+dc,c'
N
. ⎝ac,t • 1 − ac’,t ’ ⎠ = 0, ∀ c, c' ∈ C | c /= c'
t=0 t ’=t
Objective function:
Ltotal The total flowtime of the production:
∞
Ltotal =
. t • ac,t
t=0
∞
Lweighted =
. t • wp • ep,t
t=0 ∀i∈I
λi, k The output (production) rate for an item type in a specific interval can be
defined as a moving average in the series of allocations for an item:
(k+1)L
t=k L ∀p∈P xp,i,t
λi,k =
. , k = 0, 1, 2, .., nt
L
Figure 5 shows where the requirements for this objective come from, and what
implications could come from missing to apply this objective. It is clear that in this
example, missing to produce the items at the average rates that are departed from
the buffer afterwards will cause an overflow, and in this case it is illustrated that the
circle needs to be at a much slower output rate than the cube item.
The ways that this can be applied are more than one, specifically the user may
require this to be a constrain to the scheduler, by means that at no times this rate is
exceeded, which can be applied by the next inequality, or via the objective function
trying to approach a specific value, yet this does not necessarily ensure that this
value will not deviate in the final results.
λi,k ≤ λdesired
. i , k = 0, 1, 2, .., N, ∀i ∈ I
∞
2
. min λi,k − λdesired
i
∀i∈I k=0
Fig. 6 Actual diagram from example case displaying the CPU delay differences of the modeling
approaches as the number of orders increase
• The linear version (MILP) was also considered where only the linear constrains
are utilized, improving the computational demand but increasing the require-
ments for memory utilization in computational resources.
• The last one was a simpler form of the linear version (two-stage MILP) in
which constrain #4 is removed from the model, running the optimization only for
mixing the items of order that acquire the same color. This allows a much faster
response horizon, since there were no setup constrains to apply in the schedule.
In a second stage, once the allocation of items is achieved, the optimization
process is repeated, but this time it schedules the sequence of colors as a function
of minimizing the setup delay. In this way, the model manages to reduce the
solutions space and the constrains limitations. The problem with this model,
however, is that it decreases the flexibility of the solution as a trade-off for lower
CPU time because it is not capable of providing good solutions for the production
rate issue.
In Fig. 6, the implications of the different models on the CPU duration of the
computational resources is clearly shown. The graphs are in a logarithmic scale in
the Y-axis and is clearly shown that both MILP and MINLP cannot outperform the
simplified-MILP, which can cover up to a very high number of items (100 orders
are usually 18,000 items) in a relatively short period of time (20 min).
Indices:
Discrete sets: np , ni , nc , nt
Decision variables: np 2 + ni (np + 1)
Input variables: np 2 + ni (np + 1)
164 E. Bakopoulos et al.
Fig. 7 Input layer (vector) encoding mechanism displaying an example for how the tables are
reshaped into a single dimensional vector
Fig. 8 Output layer (vector) encoding mechanism displaying the output can be encoded into the
allocation per product per item type per timestep
In order to avoid this long CPU delays and demanding RAM utilization, the
utilization of data-driven (i.e., ML) approaches was investigated in order to rather
predict the output of the scheduler. First, a feed-forward neural network (FFNN)
was developed, which uses as an input information over the workload data (i.e.,
orders, items, colors, and sizes), as well as produces the sequence of orders/items
allocated into the painting line. The input layer to the model was based on the
same parameters that were used for the mathematical formulation of MIP models,
resulting in the following encoded input vector (.x), while also the decision variable
xp,i,t was described the output vector (.y) (Figs. 7 and 8):
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 165
.Lx and .Ly shown in the above equation represent the dimensions of the input and
neurons# = 2k Lx
.
params# =
. 2k Lx 2k−1 Ly + 1
∀k:layer
Another data-driven approach has been developed that treats the scheduling
output as a time series from which the next allocations can be predicted based on
known previous values of the sequence. As such, the whole production schedule can
be generated in a recursive manner, reducing the model’s prediction variables which
improves the accuracy as well as avoiding any limitations regarding the problem’s
scalability. Similar to the above-mentioned approach, the ultimate objective is the
prediction of xp,i,t , for all the given orders, items, and timesteps; however, in this
model a prediction is only applied for one timestep and is repeated for all the output
sequence. The input features of the LSTM neural network consist of a dynamic and
a static part. The dynamic part as presented below are the features that change as
moving in the time axis.
The allocation of all items over a specific timestep (carrier) is given by the
following vector:
Fig. 9 Overview of LSTM RNN I/O layers design and how the specific input is derived as well as
how the output is represented
The following defines a variable that provides the number of remaining items of
an order at that given timestep, given the sequence of previous allocation selections:
ti
.Q [p] [i] [ti ] = q [p] [i] − x [p] [i] [t]
t=0
Fig. 10 Each graph (row) shows the total number of allocations from a product over time. Each
graph contains two lines for display purposes
The deep reinforcement learning (DRL) agent was selected to solve the dynamic
scheduling problem (DSP). According to Chien and Lan [45], the DSP is susceptible
to a number of uncertainties, including machine failures, the arrival of new, urgent
jobs, and changes to job due dates. In the literature there are several articles on
the DSP [46–49]. DRL agent is also combined with DNNs and deep Q-network to
approximate of a state action value function [50, 51]. The proposed DRL agent is
combined with a discrete event simulator in order for training and testing the DRL
model. In details, the DES that was used is the Witness Horizon from Lanner [52].
The DRL and DES communicated via API, where the API is provided by Witness
Horizon. In addition, except from API files, text files were used to exchange data
among the DES and DRL (see Fig. 11). The concept that was used for the DRL
agent is to purpose task allocation to resources via the use of dispatch rules. In the
literature there are several research works that study the use of RL agent combined
with dispatch rules [53].
168 E. Bakopoulos et al.
The main concept of the Q-learning is to use the Bellman equation as a value
iteration update. The agent in a decision point t in a state st ∈ S selects an action
at ∈ A according to a policy π. Taking the action at the agent gets to a state st + 1
with transition probability p(st + 1 |st , at ) ∈ P(S × A → S) and reward rt ∈ R.
Additionally, γ is a discount factor at each timestep t. Also, a is a learning rate,
where 0 < a ≤ 1. The objective for the agent is to find the optimal policy π ∗ that
maximizes the expected sum of rewards. The Q-leaning has some limitations when
the environment is huge. For that reason, the deep Q-network (DQN) concept was
used. Coupled RL with deep learning techniques Q-tables can be replaced with
Q-function approximator with weights [55]. In order to solve the DSP problem,
due to the fact that the environment is huge, DRL DQN concept was used. Let
us denote as Q(s, a; θ i ) the approximate value using deep convolutional neural
network. Additionally, the ψ i are the weights at iteration i of the Q-network. The
experiences are denoted as et = (st , at , rt , st + 1 ) where each time t are stored to a
dataset Dt = {e1 , . . . , et }. Chosen uniformly at random an instance from the pool of
'
stored instances, a Q-learning update is applied of each experience (s, a, r, s )~U(D).
2
Li (θi ) = E(s,a,r,s ' )∼U (D)
. r + γ maxa ' Q s ' , a ' ; θi− − Q (s, a; θi )
θ i are the weights of the Q-network at the iteration i and .θi− are the network
weights used to compute the target in iteration i. Target network parameters (.θi− ) are
updated with the Q-network parameters every c step, where c is a constant number.
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 169
The state is a tuple of feature that characterizes a given input. This chapter
contains the stats of the resources (down, busy, and available), the stats of the tasks
(waiting, pending, on-going, and finished), and finally a list with the quantities or
the product orders. Moreover, an action describes the dispatch rule that is selected
by the DRL agent to propose the task allocation over resources.
Algorithm:
Three adjustable parameters, MNA, SR, DH
Initialize: MNA, SR, DH
while full schedule is not generated
Generate MNA-alternative-branches of allocations for DH-steps in
the future.
for each branch in alternatives:
Generate SR sub-branches of allocations from DH-step and for-
ward.
Calculate average score of SR sub-branches on each MNA branch.
Select alternative with the highest score.
Store allocations of the alternative for up to DH-steps
return: best alternative
For each decision tree, the algorithm returns a list with valid task-resource
allocations [61–64]. MNA and DH control the breadth and DH the depth of the
search, respectively. On the other hand, SR is used to direct the search toward
alternatives that can provide better quality solutions. Thus, the quality of the solution
depends on the selection of the MNA, DH, and SR.
4 Case Study
The multi-agent system for the scheduling agents was developed with the use of
JANUS, an efficient and flexible agent-oriented programming framework that gives
the opportunity for easy and fast deployment of virtual assets. JANUS multi-agent
system framework is compatible with the programming language SARL and also
with JAVA. In this multi-agent system, there are four main concepts that need to be
defined before the deployment of any agent: agents, events, capacities, and skills.
The agent instance stands for all the operating sequences required for a specific
batch of functionalities and operations to happen when the agent needs to operate.
Agents’ communication and behavior is controlled by events, which are predefined
patterns that allow all the agents in the framework to interact one with another.
The term capacity refers to an abstract description for an implementation in skills,
which is used to define reusable capabilities of agent patterns without defining
implementation details. Lastly, the concept of skills is a manner of implementing
a capability, which allows exchange and modification of implementations based on
own or adapted skills without modifying the agent’s behavior or the template agent’s
characteristics. To address the scheduling multi-agent system using the JANUS
framework, the scheduling agents are modeled as agents in the JANUS framework,
capable of spawning and operating under the control of a meta-agent, which is
the orchestrator agent inside the multi-agent system. The scheduling agents have
specific skills, related to the problem-solving algorithm and the meta-agent concept
was integrated, in order to realize an automated and distributed cooperation of the
different agents inside the multi-agent system, when there is a scheduling request.
The user is able to interact with the multi-agent system in the backend of a UI,
developed for the scheduling tasks visualization.
In practice, the meta-agent receives the scheduling request from the UI. This
scheduling request is modeled in an AAS, as already described in previous sections,
and the meta-agent is able to spawn the corresponding scheduling agent to solve a
particular scheduling problem. A scheduling agent is the parent “class” in JANUS
that implements events, skills, and methods, and can also consist of local variables.
Each one of the scheduling agents accommodated three MAS events:
• “Initialization,” where the scheduling agent has been spawned by the meta-agent
during the initialization of the framework and waits for a scheduling request
notification from the meta-agent. During the initialization of the agent, specific
scheduling agent parameters are defined and initialized, able to serve a specific
scheduling request type in the future.
• “Scheduling request,” where the meta-agent is requested to notify the corre-
sponding scheduling agent in order for the required scheduling computation to be
performed. After this event call, specific skills and operations are performed by a
scheduling agent in order for the scheduling algorithm to calculate the schedule.
• “Schedule response,” where the output of the scheduling task is emitted to all
other agents of the multi-agent system. When the scheduling agent finishes its
operation, the event notices every other agent in the framework that can listen to
this event.
172 E. Bakopoulos et al.
KPIs through the digital twin tab where a DES run of the resulted scheduling is
performed.
To validate the whole framework performance, discrete event simulation (DES)
was utilized. Two DES models were developed, representing the production envi-
ronments of the two departments from a bicycle production system. These DES
models were used to showcase the results of the scheduling request that the multi-
agent framework handled, as well as for the actual operation of the agent for
solving the dynamic scheduling problem. The heuristic and the MIP schedulers were
deployed for the painting department whereas the DRL scheduler was deployed
for the wheels assembly department. JANUS multi-agent system spawned all three
scheduling agents when the necessary information for accessing them is provided
within the AAS definition after the scheduling request formulated in the UI. As
such, the user could choose any of the scheduling agents and, using the toolbox
of schedulers provided in the UI, address similar or different kinds of problems.
The user sent scheduling operations to the multi-agent system in an abstract manner
without the need to specify the corresponding problem. After the scheduling request
arrival, the meta-agent was responsible to spawn the required scheduling agent.
Seamless integration between the SARL software and the individual schedulers was
achieved.
The resulted framework implementation showed great potential in achieving
multi-agent scheduling optimization. The UI (Figs. 13 and 14) allows the user
to evaluate the resulted scheduling through the use of DES. Production KPIs are
presented and through the evaluation of the system performance on each occasion,
one can decide if the resulting schedule is efficient. Manual tests were made
in collaboration with the production manager, and the results were validated for
their accuracy and precision. Hence, the proposed scheduling multi-agent system
implementation for the bicycle production industrial environment can effectively
handle the workload distribution among its different scheduling agents in order to
propose the most appropriate production sequence.
Table 1 summarizes the agent results from testing the framework over some real-
life examples of the industrial use-case. The results do not directly translate on
174 E. Bakopoulos et al.
business KPIs and are the log from the schedulers. This is why the digital twin
component is necessary to reflect how these solutions fit into the overall production
scenario and inspect the performance.
5 Conclusion
In conclusion, the multi-agent system (MAS), digital twin (DT), Asset Administra-
tion Shell (AAS) concept, and artificial intelligence (AI) technology are part of the
Industry 4.0, and more and more researchers and industrial experts aim to combine
these technologies. Digital manufacturing is an important step for industries and
researchers, where there are many gaps and challenges to overcome. Digitalization
will enable automation, increase efficiency, real-time decision-making, flexibility,
and adaptability in industries. This work proposes a MAS framework that was
developed for the bicycle industry using the concept of AAS, DT, and MAS for the
production scheduling problem. A mathematical optimization, deep reinforcement
learning, heuristic algorithm, and deep learning algorithm have been developed
to address the identified problems. The key contribution of this work is the
use of the DT to accurately simulate the production environment and increase
the efficiency of the developed AI agents. The AAS concept is also used to
guarantee interoperable data transfer within MAS. Future research directions could
be considered the continuous exploitation of the DT and AI integrations. Moreover,
the AAS technology was used to fully parameterize the agents and the production
environment on the simulator.
Acknowledgments This work has received funding from the European Union’s Horizon 2020
research and innovation program under grant agreement No. 957204 MAS4AI. The dissemination
of results herein reflects only the authors’ view, and the Commission is not responsible for any use
that may be made of the information it contains.
Table 1 Table with results from the different agents used to solve the use-case scheduling problems
Assignments no.
(average)
Reaction time Performance
Level Optimization plugin Orders Items (min) (average) Options (average)
Resource Conveyor scheduler 98 18,500 22 2 [Cores] 5% [MIP Gap]
(painting line) (simple MILP, 12 [Threads] 8000 [Hangers]
Gurobi)
Resource Conveyor scheduler 12 1500 71 2 [Cores] 7% [MIP Gap]
(painting line) (MILP, Gurobi) 32 [Threads] 700 [Hangers]
Resource Conveyor scheduler 12 1500 180 2 [Cores] 7% [MIP Gap]
(painting line) (MINLP, Gurobi) 32 [Threads] 700 [Hangers]
Resource Conveyor scheduler 20 2000 0 20 [pmax ], 16 1000 [Hangers]
(painting line) (FFNN) [imax ], 1000 27 [Constrain
[tmax ], 20 [cmax ] Violations]
Resource Conveyor scheduler 50 9000 0 50 [pmax ], 16 5000 [Hangers]
(painting line) (LSTM RNN) [imax ], 50[cmax ] 5 [Constrain
Violations]
Resource Factory scheduler 8000 [Tasks] 70 0.4% [MNA], 70% [Utility]
(preparation) 0.15% [SR], 2
[DH]
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . .
Resources (assembly Department 50 [Orders], 1000 [Jobs], 2500 [Tasks] 2 1 shift [DH] −7% [Makespan]
lines) scheduler 70–75%
[Utilization]
175
176 E. Bakopoulos et al.
References
1. Chryssolouris, G.: The design of manufacturing systems. In: Manufacturing Systems: Theory
and Practice, pp. 329–463 (2006). https://doi.org/10.1007/0-387-28431-1_5
2. Rodammer, F.A., White, K.P.: A recent survey of production scheduling. IEEE Trans. Syst.
Man Cybern. 18(6), 841–851 (1988). https://doi.org/10.1109/21.23085
3. Lawler, E.L., Lenstra, J.K., Rinnooy Kan, A.H.G., Shmoys, D.B.: Chapter 9 Sequencing
and scheduling: algorithms and complexity. In: Handbooks in Operations Research and
Management Science, vol. 4, no. C, pp. 445–522 (1993). https://doi.org/10.1016/S0927-
0507(05)80189-6
4. Tao, F., Zhang, H., Liu, A., Nee, A.Y.C.: Digital twin in industry: state-of-the-art. IEEE Trans.
Industr. Inform. 15(4), 2405–2415 (2019). https://doi.org/10.1109/TII.2018.2873186
5. Alexopoulos, K., Nikolakis, N., Chryssolouris, G.: Digital twin-driven supervised machine
learning for the development of artificial intelligence applications in manufactur-
ing. Int. J. Comput. Integr. Manuf. 33(5), 429–439 (2020). https://doi.org/10.1080/
0951192X.2020.1747642
6. Jiang, Y., Yin, S., Li, K., Luo, H., Kaynak, O.: Industrial applications of digital twins. Phil.
Trans. R. Soc. A. 379(2207) (2021). https://doi.org/10.1098/RSTA.2020.0360
7. Enders, M., Enders, M.R., Hoßbach, N.: Dimensions of digital twin applications-a lit-
erature review completed research. Accessed 16 June 2023. [Online]. Available: https://
www.researchgate.net/publication/359715537
8. Rasheed, A., San, O., Kvamsdal, T.: Digital twin: values, challenges and enablers from
a modeling perspective. IEEE Access. 8, 21980–22012 (2020). https://doi.org/10.1109/
ACCESS.2020.2970143
9. Park, K.T., Son, Y.H., Ko, S.W., Do Noh, S.: Digital twin and reinforcement learning-based
resilient production control for micro smart factory. Appl. Sci. 11(7), 2977 (2021). https://
doi.org/10.3390/APP11072977
10. Plattform Industrie 4.0 - Reference Architectural Model Industrie 4.0 (RAMI4.0) - an
introduction. https://www.plattform-i40.de/IP/Redaktion/EN/Downloads/Publikation/rami40-
an-introduction.html. Accessed 16 June 2023.
11. Wei, K., Sun, J.Z., Liu, R.J.: A review of Asset Administration Shell. In: IEEE International
Conference on Industrial Engineering and Engineering Management, pp. 1460–1465 (2019).
https://doi.org/10.1109/IEEM44572.2019.8978536
12. Arm, J., et al.: Automated design and integration of Asset Administration Shells in components
of Industry 4.0. Sensors. 21(6), 2004 (2021). https://doi.org/10.3390/S21062004
13. Wagner, C., et al.: The role of the Industry 4.0 asset administration shell and the digital twin
during the life cycle of a plant. In: IEEE International Conference on Emerging Technologies
and Factory Automation, ETFA, pp. 1–8 (2017). https://doi.org/10.1109/ETFA.2017.8247583
14. Cavalieri, S., Salafia, M.G.: A model for predictive maintenance based on Asset Administration
Shell. Sensors. 20(21), 6028 (2020). https://doi.org/10.3390/S20216028
15. Ocker, F., Urban, C., Vogel-Heuser, B., Diedrich, C.: Leveraging the Asset Administration
Shell for agent-based production systems. IFAC-PapersOnLine. 54(1), 837–844 (2021). https:/
/doi.org/10.1016/J.IFACOL.2021.08.186
16. Chryssolouris, G., Alexopoulos, K., Arkouli, Z.: Artificial intelligence in manufacturing
systems. Stud. Syst. Decis. Control. 436, 79–135 (2023). https://doi.org/10.1007/978-3-031-
21828-6_4/COVER
17. De Simone, V., Di Pasquale, V., Miranda, S.: An overview on the use of AI/ML in manufactur-
ing MSMEs: solved issues, limits, and challenges. Proc. Comput. Sci. 217, 1820–1829 (2023).
https://doi.org/10.1016/J.PROCS.2022.12.382
18. Shahgholi Zadeh, M., Katebi, Y., Doniavi, A.: A heuristic model for dynamic flexible job shop
scheduling problem considering variable processing times. Int. J. Prod. Res. 57(10), 3020–
3035 (2018). https://doi.org/10.1080/00207543.2018.1524165
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 177
19. Chen, X., An, Y., Zhang, Z., Li, Y.: An approximate nondominated sorting genetic algorithm to
integrate optimization of production scheduling and accurate maintenance based on reliability
intervals. J. Manuf. Syst. 54, 227–241 (2020). https://doi.org/10.1016/J.JMSY.2019.12.004
20. Essien, A., Giannetti, C.: A deep learning model for smart manufacturing using convolutional
LSTM neural network autoencoders. IEEE Trans. Industr. Inform. 16(9), 6069–6078 (2020).
https://doi.org/10.1109/TII.2020.2967556
21. Wang, L., Pan, Z., Wang, J.: A review of reinforcement learning based intelligent optimization
for manufacturing scheduling. Compl. Syst. Model. Simul. 1(4), 257–270 (2022). https://
doi.org/10.23919/CSMS.2021.0027
22. Negri, E., Fumagalli, L., Macchi, M.: A review of the roles of digital twin in CPS-
based production systems. Proc. Manuf. 11, 939–948 (2017). https://doi.org/10.1016/
J.PROMFG.2017.07.198
23. Jones, D., Snider, C., Nassehi, A., Yon, J., Hicks, B.: Characterising the digital twin: a
systematic literature review. CIRP J. Manuf. Sci. Technol. 29, 36–52 (2020). https://doi.org/
10.1016/J.CIRPJ.2020.02.002
24. Boschert, S., Rosen, R.: Digital twin-the simulation aspect. In: Mechatronic Futures: Chal-
lenges and Solutions for Mechatronic Systems and Their Designers, pp. 59–74 (2016). https://
doi.org/10.1007/978-3-319-32156-1_5/COVER
25. Botín-Sanabria, D.M., Mihaita, S., Peimbert-García, R.E., Ramírez-Moreno, M.A., Ramírez-
Mendoza, R.A., de Lozoya-Santos, J.: Digital twin technology challenges and applica-
tions: a comprehensive review. Remote Sens. 14(6), 1335 (2022). https://doi.org/10.3390/
RS14061335
26. Negri, E., Pandhare, V., Cattaneo, L., Singh, J., Macchi, M., Lee, J.: Field-synchronized digital
twin framework for production scheduling with uncertainty. J. Intell. Manuf. 32(4), 1207–1228
(2021). https://doi.org/10.1007/S10845-020-01685-9/FIGURES/16
27. Jhunjhunwala, P., Atmojo, U.D., Vyatkin, V.: Applying skill-based engineering using OPC-
UA in production system with a digital twin. In: IEEE International Symposium on Industrial
Electronics (2021, June). https://doi.org/10.1109/ISIE45552.2021.9576342
28. Fuchs, J., Schmidt, J., Franke, J., Rehman, K., Sauer, M., Karnouskos, S.: I4.0-compliant inte-
gration of assets utilizing the Asset Administration Shell. In: IEEE International Conference on
Emerging Technologies and Factory Automation, ETFA, pp. 1243–1247 (2019, Sept). https://
doi.org/10.1109/ETFA.2019.8869255
29. Pribiš, R., Beňo, L., Drahoš, P.: Asset Administration Shell design methodology using
embedded OPC unified architecture server. Electronics. 10(20), 2520 (2021). https://doi.org/
10.3390/ELECTRONICS10202520
30. Chen, S., Pan, Q.K., Gao, L.: Production scheduling for blocking flowshop in distributed
environment using effective heuristics and iterated greedy algorithm. Robot. Comput. Integr.
Manuf. 71, 102155 (2021). https://doi.org/10.1016/J.RCIM.2021.102155
31. Montiel, L., Dimitrakopoulos, R.: A heuristic approach for the stochastic optimization of mine
production schedules. J. Heuristics. 23(5), 397–415 (2017). https://doi.org/10.1007/S10732-
017-9349-6/FIGURES/17
32. Aghelinejad, M.M., Ouazene, Y., Yalaoui, A.: Production scheduling optimisation with
machine state and time-dependent energy costs. Int. J. Prod. Res. 56(16), 5558–5575 (2017).
https://doi.org/10.1080/00207543.2017.1414969
33. Jélvez, E., Morales, N., Nancel-Penard, P., Cornillier, F.: A new hybrid heuristic algorithm
for the precedence constrained production scheduling problem: a mining application. Omega
(Westport). 94, 102046 (2020). https://doi.org/10.1016/J.OMEGA.2019.03.004
34. Liu, C.L., Chang, C.C., Tseng, C.J.: Actor-critic deep reinforcement learning for solving job
shop scheduling problems. IEEE Access. 8, 71752–71762 (2020). https://doi.org/10.1109/
ACCESS.2020.2987820
35. Waschneck, B., et al.: Optimization of global production scheduling with deep reinforcement
learning. Proc. CIRP. 72, 1264–1269 (2018). https://doi.org/10.1016/J.PROCIR.2018.03.212
36. Villalonga, A., et al.: A decision-making framework for dynamic scheduling of cyber-physical
production systems based on digital twins. Annu. Rev. Control. 51, 357–373 (2021). https://
doi.org/10.1016/J.ARCONTROL.2021.04.008
178 E. Bakopoulos et al.
37. Zhang, M., Tao, F., Nee, A.Y.C.: Digital twin enhanced dynamic job-shop scheduling. J.
Manuf. Syst. 58, 146–156 (2021). https://doi.org/10.1016/J.JMSY.2020.04.008
38. Fang, Y., Peng, C., Lou, P., Zhou, Z., Hu, J., Yan, J.: Digital-twin-based job shop scheduling
toward smart manufacturing. IEEE Trans. Industr. Inform. 15(12), 6425–6435 (2019). https://
doi.org/10.1109/TII.2019.2938572
39. Inigo, M.A., Porto, A., Kremer, B., Perez, A., Larrinaga, F., Cuenca, J.: Towards an Asset
Administration Shell scenario: a use case for interoperability and standardization in industry
4.0. In: Proceedings of IEEE/IFIP Network Operations and Management Symposium 2020:
management in the Age of Softwarization and Artificial Intelligence, NOMS 2020 (2020,
April). https://doi.org/10.1109/NOMS47738.2020.9110410
40. Dorri, A., Kanhere, S.S., Jurdak, R.: Multi-agent systems: a survey. IEEE Access. 6, 28573–
28593 (2018). https://doi.org/10.1109/ACCESS.2018.2831228
41. Cardoso, R.C., Ferrando, A.: A review of agent-based programming for multi-agent systems.
Computers. 10(2), 16 (2021). https://doi.org/10.3390/COMPUTERS10020016
42. Dittrich, M.A., Fohlmeister, S.: Cooperative multi-agent system for production control
using reinforcement learning. CIRP Ann. 69(1), 389–392 (2020). https://doi.org/10.1016/
J.CIRP.2020.04.005
43. Egger, G., Chaltsev, D., Giusti, A., Matt, D.T.: A deployment-friendly decentralized scheduling
approach for cooperative multi-agent systems in production systems. Proc. Manuf. 52, 127–132
(2020). https://doi.org/10.1016/J.PROMFG.2020.11.023
44. Renna, P.: Flexible job-shop scheduling with learning and forgetting effect by multi-
agent system. Int. J. Ind. Eng. Comput. 10(4), 521–534 (2019). https://doi.org/10.5267/
J.IJIEC.2019.3.003
45. Chien, C.F., Bin Lan, Y.: Agent-based approach integrating deep reinforcement learning and
hybrid genetic algorithm for dynamic scheduling for industry 3.5 smart production. Comput.
Ind. Eng. 162, 107782 (2021). https://doi.org/10.1016/J.CIE.2021.107782
46. Mohan, J., Lanka, K., Rao, A.N.: A review of dynamic job shop scheduling techniques. Proc.
Manuf. 30, 34–39 (2019). https://doi.org/10.1016/J.PROMFG.2019.02.006
47. Wen, X., Lian, X., Qian, Y., Zhang, Y., Wang, H., Li, H.: Dynamic scheduling method for
integrated process planning and scheduling problem with machine fault. Robot. Comput.
Integr. Manuf. 77, 102334 (2022). https://doi.org/10.1016/J.RCIM.2022.102334
48. Yan, Y., Wang, Z.: A two-layer dynamic scheduling method for minimising the earliness and
tardiness of a re-entrant production line. Int. J. Prod. Res. 50(2), 499–515 (2011). https://
doi.org/10.1080/00207543.2010.543171
49. Muhamadin, K., Bukkur, M.A., Shukri, M.I., Osama, Elmardi, M.: A review for dynamic
scheduling in manufacturing. Type: Double Blind Peer Reviewed Int. Res. J. Publ. Glob. J.
Online. 18, 25 (2018)
50. Hu, L., Liu, Z., Hu, W., Wang, Y., Tan, J., Wu, F.: Petri-net-based dynamic scheduling
of flexible manufacturing system via deep reinforcement learning with graph convolutional
network. J. Manuf. Syst. 55, 1–14 (2020). https://doi.org/10.1016/J.JMSY.2020.02.004
51. Chang, K., Park, S.H., Baek, J.G.: AGV dispatching algorithm based on deep Q-network in
CNC machines environment. Int. J. Comput. Integr. Manuf. 35(6), 662–677 (2021). https://
doi.org/10.1080/0951192X.2021.1992669
52. WITNESS Simulation Modeling Software | Lanner. https://www.lanner.com/en-gb/
technology/witness-simulation-software.html. Accessed 16 Jun 2023.
53. Zhang, C., Song, W., Cao, Z., Zhang, J., Tan, P.S., Chi, X.: Learning to dispatch for job
shop scheduling via deep reinforcement learning. Adv. Neural. Inf. Proc. Syst. 33, 1621–1632
(2020)
54. Mnih, V., et al.: Asynchronous methods for deep reinforcement learning. PMLR, 1928–
1937 (2016) Accessed 16 June 2023. [Online]. Available: https://proceedings.mlr.press/v48/
mniha16.html
55. Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature. 518(7540),
529–533 (2015). https://doi.org/10.1038/nature14236
Digital-Twin-Enabled Framework for Training and Deploying AI Agents for. . . 179
56. Kousi, N., Koukas, S., Michalos, G., Makris, S.: Scheduling of smart intra – factory material
supply operations using mobile robots. Int. J. Prod. Res. 57(3), 801–814 (Feb. 2018). https://
doi.org/10.1080/00207543.2018.1483587
57. Katoh, N., Ibaraki, T.: Resource allocation problems. In: Handbook of Combinatorial Opti-
mization, pp. 905–1006 (1998). https://doi.org/10.1007/978-1-4613-0303-9_14
58. Chryssolouris, G., Dicke, K., Lee, M.: On the resources allocation problem. Int. J. Prod. Res.
30(12), 2773–2795 (2007). https://doi.org/10.1080/00207549208948190
59. Chryssolouris, G., Papakostas, N., Mourtzis, D.: A decision-making approach for nesting
scheduling: a textile case. Int. J. Prod. Res. 38(17), 4555–4564 (2010). https://doi.org/10.1080/
00207540050205299
60. Michalos, G., Makris, S., Mourtzis, D.: A web based tool for dynamic job rotation
scheduling using multiple criteria. CIRP Ann. 60(1), 453–456 (2011). https://doi.org/10.1016/
J.CIRP.2011.03.037
61. Lalas, C., Mourtzis, D., Papakostas, N., Chryssolouris, G.: A simulation-based hybrid back-
wards scheduling framework for manufacturing systems. Int. J. Comput. Integr. Manuf. 19(8),
762–774 (2007). https://doi.org/10.1080/09511920600678827
62. Kousi, N., Michalos, G., Makris, S., Chryssolouris, G.: Short – term planning for part supply in
assembly lines using mobile robots. Proc. CIRP. 44, 371–376 (2016). https://doi.org/10.1016/
J.PROCIR.2016.02.131
63. Michalos, G., Fysikopoulos, A., Makris, S., Mourtzis, D., Chryssolouris, G.: Multi criteria
assembly line design and configuration – An automotive case study. CIRP J. Manuf. Sci.
Technol. 9, 69–87 (2015). https://doi.org/10.1016/J.CIRPJ.2015.01.002
64. Alexopoulos, K., Koukas, S., Boli, N., Mourtzis, D.: Resource planning for the installation of
industrial product service systems. IFIP Adv. Inf. Commun. Technol. 514, 205–213 (2017).
https://doi.org/10.1007/978-3-319-66926-7_24/FIGURES/5
65. Siatras, V., Nikolakis, N., Alexopoulos, K., Mourtzis, D.: A toolbox of agents for scheduling
the paint shop in bicycle industry. Proc. CIRP. 107, 1156–1161 (2022). https://doi.org/10.1016/
j.procir.2022.05.124
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
A Manufacturing Digital Twin
Framework
1 Introduction
The manufacturing industry is continuously evolving, and digital twin (DT) technol-
ogy has become a prominent driving force in this transformation. DTs play a crucial
role in optimizing manufacturing processes, increasing productivity, and enhancing
product quality.
A digital twin (DT) is a digital representation of a physical entity or process
modeled with the purpose to improve the decision-making process in a safe and
cost-efficient environment where different alternatives can be evaluated before
implementing them. The digital twin framework (DTF) for manufacturing is a set
of components for making and maintaining DTs, which describe the current state of
a product, process, or resource.
DTs have pace momentum due to their seamless integration and collaboration
with technologies such as IoT, machine learning (ML) algorithms, and analytics
solutions. DTs and ML solutions benefit in a bidirectional way, as DTs simulate
real environments, being a source of data for training the always data-eager ML
algorithms. DTs are a source of data that would be costly to acquire in other
conditions such as private data tied to legal and ethical implications, data labeling,
complex data cleaning, abnormal data, or data gathering that require intrusive
V. Anaya ()
Information Catalyst SL, Xativa, Spain
e-mail: [email protected]
E. Alberti · G. Scivoletto
Nextworks SRL, Pisa, Italy
e-mail: [email protected]; [email protected]
capturing a holistic view of the manufacturing process and all its interconnected
elements.
Ensuring availability is essential for a data collection platform to be effective.
Manufacturing operations typically operate continuously, demanding a constant
flow of real-time data. The platform should guarantee uninterrupted data acquisition,
seamlessly handling substantial data volumes promptly. It should offer dependable
connectivity and resilient infrastructure to prevent data gaps or delays, thereby
maintaining synchronization between the DT and its physical counterpart.
To connect to a range of devices, machines, and systems, support for manufactur-
ing communication protocols is crucial. Networked devices that adhere to specific
protocols are often utilized in production environments. The data collection platform
should therefore be able to interact via well-established protocols like OPC-UA1 ,
MQTT2 , or Modbus3 . Rapid data transfer, synchronization, and seamless integration
are all made possible by this interoperability throughout the production ecosystem.
Finally, security is of utmost importance in data collection for DTs. Manu-
facturing data often includes sensitive information, trade secrets, or intellectual
property. The data collection platform must implement robust security measures,
including encryption, access controls, and data anonymization techniques, to protect
the confidentiality, integrity, and availability of the collected data. This ensures that
valuable manufacturing knowledge and insights remain protected from unauthorized
access or malicious activities.
The rest of this chapter is organized as follows: in the next subsection, we
explain DT usages in the manufacturing sector. In Sect. 2, we present the digital
twin framework. In Sect. 3, we present the case study and the methodology to
experimentally evaluate the proposed method. In Sect. 4, we discuss the conclusion.
where DT refers to digital twin, that is expressed as a function (F) aggregating: the
physical system (PS), the digital system (DS), an updating engine that synchronizes
the two words (P2V), a prediction engine that runs prediction algorithms (V2P), and
an optimization dimension containing optimizers (OPT).
One of the most relevant initiatives to standardize a DT’s main building blocks
is the one proposed by ISO 23247 [17] comprising a DT framework that partitions
a digital twinning system into layers defined by standards. The framework is based
on the Internet of Things (IoT) and consists of four main layers:
• Observable Manufacturing Elements: This layer describes the items on the
manufacturing floor that need to be modeled. Officially, it is not part of the
framework, as it already exists.
• Device Communication Entity: This layer collates all state changes of the
observable manufacturing elements and sends control programs to those elements
when adjustments become necessary.
• Digital Twin Entity: This layer models the DTs, reading the data collated by the
device communication entity and using the information to update its models.
• User Entities: User entities are applications that use DTs to make manufacturing
processes more efficient. They include legacy applications like ERP and PLM, as
well as new applications that speed up processes.
On the other hand, the Digital Twin Capabilities Periodic Table (CPT) [16] is a
framework developed by the Digital Twin Consortium to help organizations design,
develop, deploy, and operate DTs based on use case capability requirements. The
CPT is architecture and technology agnostic, meaning it can be used with any
DT platform or technology solution. The framework clusters capabilities around
common characteristics using a periodic-table approach:
The CPT framework clusters capabilities into the following main clusters:
• Data Management: This cluster includes capabilities related to data access,
ingestion, and management across the DT platform from the edge to the cloud.
• Modeling and Simulation: This cluster includes capabilities related to the
creation of virtual models and simulations of real-world entities and processes.
• Analytics and Artificial Intelligence: This cluster includes capabilities related
to the use of analytics and artificial intelligence to analyze data and generate
insights.
• Visualization and User Interface: This cluster includes capabilities related to the
visualization of digital twin data and the user interface used to interact with the
DT
• Security and Privacy: This cluster includes capabilities related to the security and
privacy of DT data and systems
• Interoperability and Integration: This cluster includes capabilities related to the
integration of DT systems with other systems and the interoperability of DT data
with other data sources.
ISO 23247 and the Digital Twin Capabilities Periodic Table are generic frame-
works that are worth taking into consideration when developing a digital twin
186 V. Anaya et al.
• DT User View Backend: It is the backend engine that according to the decision
view of the digital twin can represent the different widgets (indicators, tables, 3D
view) that were defined in design time.
• Digital Domain Model Manager: This is the main backend of the DT. It is in
charge to create new DT instances based on data model definitions and connect
them to existing simulators and other AI algorithms (such as reinforcement
learning for production scheduling, neural networks for simulating the energy
consumption of manufacturing machines). Domain Data Models contain the
digital entities that will be part of the digital twin model, that is, the machines,
resources, places, and people. The Digital Domain Model Manager will support
the decomposition of digital elements in their parts through trees, and their
connection with the physical objects through the knowlEdge real-time brokering
component.
• DT Data Aggregator: It is the backend component in charge of maintaining the
DT model synchronized with the physical counterparts and offering APIs with
188 V. Anaya et al.
the rest of the components of the architecture. One of its components is the
context broker, which is based on the FIWARE Stellio Context Broker4 .
• 3D Visualization Engine: This component can render 3D scenes of the simula-
tions when a design is provided. Their results can be embedded into dashboards
used by the operators when running simulations.
• Behavior Manager: This component is in charge of keeping a linkage with
endpoints of the algorithms that define the behavior of digital things, for instance,
a linkage to the knowlEdge HP AI component that provides a REST API to the
repository of knowledge AI algorithms that are to be tested using the DT. This
subcomponent is also in charge of keeping a repository of linkages to simulators
and other algorithms through a REST API that can be third-party solutions
provided by external providers. The behavior manager has a scheduler engine
that runs simulations according to time events or data condition rules that are
executed against the DT model that is being filled with data from the IoT devices.
4 https://stellio.readthedocs.io/en/latest/index.html
A Manufacturing Digital Twin Framework 189
DTs have become an essential tool for improving shop floor processes in the
manufacturing industry. One specific application of a DT is for scheduling process
improvement. By using a DT, manufacturers can optimize their production sched-
ules to improve efficiency and reduce costs. The following is a description of how
the DT framework was applied to a dairy company within the knowlEdge project to
improve the management, control their processes, and automatize the scheduling of
the weekly production of yoghurt.
The knowlEdge Data Collection Platform (DCP) is used to connect to the shop
floor for gathering production and demand data. The platform was integrated with
various sensors and devices to collect data in real time. The DCP was also used to
collect data from various sources, such as the company’s ERP5 . By collecting data
from various sources, manufacturers can get a complete picture of their production
and demand data. The data was passed through the data collection platform for
filtering, formatting and normalization. This assured the proper quality of data and
ensured that the DT is accurate and reliable (see Fig. 3).
4 Conclusions
Digital twin technology has the potential to revolutionize the manufacturing industry
by optimizing processes, increasing productivity, and enhancing product quality.
By leveraging advanced digital techniques, simulations, and hybrid learning-based
modeling strategies, DT technology can help overcome the challenges faced by
traditional manufacturing methods and pave the way for the next generation of smart
manufacturing.
This chapter has presented the knowlEdge DT framework, an open-source toolkit
of DT modules supporting the modeling of physical assets and processes, and
the execution of functional and AI-based simulators for the execution of what-if
scenarios for improving the decision-making process. The tool has been used to also
for the generation of synthetic data for training AI algorithms. It is composed of a
set of modules as a DT Data Modeler, 3D twin modeler, IoT Ingestion Connector,
Simulator/AI Manager and Repository, Event Scheduler, DT Live Dashboard, and
the Data Collection Platform.
The DT Framework proposed was successfully used for creating a manufacturing
DT instance for generating weekly manufacturing schedules based on a rule-based
simulator and a discrete event simulator. The company where it was applied has
improved their reactiveness to incidents occurring on the shop floor, optimizing the
rescheduling process accordingly.
As more case studies and practical implementations emerge, the true potential of
DT technology in manufacturing will become increasingly apparent, driving further
transformation and innovation in the industry.
Acknowledgment This work has received funding from the European Union’s Horizon 2020
research and innovation programs under grant agreement No. 957331—KNOWLEDGE. This
paper reflects only the authors’ views, and the Commission is not responsible for any use that
may be made of the information it contains.
References
1. Stavropoulos, P., Mourtzis, D.: Digital twins in industry 4.0. In: Design and Operation of
Production Networks for Mass Personalization in the Era of Cloud Technology, pp. 277–316.
Elsevier, Amsterdam (2022)
2. Ogunsakin, R., Mehandjiev, N., Marin, C.A.: Towards adaptive digital twins architecture.
Comput. Ind. 149, 103920 (2023)
A Manufacturing Digital Twin Framework 193
3. He, B., Bai, K.J.: Digital twin-based sustainable intelligent manufacturing: a review. Adv.
Manuf. 9, 1–21 (2021)
4. Thelen, A., Zhang, X., Fink, O., Lu, Y., Ghosh, S., Youn, B.D., et al.: A comprehensive review
of digital twin—part 1: modeling and twinning enabling technologies. Struct. Multidiscip.
Optim. 65(12), 354 (2022)
5. Kim, D.B., Shao, G., Jo, G.: A digital twin implementation architecture for wire+ arc additive
manufacturing based on ISO 23247. Manuf. Lett. 34, 1–5 (2022)
6. Shao, G., Helu, M.: Framework for a digital twin in manufacturing: Scope and requirements.
Manuf. Lett. 24, 105–107 (2020)
7. Farhadi, A., Lee, S.K., Hinchy, E.P., O’Dowd, N.P., McCarthy, C.T.: The development of a
digital twin framework for an industrial robotic drilling process. Sensors. 22(19), 7232 (2022)
8. Zhong, D., Xia, Z., Zhu, Y., Duan, J.: Overview of predictive maintenance based on digital
twin technology. Heliyon (2023)
9. Hassan, M., Svadling, M., Björsell, N.: Experience from implementing digital twins for
maintenance in industrial processes. J. Intell. Manuf., 1–10 (2023)
10. Lee, J., Lapira, E., Bagheri, B., Kao, H.A.: Recent advances and trends in predictive
manufacturing systems in big data environment. Manuf. Lett. 1(1), 38–41 (2013)
11. Rodionov, N., Tatarnikova, L.: Digital twin technology as a modern approach to quality
management. In: E3S Web of Conferences, vol. 284, p. 04013. EDP Sciences (2021)
12. Fang, Y., Peng, C., Lou, P., Zhou, Z., Hu, J., Yan, J.: Digital-twin-based job shop scheduling
toward smart manufacturing. IEEE Trans. Ind. Inf. 15(12), 6425–6435 (2019)
13. Lu, Y., Liu, C., Kevin, I., Wang, K., Huang, H., Xu, X.: Digital twin-driven smart manufac-
turing: connotation, reference model, applications and research issues. Robot. Comput. Integr.
Manuf. 61, 101837 (2020)
14. Ogunseiju, O.R., Olayiwola, J., Akanmu, A.A., Nnaji, C.: Digital twin-driven framework for
improving self-management of ergonomic risks. Smart Sustain. Built Environ. 10(3), 403–419
(2021)
15. Ivanov, D., Dolgui, A., Sokolov, B.: The impact of digital technology and Industry 4.0 on the
ripple effect and supply chain risk analytics. Int. J. Prod. Res. 57(3), 829–846 (2019)
16. Capabilities Periodic Table – Digital Twin Consortium. Digital Twin Consortium. Published
August 8, 2022. https://www.digitaltwinconsortium.org/initiatives/capabilities-periodic-table/.
Accessed 2 June 2023
17. ISO 23247-2: ISO 23247-2: Automation Systems and Integration – Digital Twin Framework
for Manufacturing – Part 2: Reference Architecture. International Organization for Standard-
ization, Geneva (2021)
18. Wajid, U., Nizamis, A., Anaya, V.: Towards Industry 5.0–A Trustworthy AI Framework for
Digital Manufacturing with Humans in Control. Proceedings http://ceur-ws.org. ISSN, 1613,
0073 (2022)
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Reinforcement Learning-Based
Approaches in Manufacturing
Environments
1 Introduction
Over the past few decades, there has been a significant surge in the digitalization and
automation of industrial settings, primarily driven by the adoption of Industry 4.0
principles. At its essence, Industry 4.0 aims to establish a world of interconnected,
streamlined, and secure industries, built upon fundamental concepts such as the
advancement of cyber-physical systems (CPS) [1–3], the Internet of Things (IoT)
[4–6], and cognitive computing [7].
Computer numerical control machines (CNCs) play a pivotal role in aligning
with the principles of Industry 4.0 [8–10], facilitating automated and efficient
manufacturing of intricate and high-quality products. They have revolutionized
various industries such as woodworking, automotive, and aerospace by enhanc-
ing automation and precision. By automating industrial processes, CNCs reduce
the need for manual labor in repetitive and non-value-added activities, fostering
collaboration between machine centers and human operators in factory settings
[11]. Moreover, CNCs’ modular design and operational flexibility empower them to
perform a wide range of applications with minimal human intervention, ensuring the
creation of secure workspaces through built-in security measures. These machines
often incorporate advanced sensing and control technologies, optimizing their
performance and minimizing downtime.
In parallel with the rapid adoption of CNCs in the market, simulation techniques
have evolved to meet the industry’s latest requirements. The emergence of the digital
2 Reinforcement Learning
Reinforcement learning (RL) [15, 16] is widely recognized as the third paradigm of
Artificial Intelligence (AI), alongside supervised learning [26, 27] and unsupervised
learning [28]. RL focuses on the concept of learning through interactive experiences
while aiming to maximize a cumulative reward function. The RL agent achieves
this by mapping states to actions within a given environment, with the objective of
finding an optimal policy that yields the highest cumulative reward as defined in the
value function.
Two fundamental concepts underpin RL: trial-and-error search and the notion of
delayed rewards. Trial-and-error search involves the agent’s process of selecting
and trying different actions within an initially uncertain environment. Through
this iterative exploration, the agent gradually discovers which actions lead to the
maximum reward at each state, learning from the outcomes of its interactions.
The concept of delayed rewards [15, 26] emphasizes the consideration of not only
immediate rewards but also the expected total reward, taking into account subse-
quent rewards starting from the current state. RL agents recognize the importance of
long-term consequences and make decisions that maximize the cumulative reward
over time, even if it means sacrificing immediate gains for greater overall rewards.
By incorporating trial-and-error search and the notion of delayed rewards,
RL enables agents to learn effective policies by actively interacting with their
environment, continuously adapting their actions based on the feedback received,
and ultimately maximizing cumulative rewards.
Reinforcement learning (RL) problems consist of several key elements that work
together to enable the learning process. These elements include a learning agent, an
environment, a policy, a reward signal, a value function, and, in some cases, a model
of the environment. Let’s explore each of these elements:
1. Learning agent: The learning agent is an active decision-making entity that
interacts with the environment. It aims to find the optimal policy that maximizes
the long-term value function through its interactions. The specific approach and
logic employed by the learning agent depend on the RL algorithm being used.
2. Environment: The environment is where the learning agent operates and
interacts. It can represent a physical or virtual world with its own dynamics and
rules. The environment remains unchanged by the actions of the agent, and the
agent must navigate and adapt to its dynamics to maximize rewards.
198 A. Fernández Martínez et al.
3. Policy: The policy determines the behavior of the learning agent by mapping
states in the environment to the actions taken by the agent. It can be either
stochastic (probabilistic) or deterministic (fixed), guiding the agent’s decision-
making process.
4. Reward signal: The reward signal is a numerical value that the agent receives as
feedback after performing a specific action in the environment. It represents the
immediate feedback obtained during state transitions. The goal of the agent is to
maximize the cumulative rewards over time by selecting actions that yield higher
rewards.
5. Value function: The value function represents the expected total reward obtained
by the agent in the long run, starting from a specific state. It takes into account
the sequence of expected rewards by considering the future states and their
corresponding rewards. The value function guides the agent in estimating the
desirability of different states and helps in decision-making.
6. Model (optional): In some cases, RL algorithms incorporate a model of the
environment. The model mimics the behavior of the environment, enabling the
agent to make inferences about how the environment will respond to its actions.
However, in model-free RL algorithms, a model is not utilized.
In a typical reinforcement learning (RL) problem, the learning agent interacts
with the environment based on its policy. The agent receives immediate rewards
from the environment and updates its value function accordingly. This RL frame-
work is rooted in the Markov decision process (MDP) [28], which is a specific
approach used in process control.
RL has been proposed as a modeling tool for decision-making in both biological
[29] and artificial systems [18]. It has found applications in various domains such
as robotic manipulation, natural language processing, and energy management.
RL enables agents to learn optimal strategies by exploring and exploiting the
environment’s feedback. Inverse RL, which is based on hidden Markov models, is
another extensively studied topic in the field. Inverse RL aims to extract information
about the underlying rules followed by a system that generate observable behavioral
sequences. This approach has been applied in diverse fields including genomics,
protein dynamics in biology, speech and gesture recognition, and music structure
analysis. The broad applicability of RL and its ability to address different problem
domains make it a powerful tool for understanding and optimizing complex systems
in various disciplines.
Through iterative interactions, the agent adjusts its policy and value function to
optimize its decision-making process and maximize cumulative rewards. Figure 1
illustrates the common interaction flow in an RL problem.
As previously mentioned, the learning agent is situated within and interacts with
its environment. The environment’s state reflects the current situation or condition,
defined along a set of possible states denoted as S. The agent moves between states
by taking actions from a set of available actions, denoted as A. Whenever the
agent chooses and performs an action α from A, the environment E undergoes a
transformation, causing the agent to transition from one state S to another S ' , where
Reinforcement Learning-Based Approaches in Manufacturing Environments 199
(S, S ' )ϵS. Additionally, the agent receives a reward γ based on the chosen action α.
The ultimate objective of the agent is to maximize the expected cumulative reward
Rt over the long term, which can be estimated and reestimated through the learning
process of the agent to include and adapt to the new knowledge acquired.
A significant challenge in RL is striking the right balance between exploration
and exploitation. On one hand, it is advantageous for the agent to exploit its existing
knowledge gained from past experiences. By selecting actions that have previously
yielded high rewards, the agent aims to maximize the cumulative reward over time.
On the other hand, exploration is crucial to enable the agent to discover new states
and potentially identify better actions, thus avoiding suboptimal policies. Different
RL algorithms employ various approaches to address this trade-off.
A fundamental characteristic of MDPs and RL is their ability to handle stochastic
influences in the state–action relationship. This stochasticity is typically quantified
by a transition function, which represents a family of probability distributions that
describe the potential outcomes resulting from an action taken in a particular state.
By knowing the transition function, the agent can estimate the expected outcomes
of applying an action in a state by considering all possible transitions and their
corresponding probabilities. This analysis allows the agent to assess the desirability
or undesirability of certain actions.
To formalize this process, a value function U is defined [16]. The value function
assigns a numerical value to each state, representing the expected cumulative reward
the agent can achieve starting from that state and following a specific policy. It
serves as a measure of the desirability or utility of being in a particular state.
The value function helps guide the agent’s decision-making process by allowing
it to compare the potential outcomes and make informed choices based on maximiz-
ing the expected cumulative reward over time.
∞
U ∗ (s) = maxE
. γ t rt
π
t=0
Indeed, the parameter γ in the value function equation is widely referred to as the
discount factor. It plays a pivotal role in regulating the importance of future events
during the decision-making process, considering the delayed nature of rewards.
By adjusting the discount factor, one can determine the relative significance of
200 A. Fernández Martínez et al.
'
'
∗ ∗
U (s) = max r (s, a) + γ
. T s, a, s U s
a
s∈S
'
'
∗
. π (s) = argmax r (s, a) + γ T s, a, s U ∗ s
a
s∈S
As mentioned above, it returns the action to be applied on each state so that once
converged it returns the best action to be applied on each state.
The value function within a MDP can be also expressed or summarized in a
matrix that stores the value associated with an action a in a given state s. This matrix
is typically called Q-matrix and is represented by
'
' '
Q∗ (s, a) = r (s, a) + γ
. T s, a, s maxQ∗ s , a
a
s∈S
Reinforcement Learning-Based Approaches in Manufacturing Environments 201
In this section, we present two distinct examples that demonstrate the advancements
made in RL, specifically in the context of deep reinforcement learning (DRL). These
examples involve the application of RL within a virtual environment, which allows
for the development of strategies that can later be translated to real systems. This
approach opens up the possibility of deploying this technology in manufacturing
environments.
One key advantage of utilizing virtual environments is that it mitigates the
significant amount of learning episodes typically required by RL agents to develop
an optimal policy. In a real manufacturing system, the time needed to explore numer-
ous strategies would make the process highly inefficient for reaching an optimal
solution. Moreover, certain strategies may introduce risks, such as safety concerns,
that cannot be easily managed or assumed in a manufacturing environment. For
instance, in the second example, robotic systems operating collaboratively may pose
safety risks, while in the first example, machines with high power consumption may
introduce operational risks.
By leveraging virtual environments [32, 33], RL techniques can be effectively
applied to develop optimal strategies while minimizing risks and reducing the time
and costs associated with experimentation in real systems. This approach enables
Reinforcement Learning-Based Approaches in Manufacturing Environments 203
The digital twin (DT) of the physical CNC presented here was developed and shared
by one of the partners along the MAS4AI Project (GA 957204) within the ICT-38 AI
for manufacturing cluster for the development of the RL framework. The DT is built
on X-Sim, and it incorporates the dynamics of the machine, its parts, and the effects
on the working piece, simulating the physical behavior of the CNC. The CNC of
our study was a machining center for woodworking processes, more specifically for
cutting, routing, and drilling.
The machine considered (not shown for confidentiality issues) consists of a work-
ing table in which wood pieces are located and held throughout the woodworking
process and a robot-head of the CNC, which is responsible for performing the
required actions to transform the raw wood piece into the wood product (Fig. 3).
The model-based machine learning (ML) agent developed in this context aims at
optimizing the path trajectories of the five-axes head of a CNC within a digital twin
environment. Currently, the DT CNC enables the 3D design of the wood pieces by
users, creating all the necessary code to control the physical CNC.
Controlling a five-axes head automatically in an optimized way is yet an
engineering challenge. The CNC must not only perform the operation required to
transform the wood piece into the desired product, but it must also avoid potential
collisions with other parts of the machine, control the tools required by the head
204 A. Fernández Martínez et al.
of the robot, and keep times short reducing unnecessary movements to enhance
productivity and save energy, while ensuring the machine integrity and safety of
operators throughout the process, as well as high-quality products.
In this context, a model-based ML agent based on a DRL framework was trained
to optimize the path trajectories of the five-axes head of the CNC in order to avoid
potential collisions while optimizing the overall time operation. The difficulty of
working in a five-dimensional space (due to the five-axis robot head) is increased by
the dimensions of the working table of the CNC, which goes up to 3110 × 1320 mm.
In the DT environment, the measurement scale is micrometers, resulting in more
than 4,105,200 million states to be explored by the agent in a discrete state–action
space only in the plane XY of the board, without considering the extra three-axes of
the robot head. This complex applicability of discrete approaches is the reason why
only a continuous action space using Deep Deterministic Policy Gradient (DDPG)
[37] is shown here.
The model-based AI CNC agent was trained to work considering different
operations. The ultimate goal of the agent is to optimize the path trajectories
between two jobs in a coordinated basis considering the five-axes of the CNC head.
For this reason, the inputs of the model are the coordinates of the initial location, i.e.,
state of the five-axes head, and the destination location or a label representing the
desired operation to be performed by the CNC. The agent returns the optimized path
trajectory to reach the goal destination by means of a set of coordinates representing
the required movements of the five-axes head. Currently, the agent has been trained
separately to perform each operation independently. In a future stage, a multi-goal
DRL framework was explored in order to enhance generalization.
Different operations and different learning algorithms were explored during the
development of the deep RL framework, including 2-D, 3-D, and 5-D movements
of the five-axes head of the CNC, different path trajectories to be optimized, and
different learning algorithms including Q-learning [15, 38], deep Q-learning (DQL)
[39], and DDPG.
As seen previously, Q-learning is a model-free, off-policy RL algorithm that
seeks to find an optimal policy by maximizing a cost function that represents the
expected value of the total reward over a sequence of steps. It is used in finite
Markov decision processes (stochastic, discrete), and it learns an optimal action-
selection policy by addressing the set of optimal actions that the agent should take
in order to maximize the total reward (Rt ). The algorithm is based on an agent, a set
Reinforcement Learning-Based Approaches in Manufacturing Environments 205
of actions A, a set of states S, and an environment E. Every time the agent selects and
executes an action a ϵ A, the environment E is transformed, and the agent transitions
from one state, s, to another, s’, with (s, s’) ϵ S, receiving a reward r according to
the action selected.
DDPG is an off-policy algorithm that simultaneously learns a Q-function and a
policy based on the Bellman equation in continuous action spaces. DDPG makes
use of four neural networks, namely an actor, a critic, a target actor, and a target
critic. The algorithm is based on the standard “actor” and “critic” architecture [40],
although the actor directly maps states to actions instead of a probability distribution
across discrete action spaces.
In order to solve the problem of exhaustively evaluating all possible actions from
a continuous action space, DDPG learns an approximator to Q(s, a) by means of a
neural network, the critic Qθ (s, a), with θ corresponding to the parameters of the
network (Fig. 4).
Qθ learns from an experience replay buffer that serves as a memory for storing
previous experiences. This replay buffer contains a set D of transitions, which
includes the initial state (s), the action taken (a), the obtained reward (r), the new
state reached (s’), and whether the state is terminal or not (d). In other words, each
206 A. Fernández Martínez et al.
transition is represented as (s, a, r, s’, d), where s, a, r, s’, and d are elements of the
set D.
In order to evaluate the performance of Qθ in relation to the Bellman equation,
the mean-squared Bellman error (MSBE) can be computed. The MSBE quantifies
the discrepancy between the estimated Q-values produced by Qθ and the values
predicted by the Bellman equation. It is typically calculated by taking the mean
squared difference between the Q-value estimate and the expected Q-value, using
the current parameters θ. The MSBE provides a measure of how well the Q-function
approximated by Qθ aligns with the optimal Q-values as defined by the Bellman
equation. Minimizing the MSBE during training helps the DRL algorithm converge
toward an optimal Q-function approximation.
1
Qθ (s, a) − y r, s ' , d
2
L=
.
N
where the Qθtarget and μφtarget networks are lagged versions of the Qθ (critic) and
μφ (actor) networks to solve the instability of the minimization of the MSBE due
to interdependences among parameters. Hence, the critic network is updated by
performing gradient descent considering loss L. Regarding the actor policy, it is
updated using sampled policy gradient ascent with respect to the policy parameters
by means of:
1
∇φ
. Qθ s, μφ (s)
N
s∈D
Finally, the target networks are updated by Poliak averaging their parameters
over the course of training:
θQtarget ← ρθ Q + (1 − ρ) θQtarget
.
φμtarget ← ρφ μ + (1 − ρ) φμtarget
shall optimize its policy considering the operational time and quality of the path.
The operational time is calculated based on the distance that the robot head needs to
travel to reach the goal destination following the proposed path. The quality of the
path is calculated based on the number of actions needed to reach the destination.
The former two aspects are favored by an extra reward of +100 to the agent (Fig.
5).
Figure 6 shows four exemplary paths found by the agent in a 2-D movement
for visualization. In this problem, the agent needs to learn how to move from
right to left without colliding with any machine parts. Since the action space is
continuous, the goal destination does not correspond to a specific coordinate but to
a subspace of the environment. The circles draw on the paths represent the subgoal
coordinates proposed by the agent (each action correspond to a new coordinate to
which the robot head is moved). From the figure, it can be seen that the yellow
and pink trajectories comprise more actions and a longer path than the blue and
green trajectories. Although these latter contain the same number of actions (three
movements), the green trajectory requires a shorter path, and thus is preferred.
208 A. Fernández Martínez et al.
Fig. 6 Exemplary trajectories for simplicity in a 2-D movement. Circles in the path correspond
to set of coordinates obtained after performing the action proposed by the framework in the
environment. The total length of the trajectory and the number of actions are considered in the
reward function. The red square corresponds with the starting point, while the red rectangle
corresponds to the final region (target)
n
H (X) = −
. p (xi ) log p (xi )
i=1
Fig. 7 Example of fabric manipulation using the developed environment for training under a
complex manipulation task. The red and blue points indicate grab and pull points, respectively
the calculated entropy for each state. The DQL procedure has been trained built on
a ResNet18 architecture as a backbone with an input shape (224, 224, 4) and the
RL has been set with the following characteristic parameters: the gamma factor has
been initially set as 0.75 and epsilon has variated from 1 to 0.3 in 8000 steps. During
the whole learning process, a log file captures the global status of the knowledge
acquired by the system.
Figure 7 shows a large example of the fabric manipulation in such a virtual
environment (clothsim) in order to reduce the wrinkledness represented the entropy,
which is estimated based on phi, theta, and z coordinates of the normal vectors of
the points.
The images show the status of the fabric during the application of actions selected
based on the values of the Q-matrix. These values are developed during the training
procedure where the system tries to stablish, for given state, which action drives the
maximum entropy reduction. This means that the entropy is the metric used during
the whole process for the wrinkle minimization in the fabric (Fig. 8).
In order to select the actions, the knowledge of the system is encoded in the
classic Q-matrix, which is inferred by the system for a given state. Such codification
is done using a 6 × 6 matrix that considers corners to manipulate and directions that
can be taken within the fabric manipulation. The final outcome of the procedure is
a set of three points: one static point to fix the fabric, a second point that represents
Reinforcement Learning-Based Approaches in Manufacturing Environments 211
Fig. 8 Different examples of Q-matrix which is estimated for each state in order to drive an action
attending at the position of its maximum value. The minimum values a very small displacement of
the corners, meaning a very low reduction of the entropy so that they show low Q-values
the corner to be manipulated, and a third point that represents the point where this
last corner has to be placed (grab point, pull point initial coordinates, pull point final
coordinates). To decide what action to take, the system evaluates the Q-matrix and
selects the action that corresponds with the maximum value on the matrix through
an argmax function. The Q-values are updated to include a reward function that
becomes positive as entropy decreases and remains negative otherwise. So, the Q-
values hold the information of the reduction of the entropy that the system expects
by the application of a given action. In a way that through the application of the
action that corresponds with the maximum Q-value, the entropy reduction is also
expected to be maximum in the long run (considering a whole episode).
To quantify the solution, we validate the use of the entropy as a metric for the
wrinkledness of the fabric and its minimization as the target of the algorithm for
the development of a strategy to reduce such wrinkledness. Fig. 9 shows how the
entropy is reduced through the actions followed attending at the Q-matrix toward an
acceptance distance from the target (plane).
The action-selection method, which serves as the system’s plan to efficiently
eliminate wrinkles from fabric, relies on a knowledge structure. This knowledge
structure can be validated by evaluating the entropy variation resulting from the
application of different actions, taking into account the information stored in the
Q-matrix. In this process, various actions are applied to a given initial state, ordered
based on their optimality as determined by the Q-matrix. By examining the entropy
212 A. Fernández Martínez et al.
Fig. 9 Example of entropy evolution during synthetic fabric manipulation. The figure shows the
initial state of the fabric and the final after different actions are applied. The red arrow (top-left
corner) indicates the action suggested by the system. The right side of the figure shows the entropy
evolution during the application of the actions. It can be seen how it decreases until a certain
threshold (orange horizontal line) is crossed, meaning that the fabric is close to an ideal (plane)
state
Fig. 10 Optimal path. Comparison of the entropy evolution when always the maximum of the
Q-matrix is taken as optimal actions (blue curve) with entropy evolution under the selection of
suboptimal actions (green and orange curves). The starting point for each transition is always
considered as the state achieved by the application of the optimal (argmax from Q-values).
Suboptimal actions drive more irregular behavior
variation, one can assess the effectiveness of the selected actions and validate the
underlying knowledge structure.
Furthermore, the entropy’s evolution along the optimal path can be compared
with a scenario where suboptimal actions are taken, disregarding the Q-matrix. This
comparison allows us to observe how the selection of the maximum Q-value drives
a more robust curve toward entropy minimization. In Fig. 10, we can see that the
entropy follows a decreasing curve toward the minimum entropy when the optimal
actions are taken, considering the evolving state of the fabric. However, when
suboptimal actions are consistently chosen for the same fabric state, the entropy
Reinforcement Learning-Based Approaches in Manufacturing Environments 213
Fig. 11 Example of real fabric manipulation results. The figure shows three different steps during
the manipulation following the actions suggested by the system. The sample manipulated is one
from the real Use Case of the MERGING Project. In the figure, it can be clearly appreciated how
the wrinkledness is reduced as the actions are applied
exhibits a more erratic behavior. This demonstration highlights how the developed
solution offers a long-term optimal solution, ensuring a continuous reduction in
entropy along the optimal path.
The strategies have been also tested in a real scenario, exploiting the information
captured through a point cloud and following the outcomes suggested by the
analysis of the Q-matrix associated with the given state. However, a complete
description of the entire work behind this demonstration in terms of hardware is
out of the scope of this chapter. Nevertheless, Fig. 11 shows the state of a fabric
from an initial configuration to a final (ordered) state after the application of two
actions to reduce the wrinkles.
4 Conclusions
systems. This approach allows for safer and more efficient exploration, enabling the
optimization of manufacturing processes and performance.
The two manufacturing scenarios presented in this chapter highlight the potential
and applicability of reinforcement learning in improving industrial processes. By
bridging the gap between digital and real environments, we strive to advance the
field of manufacturing and drive innovation in these sectors.
Overall, this research sheds light on the benefits of applying reinforcement
learning in a digital context for manufacturing industries. It underscores the
importance of leveraging digital environments to enhance training and strategy
development, ultimately leading to improved performance and efficiency in real-
world systems.
Acknowledgments This work has received funding from the European Union’s Horizon 2020
research and innovation program under the MAS4AI Project (GA 957204) under the ICT-38 2020 –
Artificial Intelligence for Manufacturing Topic. This work has also received funding from the
MERGING Project (GA 869963).
References
1. Qi, Q., Tao, F.: Digital twin and big data towards smart manufacturing and industry 4.0: 360
degree comparison. IEEE Access. 6, 3585–3593 (2018)
2. Fuller, A., Fan, Z., Day, C., Barlow, C.: Digital twin: enabling technologies, challenges and
open research. IEEE Access. 8 (2020). https://doi.org/10.1109/ACCESS.2020.2998358
3. Wang, Z.: Digital twin technology. In: Bányai, T., Petrilloand, A., De Felice, F. (eds.) Industry
4.0 – Impact on Intelligent Logistics and Manufacturing. IntechOpen (2020)
4. Alexopoulos, K., Sipsas, K., Xanthakis, E., Makris, S., Mourtzis, D.: An industrial internet of
things based platform for context-aware information Services in Manufacturing. Int. J. Comput.
Integr. Manuf. 31(11), 1111–1123 (2018). https://doi.org/10.1080/0951192X.2018.1500716
5. Kumar, S., Tiwari, P., Zymbler, M.: Internet of Things is a revolutionary approach for future
technology enhancement: a review. J. Big Data. 6, 111 (2019). https://doi.org/10.1186/s40537-
019-0268-2
6. Saad, A., Faddel, S., Mohammed, O.: IoT-based digital twin for energy cyber-physical systems:
design and implementation. Energies. 13(18), 4762 (2020)
7. Wang, Y.: A cognitive informatics reference model of autonomous agent systems (AAS). Int.
J. Cogn. Inform. Nat. Intell. 3(1), 1–16 (2009c)
8. Usländer, T., Epple, U.: Reference model of Industrie 4.0 service architectures: basic concepts
and approach. Automatisierungstechnik. 63(10), 858–866 (2015). https://doi.org/10.1515/auto-
2015-0017
9. Phuyal, S., Bista, D., Bista, R.: Challenges, opportunities and future directions of smart
manufacturing: a state of art review. Sustainable Futures. 2, 100023 (2020). https://doi.org/
10.1016/j.sftr.2020.100023. ISSN 2666-1888
10. Ahuett-Garzaa, H., Kurfess, T.: A brief discussion on the trends of habilitating Technologies for
Industry 4.0 and smart manufacturing. Manuf. Lett. 15(Part B), 60–63 (2018). https://doi.org/
10.1016/j.mfglet.2018.02.011
11. Martins, A., Lucas, J., Costelha, H., Neves, C.: CNC machines integration in smart factories
using OPC UA. J. Ind. Inf. Integr. 34, 100482 (2023)
12. Alexopoulos, K., Nikolakis, N., Chryssolouris, G.: Digital twin-driven supervised machine
learning for the development of artificial intelligence applications in manufacturing. Int. J.
Comput. Integr. Manuf. 33(5), 429–439 (2020)
Reinforcement Learning-Based Approaches in Manufacturing Environments 215
13. Grieves, M., Vickers, J.: Digital twin: mitigating unpredictable, undesirable emergent behavior
in complex systems. In: Kahlen, F.-J., Flumerfelt, S., Alves, A. (eds.) Transdisciplinary
Perspectives on Complex Systems, pp. 85–113. Springer, Cham (2017)
14. He, B., Bai, K.J.: Digital twin-based sustainable intelligent manufacturing: a review. Adv.
Manuf. 9(1), 1–21 (2021)
15. Sutton, R.S., Barto, A.G.: Reinforcement learning: an introduction, 2nd edn. The MIT Press
(2014)
16. Kaelbling, L.P., Littman, M.L., Moore, A.W.: Reinforcement learning: a survey. J. Artif. Intell.
Res. 4, 237–285 (1996). https://doi.org/10.1613/JAIR.301
17. Jang, B., Kim, M., Harerimana, G., Kim, J.W.: Q-Learning algorithms: a comprehensive clas-
sification and applications. IEEE Access. 7, 133653–133667 (2019). https://doi.org/10.1109/
ACCESS.2019.2941229
18. Muiños-Landin, S., Fischer, A., Holubec, V., Cichos, F.: Reinforcement learning with artificial
microswimmers. Sci. Rob. 6(52), eabd9285 (2021)
19. Chen, G., Luo, N., Liu, D., Zhao, Z., Liang, C.: Path planning for manipulators based on an
improved probabilistic roadmap method. Robot. Comput. Integr. Manuf. 72 (2021). https://
doi.org/10.1016/j.rcim.2021.102196
20. Pohan, M.A.R., Trilaksono, B.R., Santosa, S.P., Rohman, A.S.: Path planning algorithm using
the hybridization of the rapidly-exploring random tree and ant Colony systems. IEEE Access.
9 (2021). https://doi.org/10.1109/ACCESS.2021.3127635
21. Wei, K., Ren, B.: A method on dynamic path planning for robotic manipulator autonomous
obstacle avoidance based on an improved RRT algorithm. Sensors (Switzerland). 18(2) (2018).
https://doi.org/10.3390/s18020571
22. Kang, J.G., Choi, Y.S., Jung, J.W.: A method of enhancing rapidly-exploring random tree robot
path planning using midpoint interpolation. Appl. Sci. 11(18) (2021). https://doi.org/10.3390/
app11188483
23. Wawrzynski, P.: Control policy with autocorrelated noise in reinforcement learning for
robotics. Int. J. Mach. Learn. Comput. 5, 91–95 (2015)
24. Wawrzynski, P., Tanwani, A.K.: Autonomous reinforcement learning with experience replay.
Neural Netw. 41, 156–167 (2013).; Xie, J., Shao, Z., Li, Y., Guan, Y., Tan, J.: Deep
reinforcement learning with optimized reward functions for robotic trajectory planning. IEEE
Access. 7, 105669–105679 (2019). https://doi.org/10.1109/ACCESS.2019.2932257; Watkins,
J.C.H.: Learning from Delayed Rewards. King’s College, Cambridge (1989)
25. Damen, T., Trampert, P., Boughobel, F., Sprenger, J., Klusch, M., Fischer, K., Kübel, C.,
Slusallek, P. et al.: Digital reality: a model-based approach to supervised Learning from
synthetic data. AI Perspect Adv. 1(1), 2 (2019). https://doi.org/10.1186/s42467-019-0002-0
26. Liu, Q., Wu, Y.: Supervised learning. In: Encyclopedia of the Sciences of Learning, pp. 3243–
3245 (2012). https://doi.org/10.1007/978-1-4419-1428-6_451
27. Hinton, G., Sejnowski, T.: Unsupervised Learning: Foundations of Neural Computation. MIT
Press, Cambridge (1999). https://doi.org/10.7551/MITPRESS/7011.001.0001
28. White, D.J.: A survey of applications of Markov decision processes. J. Oper. Res. Soc. 44(11),
1073–1096 (1993). https://doi.org/10.2307/2583870
29. Gustavsson, K., Biferale, L., Celani, A., Colabrese, S.: Finding efficient swimming strategies
in a three-dimensional chaotic flow by reinforcement learning. Eur. Phys. J. E. 40, 110 (2017)
30. Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., Riedmiller,
M.: Playing Atari with deep reinforcement learning. arXiv preprint arXiv, 1312.5602 (2013)
31. Sewak, M.: Deep Reinforcement Learning: Frontiers of Artificial Intelligence. Springer (2019).
https://doi.org/10.1007/978-981-13-8285-7
32. Dröder, K., Bobka, P., Germann, T., Gabriela, F., Dietrich, F.: A machine learning-
enhanced digital twin approach for human-robot-collaboration. In: 7th CIRP Conference on
Assembly Technologies and Systems, vol. 76, pp. 187–192 (2018). https://doi.org/10.1016/
j.procir.2018.02.010
33. Zayed, S.M., Attiya, G.M., El-Sayed, A., et al.: A review study on digital twins with artificial
intelligence and internet of things: concepts, opportunities, challenges, tools and future scope.
Multimed. Tools Appl. (2023). https://doi.org/10.1007/s11042-023-15611-7
216 A. Fernández Martínez et al.
34. Jazdi, N., Ashtari Talkhestani, B., Maschler, B., Weyrich, M.: Realization of AI-enhanced
industrial automation systems using intelligent Digital Twins. Procedia CIRP. 97, 396–400
(2020)
35. Hofmann, W., Branding, F.: Implementation of an IoT- And cloud-based digital twin for real-
time decision support in port operations. IFAC-PapersOnLine. 52(13), 2104–2109 (2019)
36. Bilberg, A., Malik, A.A.: Digital twin driven human–robot collaborative assembly. CIRP Ann.
(2019). https://doi.org/10.1016/j.cirp.2019.04.011
37. Lillicrap, T.P., et al.: Continuous control with deep reinforcement learning. arXiv:1509.02971.
(2016). https://doi.org/10.48550/arXiv.1509.02971
38. Maoudj, A., Hentout, A.: Optimal path planning approach based on Q-learning algo-
rithm for mobile robots. Appl. Soft Comput. 97, 106796 (2020). https://doi.org/10.1016/
J.ASOC.2020.106796
39. Chen, X., 5-axis coverage path planning with deep reinforcement learning and fast parallel col-
lision detection (2020). Available: https://smartech.gatech.edu/handle/1853/62825. Accessed
28 Aug 2022
40. Grondman, I., Busoniu, L., Lopes, G.A.D., Babuška, R.: A survey of actor-critic reinforcement
learning: standard and natural policy gradients. IEEE Trans. Syst. Man Cybern. Part C Appl.
Rev. 42(6), 1291–1307 (2012). https://doi.org/10.1109/TSMCC.2012.2218595
41. Chen, L., Jiang, Z., Cheng, L., Knoll, A.C., Zhou, M.: Deep reinforcement learning based
trajectory planning under uncertain constraints. Front. Neurorobot. 16, 80 (2022). https://
doi.org/10.3389/FNBOT.2022.883562
42. Jiménez, P., Torras, C.: Perception of cloth in assistive robotic manipulation tasks. Nat.
Comput. 19, 409–431 (2020). https://doi.org/10.1007/s11047-020-09784-5
43. Colomé, A., Torras, C.: Dimensionality reduction for dynamic movement primitives and
application to bimanual manipulation of clothes. IEEE Trans. Robot. 34(3), 602–615 (2018).
https://doi.org/10.1109/TRO.2018.2808924
44. Cusumano-Towner, M., Singh, A., Miller, S., O’Brien, J.F., Abbeel, P.: Bringing clothing
into desired configurations with limited perception. In: Proceedings of IEEE Interna-
tional Conference on Robotics and Automation (ICRA), vol. 2011, pp. 1–8 (2011) http://
graphics.berkeley.edu/papers/CusumanoTowner-BCD-2011-05/
45. Hamajima, K., Kakikura, M.: Planning strategy for task of unfolding clothes. Robot. Auton.
Syst. 32(2–3), 145–152 (2000). https://doi.org/10.1016/S0921-8890(99)00115-3
46. Hou, Y.C., Sahari, K.S.M.: Self-generated dataset for category and pose estimation of
deformable object. J. Rob. Netw. Artif Life. 5, 217–222 (2019). https://doi.org/10.2991/
jrnal.k.190220.001
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
A Participatory Modelling Approach to
Agents in Industry Using AAS
1 Introduction
With the increasing number of assets being digitized, all of which are expected to
be interoperable with each other, the need for well-designed information models
grows. Such models describe a given entity in a machine-readable way and enable
interpreting, usage and reasoning with data from previously unknown devices. The
usual approach to make these information models is through a heavy standardization
process in which a large number of organizations all have to come to a common
understanding of definitions and approaches, after which all parties implement this
understanding in their systems [17]. These standards are usually well-scoped and
documented, making them easier to use. For developers, this provides a steady
base to build upon without fear of the implementation becoming outdated quickly
due to a newly released version of a standard. Unfortunately, these standardization
Cornelis Bouter, Michael van Bekkum, and Sjoerd Rongen contributed equally to this work.
efforts tend to be slow, often taking multiple years before agreement is reached, for
example, ISO standards take 18 to 36 months to be developed [8].
Because of their slow development nature, formal standardization processes are
not universally suitable for all applications. For example, when developing a new
product, it is not feasible to delay the product launch by months, or even years,
because an agreement needs to be reached on the information model. As such,
currently, during development, there is little alignment done. Instead, the product
developer creates a model with the information that they consider necessary. This
is similar to what has been happening in the development of linked data ontologies,
where there are few formal standards standardized by W3C (RDF, OWL, SHACL)
defining the metamodel. A large number of common use case-specific models made
by institutions unrelated to W3C have become de facto standards, such as Friend
of a Friend,1 Prov-o,2 Schema.org.3 These de facto standards are commonly used
by individuals designing their models and enable easier alignment between a large
number of use case-specific models. This results in a large number of datasets that
mostly use domain-specific ontologies, while they can still be related to each other
as done with the Linked Open Data Cloud.4
We believe these de facto standards are vital to quickly defining the semantic
model. This belief appears to be shared with organizations in the manufacturing
industry, as they have developed their own metamodel to ensure domain models
can be aligned with each other. The standardized Asset Administration Shell (AAS)
defines how to structure the information model of your asset without specifying
what your asset is or what its information model is, cf. [2]. This provides a structure
for various organizations to express interoperable models, which may become
de factoo standards. However, in the Asset Administration Shell community, we
detect a lack of tools and methodologies that allow these bottom-up developed
AAS submodels to rise to the level of de facto standards. This leaves us with
formal standards (such as those published by the Industrial Digital Twin Association
(IDTA) [7]) and lots of use case-specific models which lack the reuse of domain-
relevant models which aids their interoperability.
In this paper, we build upon earlier work for AAS modelling practices [3] and
agent modelling approaches [10] to present a set of tools, aiming to help inter-
operability and standardization. We build on lessons learned from the MAS4AI5
European project in which software agents were described using an AAS utilizing a
reusable model of describing software agents. The methodology referred to in this
paper was developed within the European DIMOFAC6 project. In this work, we aim
1 http://xmlns.com/foaf/0.1/.
2 https://www.w3.org/TR/prov-o/.
3 https://schema.org/.
4 https://lod-cloud.net/.
5 https://www.mas4ai.eu.
6 https://dimofac.eu/.
A Participatory Modelling Approach to Agents in Industry Using AAS 219
2 Background
by the speed at which information models can be reviewed and consensus can be
reached. To be able to achieve widely adopted semantic interoperability, it is needed
to facilitate also a bottom-up approach.
Currently, there is a lack of common standardization, which is universally used,
as shown in the work of [9]. To tackle this, there are multiple different approaches
employed toward developing a common method—[5] proposes a solution using
OPC-UA, [4] defines an interoperability framework for digital twins, and [6] offer
a solution using ontologies. We tackle interoperability from the perspective of what
is needed for a standardization approach to happen and how can it be implemented.
Specifically, we look at the topic of agents and how their AAS modelling process
can be standardized. We develop a set of tools, which can be used to simplify and
streamline the process of modelling with AAS. To the best of our knowledge, there
is no current common process. We place all those developments in the context of
real Industry 4.0 applications.
The general model structure follows the work we presented in [10]. We provide
a set of general and standard submodels which aim to provide a structure, which
can be followed when creating any agent model. The model is shown in Fig. 1,
in which we add the Parameterization and Configuration submodels to replace
the Communication submodel from earlier work. The purpose of this change is to
address the difference between the information locally needed for an agent to be
defined, and the data, defined by the framework where it is deployed. This creates
a concrete distinction between what type of information is needed in the submodels
and where it comes from.
A Participatory Modelling Approach to Agents in Industry Using AAS 221
Fig. 1 General agent structure, which contains seven different distinct submodels, split into two
categories
Specific Submodels [10] are those, which depend on the use case. They contain
information, which is determined by the exact situation and setting where the agent
is to execute its work.
• Task: The task submodel provides a description of the exact task that the agent
has to execute. Depending on the case there can be a single task such as
“Moving,” but it can also be a set of multiple sub-tasks.
• Resources: The resources submodel aims to wrap in one place the connections
to all resources and corresponding relevant information from them, which the
agent would need. This would, for example, contain properties such as “Machine
Capability” and “Operator Availability.”
• Results: The results submodel presents the type of results and corresponding
details that the agent provides after its task is executed. There can be multiple
results such as “Analysis,” “Plan,” etc.
3.4 Usage
The concept of the general agent structure is to serve as a base skeleton model, which
provides a clear split and indication regarding what type of information needs to be
contained in an agent model. The aim is to use the structure, when creating models
of new agents, starting from concretely specifying the submodels to filling them in.
To provide an approach for this, we have defined a methodology based on the work
of [3], described in the next section.
When creating a rich semantic information model there are several aspects to
consider. To support proper modelling practices, we suggest the usage of a well-
defined process which ensures no steps are missed.
We propose a methodology based on earlier work by [3], which we extend
to make it more applicable when modelling not only physical assets but also
software assets. The methodology is additionally extended by identifying four
phases, described below and visually represented in Fig. 2.
A Participatory Modelling Approach to Agents in Industry Using AAS 223
Fig. 2 Updated AAS Development Methodology, visualizing the four separate steps—design,
define, align, deploy
4.1 Phases
In the case of agent modelling, the standard process described above is directly
applicable thanks to the two added states—provisioning and instantiation. Further-
more, the process can be split into four distinct stages, referring to the different
implementation stages of agent development.
1. Design: The design stage is the use case specification moment. This is, for
example, the point in time when we aim to identify the types of agents which
224 N. Nikolova et al.
5.1 Functionality
The main functionalities of the repository are (1) visualization and (2) distribution,
both focusing on making the AAS models’ development and cooperation easier.
Currently, the majority of development happens behind closed doors in silos, with
distribution and information sharing only happening at the last step of the process.
It is very important to enable ways for cooperation, especially since the increased
interest in the AASs also creates accidental duplication of work.
One of the main challenges for distribution is the lack of direct sharing
possibilities when it comes to AAS models. Presently, the standard process requires
downloading from a source (such as GitHub) and running an extra program (such as
AAS package explorer) to open and view a model. This can hinder the development
A Participatory Modelling Approach to Agents in Industry Using AAS 225
process, since more steps are required and hence more time and focus are needed
to review a model. What the AAS repository enables is direct link sharing. Once a
developer uploads their model online, it is possible to get a link, which leads to a
web page visualization of the AAS. The link can be shared with other parties and
removes the need for any software installation, which can significantly simplify the
sharing and collaboration process, which are key for creating general and reusable
models.
In general, a full AAS contains multiple Submodels, which each contain
several (nested) SubmodelElements. For each of these Submodels and
SubmodelElements, semanticIds, descriptions and Concept Descriptions
should be maintained. Moreover, multi-language SubmodelElements exists,
which can have several values attached for the various supported languages.
Because of this size and complexity, a model can have, working collaboratively on
an AAS template should be supported with proper visualization. Currently, the most
commonly used tool is also the one used for model creation—the package explorer.
While this is useful software, it can be unintuitive for non-accustomed users and
increase complexity when looking at a model (especially for external parties, who
do not work with this program). The repository provides a web interface, which
removes the need for separate software tooling for viewers.
The working principle of the repository is visualized in Fig. 3. The figure visualizes
the different interactions that a user can have with the repository and the high-level
corresponding processing steps.
• Upload: A user can upload a model to the repository by filling in the metadata
and providing the corresponding .aasx file. This model will be checked for
errors by the system and if no errors are found, it would be uploaded to the
repository.
• Modify: A user can modify the metadata of a model or upload a newer .aasx
version at any point.
• Visualize: The model visualizes multiple details of the .aasx model. It provides
collapsible submodels, each containing the corresponding elements. There is a
possibility to show concept description and most importantly example data. This
makes it easier for a viewer to understand the whole process of the models. If
example data contains references to other models included in the repository, they
can easily be followed via Web links. Especially for complex composite models,
this can make a significant difference.
• Share: Any of the models can be shared via a link, which directly links to the
relevant AAS file.
• Access: Model access is managed by a combination of groups/users and roles.
Each model can have several groups or individual users assigned. Additionally,
226 N. Nikolova et al.
Fig. 3 Repository working principle, where black arrows represent interaction with the user and
grey represent internal connections
for each user or group a role must be specified, i.e., guest, member, or editor. The
combination of group/user and role then determines the effective access level.
This ensures that during each step in the development process, the models can
be shared with appropriate access levels and no sensitive information is publicly
available.
6 Discussion
Although we have presented a number of ways to improve the usage of the AAS
and to ensure semantic interoperability we do not believe this will on its own
lead to wide-spread adoption and an active ecosystem of AAS users. We believe
A Participatory Modelling Approach to Agents in Industry Using AAS 227
The increased accessibility of the repository can aid the various groups who
can be expected to share their models. The methodology is primarily intended
for user integrators leveraging the AAS for improved interoperability to support
a factory application. Another audience that may use the repository is the machine
manufacturer who wants to share instantiated and specialized templates for their
machines. These submodels similarly benefit from a bottom-up approach without a
mandatory procedure.
Collaborative Modelling
Currently, the process of making an AAS model is primarily a solitary effort in
which the modeller still needs to actively try to reach out for input. However, the
whole point of standards is that they align between different parties. As such, adding
more tools to facilitate collaborative modelling would be beneficial for the creation
of a proper AAS. The presented repository and development methodology already
aid in this, but we believe further steps could be taken to support a shared model
discussion and increase iteration speed during the development process.
7 Conclusion
The methodology and the repository described in this paper provide us with
means to develop AAS models as de facto standards for non-physical assets from
bottom-up, community-based efforts. The MAS4AI project has shown that the
methodology can easily accommodate particular use case requirements by allowing
for straightforward bottom-up extensions to the models. An industry use case in
MAS4AI on implementing planning agent software based on ISA-95 models has
shown that bottom-up standardization is instrumental. The required models were
created by applying the aforementioned methodology in a collaborative effort
by modelling experts, domain experts, and software developers and have led to
successful integration and deployment in the use case.
The use of the repository has similarly led to clear benefits in both the MAS4AI
and the DIMOFAC project: sharing the AAS model templates with all stakeholders
in an easy-to-use, intuitive way has made them more accessible to all parties and
has promoted discussion and feedback on their contents, thus ensuring more wide-
spread support for the models.
Acknowledgments The research leading to these results has been funded by the European
Commission in the H2020 research and innovation program, under Grant agreement no. 957204,
the “MAS4AI” project (Multi-Agent Systems for Pervasive Artificial Intelligence for Assisting
Humans in Modular Production Environments) and under grant agreement No. 870092, the
DIMOFAC project (Digital & Intelligent MOdular FACtories).
A Participatory Modelling Approach to Agents in Industry Using AAS 229
References
1. 4.0 PI: Structure of the asset administration shell. continuation of the development of the ref-
erence model for the Industrie 4.0 component (2016) . Tech. rep. https://industrialdigitaltwin.
org/wp-content/uploads/2021/09/01_structure_of_the_administration_shell_en_2016.pdf
2. Bader, S., Barnstedt, E., Bedenbender, H., et al.: Details of the Asset Administration Shell. Part
1—The exchange of information between partners in the value chain of Industrie 4.0 (Version
3.0RC02) (2022)
3. Bouter, C., Pourjafarian, M., Simar, L., et al.: Towards a comprehensive methodology for
modelling submodels in the industry 4.0 asset administration shell. In: 2021 IEEE 23rd
Conference on Business Informatics (CBI), vol. 02, pp. 10–19 (2021)
4. Budiardjo, A., Migliori, D.: Digital twin system interoperability framework. Tech. rep. Digital
Twin Consortium, East Lansing, Michigan (2021). https://www.digitaltwinconsortium.org/pdf/
Digital-Twin-System-Interoperability-Framework-12072021.pdf
5. Cavalieri, S.: A proposal to improve interoperability in the industry 4.0 based on the open
platform communications unified architecture standard. Computers 10(6), 70 (2021). https://
doi.org/10.3390/computers10060070, https://www.mdpi.com/2073-431X/10/6/70, number: 6
Publisher: Multidisciplinary Digital Publishing Institute
6. Huang, Y., Dhouib, S., Medinacelli, L.P., et al.: Enabling semantic interoperability of asset
administration shells through an ontology-based modeling method. In: Proceedings of the 25th
International Conference on Model Driven Engineering Languages and Systems: Companion
Proceedings. Association for Computing Machinery, New York, NY, USA, MODELS ’22, pp.
497–502 (2022). https://doi.org/10.1145/3550356.3561606
7. IDTA: Industrial digital twin association (2023). https://industrialdigitaltwin.org/en/
8. ISO: Target date planner (2023). https://www.iso.org/files/live/sites/isoorg/files/developing_
standards/resources/docs/std%20dev%20target%20date%20planner.pdf
9. Melluso, N., Grangel-González, I., Fantoni, G.: Enhancing industry 4.0 standards inter-
operability via knowledge graphs with natural language processing. Comput. Ind. 140,
103676 (2022). https://doi.org/10.1016/j.compind.2022.103676, https://www.sciencedirect.
com/science/article/pii/S0166361522000732
10. Nikolova, N., Rongen, S.: Modelling agents in industry 4.0 applications using asset admin-
istration shell. In: Proceedings of the 15th International Conference on Agents and Artificial
Intelligence (2023). https://doi.org/DOI:10.5220/0011746100003393
11. Ocker, F., Urban, C., Vogel-Heuser, B., et al.: Leveraging the asset administra-
tion shell for agent-based production systems. IFAC-PapersOnLine 54(1), 837–844
(2021). https://doi.org/https://doi.org/10.1016/j.ifacol.2021.08.186, https://www.sciencedirect.
com/science/article/pii/S2405896321009563. 17th IFAC Symposium on Information Control
Problems in Manufacturing INCOM 2021
12. OWL Working Group: Owl 2 web ontology language document overview (2012). Tech. rep.,
W3C. https://www.w3.org/TR/owl2-overview/
13. Plattform Industrie 4.0: Details of the asset administration shell from idea to implemen-
tation (2019). https://www.plattform-i40.de/IP/Redaktion/EN/Downloads/Publikation/vws-in-
detail-presentation.pdf
14. RDF Data Shapes Working Group: Shapes constraint language (SHACL) (2017). Tech. rep.,
W3C. https://www.w3.org/TR/shacl/
15. RDF Working Group: RDF 1.1 primer (2014). Tech. rep., W3C. https://www.w3.org/TR/rdf11-
primer/
16. Rongen, S., Nikolova, N., van der Pas, M.: Modelling with AAS and RDF in industry
4.0. Comput. Ind. 148, 103910 (2023). https://doi.org/https://doi.org/10.1016/j.compind.2023.
103910, https://www.sciencedirect.com/science/article/pii/S016636152300060X
230 N. Nikolova et al.
17. Toussaint, M., Krima, S., Feeney, A.B., et al.: Requirement elicitation for adaptive
standards development. IFAC-PapersOnLine 54(1), 863–868 (2021). https://doi.org/
https://doi.org/10.1016/j.ifacol.2021.08.101, https://www.sciencedirect.com/science/article/
pii/S2405896321008508. 17th IFAC Symposium on Information Control Problems in
Manufacturing INCOM 2021
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
I4.0 Holonic Multi-agent Testbed
Enabling Shared Production
1 Introduction
that can be rapidly provisioned and released with minimal management effort or
interaction between service operators and providers” [22].
SmartFactoryKL shares their vision with the terminology Production Level 4
and Shared Production (SP) [2] that represents reconfigurable supply chains in a
federation of trusted partners that dynamically share manufacturing services over
a product lifecycle through standardized Digital Twins. Hence, it represents an
extension of Cloud Manufacturing.
This vision requires sharing data on products and resources. Production data
contains sensitive information, leading to fear of data misuse. Thus, additionally
to interoperability and standardization, data sovereignty is one major aspect when
building a manufacturing ecosystem. In [15], a technology-dependent solution is
presented. The authors propose using Asset Administration Shells (AAS) and the
concepts of Gaia-X to exchange data in a self-sovereign interoperable way. In this
sense, AAS aims to implement a vendor-independent Digital Twin [10], while Gaia-
X creates a data infrastructure meeting the highest standards in terms of digital
sovereignty to make data and services available. Key elements are compliance with
European Values regarding European Data Protection, transparency and trustability
[3]. Nevertheless, the authors of [15] focus more on the business layer and the data
exchange process rather than intelligent decision-making processes. Due to their
characteristics, Multi-Agent System (MAS) seems to be a promising natural choice
for implementing the logic and interactions among different entities.
In our contribution, we present the structure of MAS for a modern production
system to cope with the upper mentioned challenges. We focus on the intra-
organizational resilience of the shop-floor layer to provide the necessary flexibility
to enable an SP scenario. In Sect. 2, the considered concepts for the usage of MAS
in industrial applications are described. Section 3 describes the architecture of the
applied MAS and Sect. 4 gives an overview about the characteristics of a plant
specific holonic MAS and implements a prototype in the demonstrator testbed at
SmartFactoryKL .
from the specific implementation with the aim of increasing the flexibility. As
visualized in Fig. 1, CPPMs can be combined into a CPPS to perform control tasks
and react to information independently. CPPS connects the physical and digital
worlds and react to information from the environment and environmental influences
[8].
CPPMs require a need for a vendor-independent self-description to perform
production planning and control. AAS represents an approach to achieve this
standardized digital representation, where submodels (SM) are used to describe
domain-specific knowledge [10]. The Plattform Industrie 4.0 proposes an infor-
mation model composing the concepts of capabilities, skills and services as
machine-readable description of manufacturing functions to foster adaptive produc-
tion of mass-customizable products, product variability, decreasing batch size, and
planning efficiency. The Capability can be seen as an “implementation-independent
specification of a function [.. . .] to achieve an effect in the physical or virtual world”
[25]. Capabilities are meant to be implemented as skills and offered as services in a
broader supply chain. From the business layer, the service represents a “description
of the commercial aspects and means of provision of offered capabilities” [25].
From the control layer, Plattform Industrie 4.0 defines a Production Skill as
“executable implementation of an encapsulated (automation) function specified
by a capability” [25] that provides standardized interfaces and the means for
parametrization to support their composition and reuse in a wide range of scenarios.
The skill interface is mostly realized with OPC UA, since it has proven itself in
automation technology. Topics such as OPC UA and AAS can lead to confusion
regarding the separation of concerns. AAS is used as linkage to the connected
world and lifecycle management in adherence to yellow pages, whereas OPC UA is
applied for operative usage.
I4.0 Holonic Multi-agent Testbed Enabling Shared Production 235
The study of MAS began within the field of Distributed Artificial Intelligence. It
investigates the global behavior based on the agent’s fixed behavior. The studies
compromise coordination and distribution of knowledge. In this context, Leitão
and Karnouskos define an agent as “[. . .] an autonomous, problem-solving, and
goal-driven computational entity with social abilities that is capable of effective,
maybe even proactive, behavior in an open and dynamic environment in the
sense that it is observing and acting upon it in order to achieve its goals” [20].
MAS is a federation of (semi-)autonomous problem solvers that cooperate to
achieve their individual, as well as global system’s goals. To succeed, they rely on
communication, collaboration, negotiation, and responsibility delegation [20]. MAS
was motivated by subjects like autonomy and cooperation as a general software
technology, while the emergence in the manufacturing domain has been growing
recently.
The holonic concept was proposed by Koestler to describe natural beings that
consist of semi-autonomous sub-wholes that are interconnected to form a whole
[16]. Holonic Manufacturing System (HMS) is a manufacturing paradigm proposed
at the beginning of the 1990s as an attempt to improve the ability of manufacturing
systems to deal with the evolution of products and make them more adaptable
to abnormal operating conditions [7]. Holonic production systems are fundamen-
tally described in the reference architecture PROSA, with the aim of providing
production systems with greater flexibility and reconfigurability [32]. A Holon
is an autonomous, intelligent, and cooperative building block of a manufacturing
system for transformation, transportation, storing and / or validating information
and physical objects [33]. As shown in Fig. 2, a Manufacturing Holon always has
an information processing part and often a physical processing part [4]. Holons
join holarchies that define the rules for interaction between them. Each Holon can
be simultaneously a part of several holarchies and as well as a holarchy itself. This
enables very complex and flexible control structures, also called flexible hierarchies.
It is important to note that the cooperation process also involves humans, who might
enter or exit the Holon’s context [4]. In summary, HMS can be seen as an analogy
to CPPS, where skills provide the control interface to the physical processing parts.
Physical Control
Physical
Processing
Physical Processing
Holonic MAS seems a promising pattern to wrap the factory’s granularity and
build complex and resilient systems. It is important to note that the technology
to communicate with a customer and other factories might change over time or
might differ for individual customers or factories. Consequently, to be technology-
independent, our MAS does not include an explicit connection technology. MAS
accumulates some of its resource capabilities into services and provides these
services to the external world following the principles of service-oriented systems
[9]. Besides, the holonic MAS supervises a production system to plan and execute
the production and connects a factory to an SP network. Inspired by the upper
mentioned concepts, a modern factory is represented in Fig. 3.
With reference to the described aspects, the system consists of three main
Holons as displayed in Fig. 3. The basic structure follows the general idea of
PROSA as described in Sect. 2. The difference between the presented MAS and
their architecture is three-fold. First, the tasks of the management of the products
are fulfilled by AAS instead of having a Product Holon in PROSA. A more detailed
discussion about this change is discussed in 5. Second, the management of orders by
Shared
Production
Network Service Product Resource CPPM
OPC UA
Holon Holon Holon
Skills
D illi
Drilling
Identification Identification
Identification
Required Service Capabilities
Assured Services
Production Plan Bill Of Material
Production Logs
the Order Holon is shifted from the Order Holon to the Product Holon. The reason
for the new name is that our Product Holon takes care of orders and connects to
Product AAS and thus encapsulates two tasks. Third, another Holon called Service
Holon is added as an external representation of the factory layer as additional use
case to PROSA which does not examine a SP scenario in detail and represents the
factory as centralized HMS which spawns Order Holons based on the requests.
However, we prefer to decouple the task of spawning Holons and the representation
of the factory to achieve a higher resilience and flexibility.
Besides these changes, all three Holons are equipped with their own AAS to
expose their self-description. This includes an SM Identification to identify the
Holon in the MAS and an SM towards the interfaces of the Holon to be able
to communicate with different communication technologies and to implement the
communication technology independence. In addition, the SM Topology of each
Holon describes the internal structure. Part of the structure are all aggregated Sub-
Holons of the Holon. This SM eases the initialization of each Holon, especially in
case of a restart.
In the following two subsections, we present details about the Service Holon in
Sect. 3.2 and about the Product Holon in Sect. 3.3. For the Resource Holon, we will
give a more detailed description in Sect. 4. The Resource Holon executes production
steps by controlling and managing all resources. Each resource is via OPC UA
connected to one Resource Holon, which controls its execution. For that purpose,
the Resource Holon uses the Resource AAS to store information about the resource.
This includes an SM Identification, the provided capabilities of the resource and the
SM Bill of Material. In addition, the Resource Holon deals with tasks like lifecycle
management of resources, human-machine interaction, resource monitoring and
handling multiple communication technologies between the resources.
The Service Holon displayed on the left side of Fig. 3 manages and provides the
services of the factory to the connected world. One example of a service is the
3D-printing service, which offers the capability of producing customized products
using a fused deposition modeling process. Besides offering services, the Service
Holon enables the MAS to process external tasks like ordering a specific product
or executing a provided service. Besides that, the Service Holon takes care of
the disposition of external services, products or substitutes. Therefore, the Service
Holon has the ability to communicate with the SP network via an asynchronous
event-based communication based on I4.0L. This implements VDI/VDE 2193
[34, 35] as the chosen communication standard. A more detailed use case for the
connection to SP networks is applied in [15].
In the context of external SP Networks, the Service Holon represents the interface
to the factory. For this reason, the Service Holon takes care of the Factory AAS. This
AAS uses a unique identifier, a name and other general information to identify the
240 A. T. Bernhard et al.
The Product Holon takes care of the production process and subdivides production
tasks into different subtasks. In this context, the holonic approach takes effect.
Each Holon is responsible for one task and spawns for every subtask a Sub-Holon.
Together, they display the production process in a tree-like manner. To handle these
incoming tasks or derived subtasks, each Holon triggers an internal execution at the
Resource Holon or requests an external execution from the Service Holon.
In the case of an internal execution, each Holon needs to check if the execution
is feasible. Therefore, the Product Holon first matches the capabilities given by the
resources to the desired capability to fulfill the (sub-)task. If both capabilities fit
together, a feasibility check on the resources is triggered to simulate if a resource
is able to perform the task under posed conditions (e.g., if the process can supply
the desired product quality and evaluate estimated time, costs and consumption).
After a successful feasibility check, the Product Holon spawns an AAS as Digital
Twin for the product. The Product AAS contains information towards the product
identification like a name and a unique identifier. If the AAS contains some subtasks,
which require an external execution, the Product AAS contains a description of
all required external services to execute the subtask. After starting the production
process, the Product Holon further controls the process by triggering production
steps or monitoring the current production state. To monitor the process, the Product
Holon updates the corresponding Product AAS by adding logging data to the
production log.
To illustrate the execution of the Product Holon, a model truck as sample product
is ordered via the Service Holon. The truck is assembled out of two different
semitrailers. The semitrailer_truck consists of a 3D-printed cabin pressed on a brick
chassis called cab_chassis. Similar to the semitrailer_truck, the other semitrailer is
built by mounting a 3D-printed or milled trailer onto a semitrailer_chassis. Figure 4
shows the corresponding product structure tree.
Each of the displayed components and component assemblies relates to a
production step to produce the respective component. First, the components need
to be manufactured, then the semitrailers are assembled from the components and at
the end, the trailer is mounted on the cab_chassis to assemble the full truck. For each
of the given components, an own Product AAS is spawned. For example, the truck
Holon spawns on the highest level of the Bill of Material two semitrailer Holons.
I4.0 Holonic Multi-agent Testbed Enabling Shared Production 241
Both Holons independently produce their related semitrailers and after completion,
they report back to the truck Holon, which then controls the production of the full
truck by controlling the assembly step of both semitrailers.
Holons schedule and perform the concrete tasks. Furthermore, Island Resource
Holons encapsulate the sum of all Sub-Holons, by providing proxy functionality.
SARL Holons consist of a collection of Behaviors and Holon Skills. Island Resource
Holons provide hierarchies and coordinate Sub-Holons, whereas CPPM Holons, as
the smallest entity, are connected to physical assets. Figure 5 summarizes Island
Resource Holon’s Behaviors and Skills. Each Island Resource Holon has an AAS
that describes the Holon’s properties, configures and parametrize the system. The
key aspects of AAS are to provide information on how to find a Holon, how a
Holon is constructed, and how the Holon’s interfaces are defined. This information
is available in the SM Topology and Interface (see Sect. 3.1) and is also provided
to Sub-Holons. The AAS Skill extracts the Holons’ information as well as the
environmental information. The Update Behavior is used to gather information
about the contained Holons’ states, update the topology when MAS changes,
and synchronize this information with AAS. Since the Island Resource Holon
manages Sub-Holons, the Lifecycle Skill allows to dynamically create, destroy
or reboot Holons in its own holonic context. As SARL Janus provides message
channels for all Holons in runtime, the communication with other Holons requires
an external communication interface. The Inter-Holon Behavior allows the external
communication to Holons in other runtimes, e.g., message exchange between the
Resource Holon and the Product Holon via the open standard communication
middleware Apache Kafka. For communication and understanding, an I4.0 Message
Skill supports accordance with a standardized message model according to the
I4.0 Holonic Multi-agent Testbed Enabling Shared Production 243
VDI/VDE 2193-1 [35]. Communication, collaboration and negotiation are the key
components for a successful process. Island Resource Holon responds to produc-
tion requests of the Product Holon, initialize negotiations and sends production
requests to Sub-Holons. In Negotiation Behavior, the Island Resource Holon verifies
incoming messages. Depending on the request, Island Resource Holon forces Sub-
Holons to follow a task or request the possible execution. The former is used
when static optimization is applied, i.e., a global schedule shall be executed. An
example of global scheduling is demonstrated in [13], in which we schedule value-
adding as well as non-value-adding processes. The latter implements the bidding
protocol to foster dynamic optimization. Negotiation behavior defines the duration
of auctions and chooses incoming offers o based on max operator, i.e., for n
incoming offers, the chosen offer calculates with oi = max (o1 , . . . , on ). Next to
software systems, Holons may interact with humans, intending a special treatment
in terms of prioritization. Therefore, Human Behavior considers human knowledge
and adjustments. We are not trying to accomplish fully automated plants and exclude
humans from production. Instead, we want to support human decisions to benefit
from experiences and intuition, as well as building factories for humans [36].
For CPPM Resource Holons some building blocks like the Asset Administration
Shell Skill, the I4.0 Message Skill and the Human Behavior overlap (see Fig. 6).
For interactions, the CPPM Resource Holon provides three Behaviors: Requirement
Check, Bidding and Neighbor. Requirement Check acts in loose adherence to a
method call to achieve control structures in a hierarchy. In case, Island Resource
Holon fosters execution, the CPPM Resource Holon verifies if it can follow the
call and starts or queues the task. In Bidding Behavior, the Bidding Skill is used
to calculate a bid in the range between 0 and 1. The bid determines the desire to
perform the job. This fosters dynamic optimization, while taking processing time,
changeover times, deadlines, availability and resource’s possible operations into
account. The calculation process is achieved using a Reinforcement Learning algo-
rithm. The basis of the Reinforcement Learning algorithm is described in [26], while
244 A. T. Bernhard et al.
a modified variant will be published in future work. The last interaction pattern
is the neighbor behavior. CPPMs sense their environment, thus, CPPM Resource
Holon can omit hierarchies and directly communicate with their physical neighbor
to perform a complex task in a collaborative way. An example is a Pick&Place
operation that usually requires the supply by transportation means. The CPPM
Resource Holon has additional functions regarding the control and monitoring
of Production Skills. The Execution Behavior builds an event-based sequence to
reliably execute a Production Skill. In this context, it represents pipelines to set
the Production Skill’s parameters, verify compliance of all preconditions, tracks the
execution state and manages postconditions. Therefore, the CPPM Resource Holon
uses the OPC UA Skill, which allows access to a Production Skill interface directly
deployed on the CPPM. Furthermore, CPPM Resource Holon has a Monitoring
Behavior, which is used to check relevant sensor data, track the system status and
update system-critical information. In the future, anomaly detection and supervision
will also be implemented in this behavior.
Produktionsinsel_SYLT Assembly
Module
Robot
Module
Conveyor
Module
3D-Printer
Module Quality
Module
Connector
Module
ExecutionStep
Response
Request
Request
Response
Response
Request
Request
Quality Control
Connector Holon Produktionsinsel_SYLT Resource Holon
Holon
GetProduct
Response
Response
Request
Response
Response
Request
Request
GetProduct Robot Assembly Printer
Transport Holon
Holon Holon Holon
5 Conclusion
This chapter presents a MAS approach in the manufacturing domain, which enables
a factory to control its resources, to define and manage products and to provide
services to other SP participants. The MAS is based on a holonic approach and is
subdivided into Holons, each taking care of one of these tasks. For us, a Holon
is treated as a special type of Agent with additional characteristics regarding
recursiveness and connection to the hardware. The MAS collaborates with and uses
modern Industry 4.0 technologies such as Production Skills, AAS or OPC UA.
The presented MAS is enrolled on a demonstrator testbed at SmartFactoryKL ,
which is part of an SP scenario to produce model trucks. We divide our manufac-
I4.0 Holonic Multi-agent Testbed Enabling Shared Production 247
turing system into three Holons. The Service Holon provides and retrieves services
from the connected world. The Product Holon deals with modular encapsulated
products to manage dependencies between individual parts and assemblies as
well as controlling the production process. The Resource Holon encapsulates the
layer of the production testbed and connects the virtual with the physical world.
To guarantee autonomy, our Resource Holons use descriptions of AAS to gain
knowledge about the environment and use CPPM’s Production Skills to perform an
effect in the physical world. We achieve a flexible and resilient system by providing
communication patterns that allow hierarchical and heterarchical modularization.
However, the current state of our MAS is subject to different limitations. This
means that it will be extended in the future to fulfill different other features and will
solve different topics. One topic is to put more emphasize on product’s lifecycle,
while providing more complex planning systems to extract product’s features, match
capabilities and trace tender criteria. Another extension is planned on the monitoring
system to embed a factory wide monitoring system to combine a supervision of
the production process, factory level information like assured services and resource
data. The last topic is to provide more generalized holonic patterns and give more
insights about the Service Holon and the Product Holon.
As a result, we want to compare our architecture to other MAS systems, with a
special focus on the applied technologies in our systems. One of these technologies
is AAS. In comparison to full agent-based solutions, we typically replace one Holon
(e.g., in PROSA the Product Holon). As a downside, this leads to more applied
technologies in the system (due to the different technology stack) and thus to a
more complex architecture. However, the AAS as manufacturing standard supports
the interoperability between other factories and a simple data exchange format.
Furthermore, we use one standardized data format to express all our knowledge
to ease the internal usage of data via a system-wide interface.
Besides AAS as data format, we use SARL as Agent Framework. SARL itself
is a domain-specific language, which leads to a couple of general advantages and
disadvantages, as explained in [6]. We want to take up some of the listed problems
and advantages and add a few more SARL language-specific arguments. First,
SARL is especially designed to build MAS and includes an own metamodel to
define the structure of a Holon. Besides that, SARL offers concepts to encapsulate
certain functionality in Behaviors and Skills and leads to a modular system. One
special feature of SARL is that Holons are able to control the lifecycle of other
Holons, which is quite close to our applied concept of the MAS. Although SARL is
functional suitable for us, SARL also has different disadvantages. For example, it is
hard to find a documentation and help in the community if SARL specific problems
occur. Unfortunately, developing SARL code is exhausting since general supported
development environments do not always react in our desired response time.
Another difference between our MAS concept and other MAS concepts is
the granularity of applied Holons. In many cases, each device (e.g., a robot or
even in a smaller granularity like a sensor) has an own Holon. In our approach,
a Holon connects to one CPPM, which encapsulates single resources like robot
arms or 3D-printers. In this approach, a Holon accesses each CPPM by calling
248 A. T. Bernhard et al.
their provided skills. Having Holons on the device level leads to more holonic
communication, and thus more resources and effort is required to handle the
holonic communication. Moreover, MAS does not need to operate in real-time to
perform actions without a delay. This is why, we decided to encapsulate internal
communication inside a CPPM and keep time-critical and safety-critical tasks in
the physical processing parts. Furthermore, Holons are independent of machine-
specific control technologies, which increase the flexibility of the system towards
resource-specific technologies. Finally, we want to mention that even for small
holonic MAS, communication quickly becomes complex and lacks transparency.
Using standardized technologies like OPC UA and AAS regains this transparency
and supports application in the factory.
Acknowledgments This work has been supported by the European Union’s Horizon 2020
research and innovation program under the grant agreement No 957204, the project MAS4AI
(Multi-agent Systems for Pervasive Artificial Intelligence for Assisting Humans in Modular
Production Environments).
References
1. Beheshti, R., Barmaki, R., Mozayani, N.: Negotiations in holonic multi-agent systems. Recent
Advances in Agent-Based Complex Automated Negotiation pp. 107–118 (2016). https://doi.
org/10.1007/978-3-319-30307-9_7
2. Bergweiler, S., Hamm, S., Hermann, J., Plociennik, C., Ruskowski, M., Wagner,
A.: Production Level 4” Der Weg zur zukunftssicheren und verlässlichen Produk-
tion (2022), https://smartfactory.de/wp-content/uploads/2022/05/SF_Whitepaper-Production-
Level-4_WEB.pdf, (Visited on 18.05.2023)
3. Braud, A., Fromentoux, G., Radier, B., Le Grand, O.: The road to European digital sovereignty
with Gaia-X and IDSA. IEEE network 35(2), 4–5 (2021). https://doi.org/10.1109/MNET.2021.
9387709
4. Christensen, J.: Holonic Manufacturing Systems: Initial Architecture and Standards Directions.
In: First European Conference on Holonic Manufacturing Systems, Hannover, Germany, 1
December 1994 (1994)
5. Feraud, M., Galland, S.: First comparison of SARL to other agent-programming languages
and frameworks. Procedia Computer Science 109, 1080–1085 (2017). https://doi.org/10.1016/
j.procs.2017.05.389
6. Fowler, M.: Domain-specific languages. Pearson Education (2010)
7. Giret, A., Botti, V.: Holons and agents. Journal of Intelligent Manufacturing 15(5), 645–659
(2004). https://doi.org/10.1023/B:JIMS.0000037714.56201.a3
8. Hermann, J., Rübel, P., Birtel, M., Mohr, F., Wagner, A., Ruskowski, M.: Self-description
of cyber-physical production modules for a product-driven manufacturing system. Procedia
manufacturing 38, 291–298 (2019). https://doi.org/10.1016/j.promfg.2020.01.038
9. Huhns, M.N., Singh, M.P.: Service-oriented computing: Key concepts and principles. IEEE
Internet computing 9(1), 75–81 (2005). https://doi.org/10.1109/MIC.2005.21
10. Industrial Digital Twin Association: Specification of the Asset Administration Shell Part
1: Metamodel, https://industrialdigitaltwin.org/wp-content/uploads/2023/06/IDTA-01001-3-
0_SpecificationAssetAdministrationShell_Part1_Metamodel.pdf, (Visited on 28.06.2023)
11. Industrial Digital Twin Association: Specification of the Asset Administration Shell Part
2: Application Programming Interface, https://industrialdigitaltwin.org/wp-content/uploads/
I4.0 Holonic Multi-agent Testbed Enabling Shared Production 249
2023/06/IDTA-01002-3-0_SpecificationAssetAdministrationShell_Part2_API_.pdf, (Visited
on 28.06.2023)
12. Jungbluth, S., Barth, T., Nußbaum, J., Hermann, J., Ruskowski, M.: Developing a skill-based
flexible transport system using OPC UA. at-Automatisierungstechnik 71(2), 163–175 (2023).
https://doi.org/10.1515/auto-2022-0115
13. Jungbluth, S., Gafur, N., Popper, J., Yfantis, V., Ruskowski, M.: Reinforcement Learning-
based Scheduling of a Job-Shop Process with Distributedly Controlled Robotic Manipulators
for Transport Operations. IFAC-PapersOnLine 55(2), 156–162 (2022). https://doi.org/10.1016/
j.ifacol.2022.04.186, https://www.sciencedirect.com/science/article/pii/S2405896322001872,
14th IFAC Workshop on Intelligent Manufacturing Systems IMS 2022
14. Jungbluth, S., Hermann, J., Motsch, W., Pourjafarian, M., Sidorenko, A., Volkmann, M.,
Zoltner, K., Plociennik, C., Ruskowski, M.: Dynamic Replanning Using Multi-agent Systems
and Asset Administration Shells. In: 2022 IEEE 27th International Conference on Emerg-
ing Technologies and Factory Automation (ETFA). pp. 1–8 (2022). https://doi.org/10.1109/
ETFA52439.2022.9921716
15. Jungbluth, S., Witton, A., Hermann, J., Ruskowski, M.: Architecture for Shared Production
Leveraging Asset Administration Shell and Gaia-X (2023), unpublished
16. Koestler, A.: The ghost in the machine. Macmillan (1968)
17. Kolberg, D., Hermann, J., Mohr, F., Bertelsmeier, F., Engler, F., Franken, R., Kiradjiev, P.,
Pfeifer, M., Richter, D., Salleem, M., et al.: SmartFactoryKL System Architecture for Industrie
4.0 Production Plants. SmartFactoryKL, Whitepaper SF-1.2 4 (2018)
18. Kuhn, T., Schnicke, F.: BaSyx, https://wiki.eclipse.org/BaSyx, (Visited on 07.07.2022)
19. Leitão, P., Colombo, A.W., Restivo, F.J.: ADACOR: A collaborative production automation
and control architecture. IEEE Intelligent Systems 20(1), 58–66 (2005). https://doi.org/10.
1109/MIS.2005.2
20. Leitao, P., Karnouskos, S. (eds.): Industrial Agents: Emerging Applications of Software Agents
in Industry. Kaufmann, Morgan, Boston (2015). https://doi.org/10.1016/C2013-0-15269-5
21. Leon, F., Paprzycki, M., Ganzha, M.: A review of agent platforms. Multi-paradigm Modelling
for Cyber-Physical Systems (MPM4CPS), ICT COST Action IC1404 pp. 1–15 (2015)
22. Liu, Y., Wang, L., Wang, X.V., Xu, X., Jiang, P.: Cloud manufacturing: key issues and future
perspectives. International Journal of Computer Integrated Manufacturing 32(9), 858–874
(2019). https://doi.org/10.1080/0951192X.2019.1639217
23. Neubauer, M., Reiff, C., Walker, M., Oechsle, S., Lechler, A., Verl, A.: Cloud-based evalu-
ation platform for software-defined manufacturing: Cloud-basierte Evaluierungsplattform für
Software-defined Manufacturing. at-Automatisierungstechnik 71(5), 351–363 (2023). https://
doi.org/10.1515/auto-2022-0137
24. Pal, C.V., Leon, F., Paprzycki, M., Ganzha, M.: A review of platforms for the development of
agent systems. arXiv preprint arXiv:2007.08961 (2020). https://doi.org/10.48550/arXiv.2007.
08961
25. Plattform Industrie 4.0: Information Model for Capabilities, Skills & Services, https://www.
plattform-i40.de/IP/Redaktion/EN/Downloads/Publikation/CapabilitiesSkillsServices.html,
(Visited on 06.06.2023)
26. Popper, J., Ruskowski, M.: Using Multi-agent Deep Reinforcement Learning for Flexible Job
Shop Scheduling Problems. Procedia CIRP 112, 63–67 (2022). https://doi.org/10.1016/j.procir.
2022.09.039
27. Rodriguez, S., Gaud, N., Galland, S.: SARL: a general-purpose agent-oriented programming
language. In: 2014 IEEE/WIC/ACM International Joint Conferences on Web Intelligence (WI)
and Intelligent Agent Technologies (IAT). vol. 3, pp. 103–110. IEEE (2014). https://doi.org/
10.1109/WI-IAT.2014.156
28. Ruskowski, M., Herget, A., Hermann, J., Motsch, W., Pahlevannejad, P., Sidorenko, A.,
Bergweiler, S., David, A., Plociennik, C., Popper, J., et al.: Production bots für production
level 4: Skill-basierte systeme für die produktion der zukunft. atp magazin 62(9), 62–71 (2020).
https://doi.org/10.17560/atp.v62i9.2505
250 A. T. Bernhard et al.
29. Smith, R.G.: The contract net protocol: High-level communication and control in a distributed
problem solver. IEEE Transactions on computers 29(12), 1104–1113 (1980). https://doi.org/
10.1109/TC.1980.1675516
30. Trunzer, E., Calà, A., Leitão, P., Gepp, M., Kinghorst, J., Lüder, A., Schauerte, H., Reiffer-
scheid, M., Vogel-Heuser, B.: System architectures for Industrie 4.0 applications: Derivation
of a generic architecture proposal. Production Engineering 13, 247–257 (2019). https://doi.org/
10.1007/s11740-019-00902-6
31. Valckenaers, P.: Perspective on holonic manufacturing systems: PROSA becomes ARTI.
Computers in Industry 120, 103226 (2020). https://doi.org/10.1016/j.compind.2020.103226
32. Van Brussel, H., Wyns, J., Valckenaers, P., Bongaerts, L., Peeters, P.: Reference archi-
tecture for holonic manufacturing systems: PROSA. Computers in Industry 37(3), 255–
274 (1998). https://doi.org/10.1016/S0166-3615(98)00102-X, https://www.sciencedirect.com/
science/article/pii/S016636159800102X
33. Van Leeuwen, E., Norrie, D.: Holons and holarchies. Manufacturing Engineer 76(2), 86–88
(1997). https://doi.org/10.1049/me:19970203
34. VDI/VDE-Gesellschaft Mess- und Automatisierungstechnik: Language for I4.0 components
- Interaction protocol for bidding procedures, https://www.vdi.de/en/home/vdi-standards/
details/vdivde-2193-blatt-2-language-for-i40-components-interaction-protocol-for-bidding-
procedures, (Visited on 01.06.2023)
35. VDI/VDE-Gesellschaft Mess- und Automatisierungstechnik: Language for I4.0 Components
- Structure of messages, https://www.vdi.de/en/home/vdi-standards/details/vdivde-2193-blatt-
1-language-for-i40-components-structure-of-messages, (Visited on 01.06.2023)
36. Zuehlke, D.: SmartFactory–Towards a factory-of-things. Annual reviews in control 34(1), 129–
138 (2010). https://doi.org/10.1016/j.arcontrol.2010.02.008
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
A Multi-intelligent Agent Solution in the
Automotive Component–Manufacturing
Industry
1 Introduction
2. Raw material suppliers, their characteristics, their arrival times, their quantities,
etc. are usually stored in the suppliers’ departmental repositories.
3. Nominal product characteristics can be found in the engineering department’s
drawings and repositories.
4. Quality-control results are usually stored either in the quality-control machine’s
memory or in the quality-control department’s repositories.
This paper suggests that an improvement to production must balance various
choices, not only technical but also economic. In existing production environments,
several criteria are included and considered as part of determining the best solution
to problems like storage size (e.g., economic costs, logistic costs), operator costs,
production time, energy consumption, and more. When applying artificial intel-
ligence (AI) in the manufacturing process, the criteria should be similar: Several
agents with different goals should interact to determine the most holistic solution.
To this end, this paper illustrates that the multiagent method, based on a
framework of distributed examples, leads to interactions between these agents in
ways that improve the entire plant production system. Although the agents could
work in isolation, the MAS4AI (the multiagent system for AI) is tasked with making
them work together seamlessly and providing a better solution than they would
produce by working separately.
The second goal is to address scalability in the implementation of AI solutions
in the manufacturing industry: Agents need to be customized and specific to each
manufacturing process (i.e., grinding) yet at the same time be generic enough for
all machines and industries using this manufacturing process (i.e., all grinding
machines in all shop floors). The MAS4AI proposal is for an agent-based Asset
Administration Shell (AAS) approach to be deployed in every industry that wants
to use a particular agent. In this way, an agent is composed of agent logic and the
AAS. See the specifics for a case of using it at the Fersa plant, as depicted in Fig. 2.
A Multi-intelligent Agent Solution in the Automotive Component... 253
2 Experimental Development
The challenge was that the Fersa use case generated a wide variety of heterogeneous
data from different data sources. The decisions must take the following factors,
among others, into consideration:
• Raw material characteristics
• BEAIN12 and RIFA6 machine parameters and historical data
• Quality-control stations’ data
• Assembly stations’ data
• Sensors’ data
• Scheduling plans
These data sources also appear in different types of formats:
• Portable data formats (PDFs)
• Comma-separated values (CSVs)
• Tabular data formats such as Extensible Markup Language (XML) and Excel
(XLS)
On top of that, the data come from various locations:
• Shop floors (e.g., machine parameters)
• Suppliers (e.g., raw material characteristics)
• Management tools (e.g., ERPs and scheduling tools)
The first step in implementing Fersa’s MAS4AI is to ensure that the data are
properly connected, aggregated, and filtered.
Figure 5 presents the proposed solution to aggregating all the data sources: a—
machine data, including ranges and grinding-wheel parameters; b—raw material
data, including nominal and real data; c—other agents’ data; and d—assembly and
control stations’ data. The central part shows the repositories where the aggregation
takes place in MAS4AI for Fersa.
One of the main challenges is to find the most suitable griding wheel (GW) for a
certain production process (reference and quantity). On one hand, a too-big GW
will not be completely used; it will return to the warehouse as partially worn and be
difficult to reuse. On the other hand, a too-small GW will stop functioning before
the production cycle has finished, forcing a GW change, which stops production.
Our solution is to have the tooling agent check the Fersa GW warehouse in order
to select the most suitable GW for the current production cycle (see selection criteria
in 2.4). Most of the information needed by the agent is provided by the grinding
256 L. Usatorre et al.
machine agent, which receives and optimizes the machine parameters of the current
production cycle. Figure 6 shows the current workflow of this part of the Fersa
system.
The reference for the production order and the suggested machine parameters for
that production should be followed to determine the optimal grinding wheels. The
length of change time for the GWs should always be minimized. The AAS type for
the Fersa system was the resource one.
The main content of this agent can be found in the AAS of the GW agent,
specifically in the submodel’s production information. Figure 7 illustrates the
reference for the production, the inputs used for the execution of the agent, and
the list of GWs obtained for the tooling agent ASS.
A Multi-intelligent Agent Solution in the Automotive Component... 257
Wheel (GW) diameter and positions have been fully developed and are in a testing
phase, the optimization model for the initial parameters is still undergoing testing
because it still needs historical data for past batch references that were not collected
by Fersa before the project started. For the same reason, the model that predicts
deviations from the machine parameters for a fixed batch is also still undergoing
testing. Data are currently being collected to enable new models to be added and
tested over the following months.
The main content of this agent can be found in the machine agent’s AAS, specif-
ically in the submodel’s production information. Figure 8 presents the reference
A Multi-intelligent Agent Solution in the Automotive Component... 259
for the production, the inputs used for the execution of the agent, and the outputs
obtained for the machine agent ASS.
In this case, the challenge is to determine the time required to produce a certain batch
by factoring in the material, the tooling, and the machine parameters. Additional
parameters to be considered include the number of wheel changes needed and
the estimated time to complete these changes. The proposed solution considers
all these inputs. Such inputs refer to item information: the item number and the
catalog reference for the batch, the number of pieces to make (included in the
production table from the Microsoft SQL Server (MSSQL) repository), Fersa’s
current inventory, the batch schedule, and information on the production once
the item has entered the machine. The agent applies two logics: one in real time
featuring the current status of the machine and another that was calculated before
production. The output of the agent is based on the estimated time required to finish
a production. The AAS for preschedule agent is the production-planning type.
The main content of this agent can be found in the prescheduling agent’s AAS,
in the submodel’s production information. Figure 9 presents the reference of the
production order, the inputs used for the execution of the agent (e.g., the cone
reference and the quantity), and the outputs obtained (e.g., the number of GW
changes or the estimated time required to produce the batch) for the prescheduling
agent ASS.
2.5 Holon
3 Conclusion
This paper presents the ontology, semantics, and data architecture that permits
multiagent interaction, and the RAMI 4.0 model was selected as the basis for
designing and implementing the presented approach.
The presented data architecture, depicted in Fig 11, permits the data analysis
of raw materials, finished products, tooling characteristics and statuses, machine
parameters, and external conditions, to minimize the influence of intuition and
262 L. Usatorre et al.
Acknowledgments This work received funding from the European Union’s Horizon 2020
research and innovation program under grant agreement number 957204 (H2020-ICT-38-2020)
(MAS4AI).
References
1. Sakurada, L., Leitao, P., De la Prieta, F.: Agent-based asset administration shell approach for
digitizing industrial assets. IFAC-PapersOnLine. 55(2), 193–198, ISSN 2405-8963, (2022).
https://doi.org/10.1016/j.ifacol.2022.04.192
2. Galland S, Gaud N, Rodriguez S, et al.: Janus: another yet general-purpose multiagent platform.
In: Proceedings of 7th Agent-Oriented Software Engineering Technical Forum (TFGASOSE-
10); 2010 Dec; Paris, France
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Integrating Knowledge into
Conversational Agents for Worker
Upskilling
1 Introduction
The supply and demand for labor, wherein employees give the supply and employers
provide the demand, are referred to as the labor market or job market. They are a
crucial part of every economy and are closely connected to the markets for capital,
goods, and services. According to a recent International Labour Organization (ILO)
R. Alonso ()
R2M Solution s.r.l., Pavia, Italy
Programa de Doctorado, Centro de Automática y Robótica, Universidad Politécnica de
Madrid-CSIC, Madrid, Spain
e-mail: [email protected]
D. Dessí
Knowledge Technologies for the Social Sciences Department, GESIS – Leibniz Institute for the
Social Sciences, Cologne, Germany
e-mail: [email protected]
A. Meloni · M. Murgia
Mathematics and Computer Science Department, University of Cagliari, Cagliari, Italy
e-mail: [email protected]; [email protected]
R. R. Diego
Mathematics and Computer Science Department, University of Cagliari, Cagliari, Italy
R2M Solution s.r.l., Pavia, Italy
e-mail: [email protected]; [email protected]
assessment1 , the present global economic slump is likely to drive more employees
to take poorer quality, poorly paid employment that lack job security and social
safety, exasperating disparities that have already been made worse by the COVID-
19 crisis. According to the same report, only 1.0% of the world’s employment is
expected to expand in 2023, which is less than half of the rate in 2022. The number
of unemployed people worldwide is projected to increase by almost 3 million to
208 million in 2023, representing a 5.8% global unemployment rate. The limited
labor supply in high-income nations is partly to blame for the moderate extent of
this predicted increase. This would reverse the drop in unemployment seen globally
between 2020 and 2022.
The COVID-19 epidemic has, therefore, recently had a significant effect on
the world labor market. Additionally, new difficulties have arisen that are also
negatively affecting the workplace, such as rapid increases in inflation, disruptions
in the supply chain, and the conflict in Ukraine.2 It is vital to take action to advance
social justice by tackling issues like young employment, informality, productivity,
and gender parity. To create a lasting and inclusive recovery that leaves no one
behind, policymakers, companies, employees, and civil society must collaborate.
This entails spending money on education and training, enhancing social safety nets,
encouraging good work, and moving forward.
The industrial sector, and in particular the manufacturing sector, is not immune
to this situation, since it is a very competitive labor market in which it is difficult
to recruit top talents that are experienced with new technologies such as robotics
or AI. The sector also suffers the fast pace of innovative technologies, making it
difficult for both companies and employees, without adequate training and practice,
to be up-to-date and competitive. The manufacturing sector needs to retain talent
and adapt to new needs, through professional development activities and investment
in employee growth, and at the same time be able to find quick solutions, such as
cross-training, to be able to cover leaves of absence or production peaks. This calls
for new training, retention, and recruitment strategies.
Since 2021, several industries and jobs have been predicted to grow by Career-
Builder,3 a 1995-founded employment website with operations in 23 countries and
a presence in over 60 markets. CareerBuilder offers labor market information,
talent management software, and other recruitment-related services. The listed
employment spans a variety of industries, including leisure, arts, renewable energy,
personal services, healthcare, and information technology. It is possible to upload
and build a resume and look for one of many jobs posted by different organizations.
1 https://www.ilo.org/wcmsp5/groups/public/---dgreports/---inst/documents/publication/
wcms_865332.pdf
2 https://ilostat.ilo.org/assessing-the-current-state-of-the-global-labour-market-implications-for-
achieving-the-global-goals/.
3 https://www.careerbuilder.com/.
Integrating Knowledge into Conversational Agents for Worker Upskilling 267
4 https://www.monster.com/.
5 https://www.offerzen.com.
6 https://www.linkedin.com/.
7 https://www.glassdoor.com/.
8 https://www.jobstreet.com/.
9 https://www.ziprecruiter.co.uk.
10 https://www.dice.com/.
11 https://www.g2.com/deals.
12 https://www.indeed.com/employers/hiring-platform.
13 https://hired.com/.
14 https://cord.co/.
15 https://circaworks.com/.
16 https://www.naukri.com/.
268 R. Alonso et al.
2 Related Work
In this section, we will describe some state-of-the-art conversational agents that can
be employed, among others, also within the labor domain.
ChatGPT Released in November 2022, ChatGPT20 is an artificial intelligence (AI)
chatbot created by OpenAI. It is a form of generative AI that allows users to enter
prompts to receive humanlike images, text, and videos created by the AI. It is
constructed on top of the foundational large language models (LLMs) GPT-3.5 and
GPT-4 and has been fine-tuned utilizing both supervised and reinforcement learning
strategies. OpenAI claims that “Reinforcement Learning from Human Feedback”
(RLHF) was used to train ChatGPT. The model initially underwent supervised fine-
tuning, in which OpenAI trainers acted as both a human user and an AI bot. To
18 ISTAT: https://www.istat.it/en/.
19 https://www.dbpedia.org/.
20 https://openai.com/blog/chatgpt.
270 R. Alonso et al.
fine-tune the model for conversational usage, the trainers used this to develop a
dialogue sequence that mimicked how people converse in real life. Later, ChatGPT
was enhanced by developing a reward model to be used for the following phase of
reinforcing learning. In order to produce responses, this entailed AI trainers engaged
with the tool. Afterward, the responses were ranked according to their quality.
With this information, there was a further fine-tuning phase called Proximal Policy
Optimization. ChatGPT employs deep learning algorithms to assess incoming text
prompts and produce responses based on patterns in the data it has been trained on.
It can comprehend subtle language differences and deliver responses of the highest
caliber because it has been trained on a huge corpus of literature, including books,
papers, and websites. Users can give the chatbot feedback by clicking the “thumbs
up” or “thumbs down” icons next to each response in order to help it learn. Users can
also offer more textual comments to enhance and perfect upcoming dialogue. Users
can ask a wide range of topics on ChatGPT, from straightforward inquiries to more
difficult ones like, “What is the meaning of life?”. ChatGPT is skilled in STEM
fields and has the ability to troubleshoot or write code. There is no restriction on the
kinds of queries that can be asked to ChatGPT. ChatGPT uses data only through
2021; therefore, it is unaware of events and data after that point. Additionally,
because it is a conversational chatbot, users can request additional details or ask
that it try again when producing content.
A list of limitations of ChatGPT is described in the following. It does not properly
comprehend how intricate human language is. Words are generated using ChatGPT
based on input. As a result, comments could come off as superficial and lacking
in profundity. Moreover, ChatGPT could respond incorrectly if it does not fully
comprehend the question. On top of that, responses may come off as artificial
and robotic. The training data covers up to 2021 and, therefore, ChatGPT has no
knowledge of what happened later. As ChatGPT is still being trained, providing
comments when a response is erroneous is advised. ChatGPT may misuse terms
like “the” or “and.” Due to this, information must still be reviewed and edited by
humans in order to make it read more naturally and sound more like human writing.
ChatGPT does not cite sources and does not offer interpretation or analysis of any
data or statistics. It is unable to comprehend irony and sarcasm. It can focus on the
incorrect portion of a question and be unable to shift. For instance, if we ask, “Does
a horse make a good pet based on its size?” and then we ask “What about a cat?”.
Instead of providing information regarding keeping the animal as a pet, ChatGPT
might only concentrate on the animal’s size. Because ChatGPT is not divergent, it
cannot shift its response to address several questions in one.
Bing It is a chatbot created by Microsoft. Bing is integrated with Microsoft’s
Bing search engine and is made to respond to users’ inquiries in a way that is
specific to their search history and preferences.21 The Bing AI chatbot may help
with a variety of tasks, including question answering, providing recommendations,
21 https://www.bing.com/?/ai.
Integrating Knowledge into Conversational Agents for Worker Upskilling 271
22 https://blogs.bing.com/search/february-2023/The-new-Bing-Edge-%E2%80%93-Learning-
from-our-first-week.
23 https://writesonic.com/chat.
272 R. Alonso et al.
Twitter, etc. In the context of worker upskilling, it can be used to find online courses
that can be attended to learn or improve specific skills. One of its main limitations
is that each user has a budget of 10k words, and when finished, a premium account
needs to be activated.
Copilot Copilot24 is a language model that slightly differs from the others because
it targets computer scientists and developers and their daily task of writing new
source code. In fact, this model is trained on billions of lines of code. It allows
one to write code faster, suggests more lines of code, proposes implementations
of methods, and enables developers to focus on the overall approach instead of
dispersing energies on trivial tasks. However, this tool is not perfect and it has
several limitations; it often delivers code and method implementations that do not
make sense and, additionally, it does not test the code that is written. Therefore,
the developer has to verify that the automatically generated code is functional,
as well as the quality of the delivered results. This is a crucial aspect for the
accountability of the work, and for ensuring the quality of the developed software
to customers. Nevertheless, Copilot is an interesting tool for the labor domain
considering computer science-related areas and might be a plus in a large variety
of companies in which software is mostly used for simple tasks.
ELSA-Speak English Language Speech Assistant25 (ELSA) is an application
based on artificial intelligence and speech recognition technologies designed to
support users in learning English. It can be used to improve English pronunciation
and provides lessons and exercises. The application provides real-time feedback to
the users to enhance their speaking skills. ELSA-Speak can be employed within the
labor market to prepare for job interviews (for example, candidates might be more
confident to have an interview in English), to help non-native English speakers in
their work, and to improve their English skills in order to increase opportunities for
career advancement.
For some time now, different initiatives have been presented to help students, the
workforce, and companies to meet the needs of the labor market. All of them
have in common that they are initiatives supported by government agencies, are
updated periodically, and offer the data in a public and open way. In this section,
we will focus on three of the most relevant databases and information systems on
occupations and competencies.
24 https://github.com/features/copilot.
25 https://elsaspeak.com/en/.
Integrating Knowledge into Conversational Agents for Worker Upskilling 273
30 https://noc.esdc.gc.ca/Oasis/OasisWelcome.
31 https://www.statcan.gc.ca/en/subjects/standard/noc/2021/introductionV1.
Integrating Knowledge into Conversational Agents for Worker Upskilling 275
5 Proposed Solution
The goal of the solution we have proposed in recent papers [1–3] is to make available
to users, who are not always technologically prepared, the information contained
within any of the resources described in Sect. 4 in the form of natural language and
to help them analyze which skills and personal knowledge are most useful for their
application in any job.
One of the worker upskilling resources covered in Sect. 4, the O*NET database,
contains many types of information. Our system uses the titles of the principal
job categories that cover all the jobs in the North American job market, their
descriptions, and the alternative titles for each one. Abilities, Skills, Knowledge,
Work Activities, Tasks, Technology Skills, and Tools are the information that
characterizes each job category. The number of elements for each job is constant
for the first four categories and variable for the last three. The items in the first five
present a score relative to their importance to the job they refer to, while the last two
are lists of tools or technologies commonly used in each job category.
More in detail, the system we propose performs the following tasks:
1. Determine how much the user’s resume is appropriate for the chosen job
category.
The system uses NLP techniques to analyze the user’s resume and to extract
its most important information. The database entities and the extracted content
from the CV are compared using semantic similarity. The resume score for a job
is the sum of the scores of the discovered elements divided by the maximum score
obtainable for the job itself and a corrective factor required to avoid penalizing
jobs with few items in the entities with a variable number of elements. The system
draws attention to the lacking aspects of the CV and suggests ways for users to
increase their knowledge of each one.
2. Determine the user’s suitability for the selected job.
The system prompts the user to enter his/her self-assessed level (none,
little, medium, good, excellent) in terms of Abilities, Knowledge, Skills, Work
Activities, and Tasks entities taken from the database and associated with the
selected job. The system returns a percentage score, where 60% is the suitability
threshold and a list of lacking knowledge or abilities that the user might improve
to raise his/her score.
3. Determine what occupation the user is best suited for.
The system prompts the user to enter his/her self-assessed level (none, little,
medium, good, excellent) in terms of Abilities, Knowledge, Skills, and Work
Activities entities taken from the database and returns a list with five suggested
job categories.
4. Perform conversation with the user.
To integrate the knowledge provided by O-NET into the conversational agent,
we first chose a reduced set of questions that it must be able to answer. The base
question templates are:
276 R. Alonso et al.
of personal data and in some cases in profiling. The topic related to the concerns
about user data’s inappropriate use in user-chatbot interactions is being researched
and surveyed [4, 8], and still there are several debates about it at the international
level (e.g., in April 2023 Italy was the first Western country to block ChatGPT32 ).
This calls for a critical analysis of the different uses of chatbots for upskilling and,
above all, an alignment with the initiatives for the maintenance of the privacy of
workers and users, from the General Data Protection Regulation to the Artificial
Intelligence Act.
Regarding benefits and impact, at the scientific and technical level, and as can be
seen in the state of the art, advances in conversational agents and language models
are clear. All these advances have an impact on the development of solutions such
as the one we present, for example, by providing better interaction capabilities or
even the possibility of answering open-domain questions. At the economic level,
as also discussed in the introduction, there is a demand for workers and a demand
for knowledge for specific activities, and upskilling agents can facilitate this task.
At the social level, it impacts workers and the unemployed, in terms of analyzing
their strengths, detecting training or internship needs, and adapting their CVs to fit
specific jobs.
From the worker’s perspective, conversational AI agents based on standardized
and reliable databases of occupational information have direct impacts and benefits.
On the one hand, these conversational solutions have a direct impact on workers’
knowledge of their work activity. The workers themselves can at any time obtain
reliable information about their job and learn about skills, activities, and knowledge
relevant to the task they are carrying out. Moreover, the worker can ask the
conversational agent which are the usual technical skills for a certain occupation
or activity to imagine where they might be oriented.
On the other hand, it is beneficial for their training and career. The employees can
detect points in which their profile would benefit from training and at the same time
detect training needs and create training plans to adapt to the needs of their position,
or to another position they plan to move to. This is one of the other benefits of these
conversational agents since the workers can compare themselves with professions
and positions in which they would be interested in changing or being promoted.
From a more organizational and business point of view, there are also a couple
of areas where AI-based conversational agents such as the ones mentioned in this
article can be beneficial. The part of assistance in hiring processes is a point in which
these agents will have a direct impact: from providing candidates with information
about the position, to pre-screening or determining the suitability of a candidate
by using conversational agents to compare the candidate’s CV with the baseline of
the occupational reference database. Conversational agents are also beneficial in the
detection of training needs for specific positions or the creation of group training
plans, based on the needs of multiple workers.
32 https://www.bbc.com/news/technology-65139406.
278 R. Alonso et al.
7 Conclusions
This chapter explores the potential uses, benefits, and limitations of conversa-
tional agents within the labor market. The chapter guides the reader through
the state-of-the-art research on this topic and presents which technologies are
currently employed. Furthermore, the chapter analyzes conversational agents that
are disrupting the labor market, where AI tools are expected to be employed
for a large variety of tasks such as learning new skills, supporting document
writing, code development, language skills improvement, and so on. Along with
these conversational agents, we also introduce the reader to valuable sources of
information which can be leveraged to build AI systems on top. Finally, we present
how potential solutions can be used for specific use cases and discuss how all these
technologies can impact the labor market.
As main lessons, we would like to mention the need for both the workforce and
companies to adapt to the rapidly changing world of new technologies and how
this can be beneficial, for example, by taking advantage of it to better understand
occupations, retain talent, hire, gain job knowledge, or adapt to new activities. It is
also important to consider the need to efficiently integrate large language models
and validated and reliable databases. Finally, it is essential to consider and analyze
the privacy implications of these solutions. It is possible to develop privacy-aware
conversational systems for the upskilling of workers, but there is a possibility that
certain solutions may abuse data. For this reason, it is critical to provide these
systems with consideration of Human Centricity aspects and all the factors that
compose it, including privacy.
In the medium term, it is expected that these technologies will be integrated
into the regular workflow of employees and will have a positive impact on
employee engagement and professional development. For this reason, we expect
that this chapter will foster several visions to the reader about the role of AI-based
conversational agents in the labor domain and worker upskilling which might bring
unprecedented development in the field.
Integrating Knowledge into Conversational Agents for Worker Upskilling 279
Acknowledgments This work has been partly supported by EU H2020 research and innovation
programme project STAR—Safe and Trusted Human Centric Artificial Intelligence in Future
Manufacturing Lines (Grant n. 956573).
We acknowledge financial support under the National Recovery and Resilience Plan (NRRP),
Mission 4 Component 2 Investment 1.5—Call for tender No.3277 published on December 30,
2021 by the Italian Ministry of University and Research (MUR) funded by the European Union—
NextGenerationEU. Project Code ECS0000038—Project Title eINS Ecosystem of Innovation
for Next Generation Sardinia—CUP F53C22000430001—Grant Assignment Decree No. 1056
adopted on June 23, 2022 by the Italian Ministry of University and Research (MUR)
References
1. Alonso, R., Dessí, D., Meloni, A., Recupero, D.R.: Incorporating knowledge about employa-
bility into conversational agent. Appl. Intelligence submitted (2023)
2. Alonso, R., Dessí, D., Meloni, A., Recupero, D.R.: A novel approach for job matching and skill
recommendation using transformers and the o*net database. Big Data Res. submitted (2023)
3. Alonso, R., Dessí, D., Meloni, A., Reforgiato Recupero, D.: A general and NLP-based archi-
tecture to perform recommendation: A use case for online job search and skills acquisition.
In: Proceedings of the 38th ACM/SIGAPP Symposium On Applied Computing. ACM Special
Interest Group on Applied Computing (SIGAPP) (2023)
4. Belen Saglam, R., Nurse, J.R.C., Hodges, D.: Privacy concerns in chatbot interactions: When
to trust and when to worry. In: Stephanidis, C., Antona, M., Ntoa, S. (eds.) HCI International
2021 - Posters, pp. 391–399. Springer International Publishing, Cham (2021)
5. Fadhil, A., Gabrielli, S.: Addressing challenges in promoting healthy lifestyles: the al-chatbot
approach. In: Proceedings of the 11th EAI International Conference on Pervasive Computing
Technologies for Healthcare, pp. 261–265 (2017)
6. Hoffmann, E.: International Statistical Comparisons of Occupational and Social Structures,
pp. 137–158. Springer US, Boston, MA (2003). https://doi.org/10.1007/978-1-4419-9186-7_
8
7. Hussain, S., Ameri Sianaki, O., Ababneh, N.: A survey on conversational agents/chatbots
classification and design techniques. In: Web, Artificial Intelligence and Network Applications:
Proceedings of the Workshops of the 33rd International Conference on Advanced Information
Networking and Applications (WAINA-2019) 33, pp. 946–956. Springer (2019)
8. Ischen, C., Araujo, T., Voorveld, H., van Noort, G., Smit, E.: Privacy concerns in chatbot
interactions. In: Følstad, A., Araujo, T., Papadopoulos, S., Law, E.L.C., Granmo, O.C., Luger,
E., Brandtzaeg, P.B. (eds.) Chatbot Research and Design, pp. 34–48. Springer International
Publishing, Cham (2020)
9. Lee, M., Ackermans, S., Van As, N., Chang, H., Lucas, E., IJsselsteijn, W.: Caring for Vincent:
a chatbot for self-compassion. In: Proceedings of the 2019 CHI Conference on Human Factors
in Computing Systems, pp. 1–13 (2019)
10. Lee, T., Zhu, T., Liu, S., Trac, L., Huang, Z., Chen, Y.: CASExplorer: A conversational
academic and career advisor for college students. In: The Ninth International Symposium
of Chinese CHI, pp. 112–116 (2021)
11. Maddigan, P., Susnjak, T.: Chat2vis: Generating data visualisations via natural language using
ChatGPT, codex and gpt-3 large language models. IEEE Access (2023)
12. Meloni, A., Angioni, S., Salatino, A., Osborne, F., Reforgiato Recupero, D., Motta, E.:
Integrating conversational agents and knowledge graphs within the scholarly domain. IEEE
Access 11, 22468–22489 (2023). https://doi.org/10.1109/ACCESS.2023.3253388
13. Oviedo-Trespalacios, O., Peden, A.E., Cole-Hunter, T., Costantini, A., Haghani, M., Kelly, S.,
Torkamaan, H., Tariq, A., Newton, J.D.A., Gallagher, T., et al.: The risks of using ChatGPT to
obtain common safety-related information and advice. Available at SSRN 4346827 (2023)
280 R. Alonso et al.
14. Reimers, N., Gurevych, I.: Sentence-Bert: Sentence embeddings using Siamese Bert-networks.
In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Process-
ing. Association for Computational Linguistics (2019). https://arxiv.org/abs/1908.10084
15. Singh, S., Beniwal, H.: A survey on near-human conversational agents. Journal of King
Saud University - Computer and Information Sciences 34(10, Part A), 8852–8866 (2022).
https://doi.org/10.1016/j.jksuci.2021.10.013. https://www.sciencedirect.com/science/article/
pii/S1319157821003001
16. Vo, N.N., Vu, Q.T., Vu, N.H., Vu, T.A., Mach, B.D., Xu, G.: Domain-specific NLP system to
support learning path and curriculum design at tech universities. Comput. Educ. Artif. Intell.
3, 100042 (2022)
17. Xu, A., Liu, Z., Guo, Y., Sinha, V., Akkiraju, R.: A new chatbot for customer service on social
media. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems,
pp. 3506–3510 (2017)
18. Zhou, M.X., Mark, G., Li, J., Yang, H.: Trusting virtual agents: The effect of personality. ACM
Trans. Interact. Intell. Syst. (TiiS) 9(2-3), 1–36 (2019)
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative Commons license and
indicate if changes were made.
The images or other third party material in this chapter are included in the chapter’s Creative
Commons license, unless indicated otherwise in a credit line to the material. If material is not
included in the chapter’s Creative Commons license and your intended use is not permitted by
statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder.
Advancing Networked Production
Through Decentralised Technical
Intelligence
1 Introduction