=Paper=
{{Paper
|id=Vol-3214/WS5Paper7
|storemode=property
|title=On Exploring the Possibilities and the Limits of AI for an Interoperable and Empowering Industry 4.0
|pdfUrl=https://ceur-ws.org/Vol-3214/WS5Paper7.pdf
|volume=Vol-3214
|authors=Francesco Lelli
|dblpUrl=https://dblp.org/rec/conf/iesa/Lelli22
}}
==On Exploring the Possibilities and the Limits of AI for an Interoperable and Empowering Industry 4.0==
On Exploring the Possibilities and the Limits of AI for an
Interoperable and Empowering Industry 4.0
Francesco Lelli1
1
Tilburg University, Warandelaan 2, Tilburg, The Netherlands
Abstract
This paper aims to raise awareness on certain interoperability issues as we intend to shape
industry 5.0 in order to enable a human-centric resilient society. We advocate that the need of
sharing small and specific data will become more intensive as AI-based solutions will
become more pervasive. Consequently, dataspaces should be carefully designed to address
this need. We advance the conversation by presenting a case study from HR demonstrating
how to predict the possibility of an employee experiencing attrition. Our experimental results
show that we need more than 500 samples for developing a machine learning model to be
sufficiently capable to generalize the problem. Consequently, our experimental results show
the feasibility of the idea. However, in small and medium sized companies this approach
cannot be implemented due to the limited number of samples. At the same time, we advocate
that this obstacle may be overcome if multiple companies will join a shared dataspace, thus
raising interoperability issues.
Keywords 1
Industry 4.0, Industry 5.0 interoperability, machine learning, AI, HR, attrition.
1. Introduction
Industry 4.0 envisions a human machine interaction that goes beyond the classical paradigms.
Humans supervise semi-autonomous interoperable and reconfigurable machines that participate to the
production line. In this way, humans become more and more part of a socio-technical production
system and develop an intricate relationship with their (smart) tools and (smart) products.
AI is playing a critical role in advancing this vision [1], however, the path is more intricate then
expected. Contrary to what is presented in [2], when we do not consider IoT devices, the so-called
“big data” in many contexts presents itself as a collection of islands of “small data”. These islands are
scattered, fragmented, and unusable alone. Consequently, this raises the need of advancing and
advocating for interoperability in several mini and micro contexts for sharing their data for creating
meaningful and actionable information.
This paper presents a concrete example of this situation by considering a case study of human
resources. We intend to investigate the possibility of an employee to experience attrition in his
working environment by using a dataset of 1470 cases. The information collected are typical HR data
such as number of years in the position, distance from work, job title etc. As we will see in section 3,
each single datapoint is uncorrelated with the possibility of an employee to experience attrition.
Consequently, it is impossible for a human to devise a meaningful heuristic. However, our results
(presented in section 4) show that it is possible to build a neural network that achieves 80+% accuracy
in a validation set and 100% in training set. At the same time, these kinds of solutions require a
relatively large amount of data and companies may not have 1000+ employees that cover similar roles
[3]. Consequently, a similar solution may not be applicable to simpler contexts such as, for example,
in SMEs. However, this problem may be overcome if we consider an entire industry vertical where
Proceedings of the Workshop of I-ESA’22, March 23–24, 2022, Valencia, Spain
EMAIL: F.Lelli@tilburguniversity.edu
ORCID: 0000-0003-1900-9171
© 2022 Copyright for this paper by its authors.
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR
Wor
Pr
ks
hop
oceedi
ngs
ht
I
tp:
//
ceur
-
SSN1613-
ws
.or
0073
g
CEUR Workshop Proceedings (CEUR-WS.org)
the participants will decide to share some of their information to the larger group. Let us consider, for
example, the case of a hotel: One single unit may not have a sufficient amount of employees for
applying a machine learning approach. However, a collection of hotels in a specific region may reach
a meaningful threshold number and therefore benefit of tools that leverage an AI model, like one
proposed in this paper, for supporting HR in conducting their everyday activities.
The rest of this paper is organized as follows: in section (ii) we will contextualize this research
presenting a few related works, in section (iii) we will introduce the dataset and the methodology used
for developing the predictor, in section (iv) we will analyze the results, and in section (v) we will
conclude by presenting the interoperability issues that we are currently facing in developing such
solutions.
2. Related work
According to authors in [4], Industry 4.0 has embraced AI technology as one of the key enablers.
At the same time, also predicting employee attrition using machine learning is not a novel idea and, in
some cases, raises also ethical concerns due to the possibility to inject AI-based biases in the decision
process [5]. This is not a show stop for research. A recent survey presented by authors in [6]
summarizes the work of 100 papers that are trying to propose AI techniques for solving HR issues.
Authors in [3] also report the successful integration of AI tools in HR practices in a large bank like
Credit Suisse.
The dataset that we are using in this paper has been publicly released by IBM [8] and a few
exploratory research have been already performed. This is the case of authors in [7] where they
suggest a simple ML approach achieving a recall rate of 0.54. The same data have been analyzed also
using deep learning techniques as presented by authors in [8]. In this specific case they have been able
to achieve a 91% accuracy.
In this paper, we present a neural network model, enhanced by genetic algorithm for finetuning the
model to the specific dataset. Our model achieves a better accuracy; however, it is not the point of this
paper. In this respect we intend to outline the importance of a proper dataset for developing such
tools. In addition, we consider the difficulties of implementing such tools without creating a shared
dataspace [9, 10] where companies engage in data sharing for developing win-win solutions.
3. Method
The core idea of this research is to develop a machine learning predictor that uses a dataset
containing typical information collected from HR. Clearly, the possibility to model the problem is
dependent by the quality and quantity of available information. Consequently, we can investigate the
learning capability of the model reducing the size of samples. In this way, we will be able to observe
the degradation of performances and define a threshold where the system becomes instable and
therefore useless for addressing the specific problem of predicting the possibility that an employee
will experience attrition. In the next subsection (3.1) we will present the dataset and in the consequent
subsection we will describe the selected ML approach.
3.1. Dataset
The dataset used in this study is a collection of 35 different datapoints from a sample of 1470
employees. It does not contain missing information and all the employees have been classified in
terms of experiencing attrition or not experiencing it. Figure 1 presents a normalized version of the
datapoints and gives an intuitive idea of the information collected.
Figure 1: Datapoints and their distribution
Figure 2: Correlation among datapoints of the dataset
In particular, a histogram of each variable is presented. During the normalization process we
removed 87 samples as their values were deviating more than 3 times from the norm (i.e. the
threshold was 3 times sigma).
At the same time, Figure 2 analyzes the correlations that each datapoint has with the others. The
whiter the color (as opposed to purple, see the column at the right), the more a datapoint is correlated
with the others. In addition, for our specific case, we are interested in predicting the datapoint
“attrition” that is represented in the second row of the matrix using the rest of the variables contained
in the dataset.
We can observe that none of the datapoints in the dataset is significantly correlated to the variable
that we would like to predict. This may explain why detecting employees that are experiencing
attrition is a particularly challenging task given the information that the usual HR department collects
regarding their employees. In the coming section, we will present a neural network model used in
conjunction with a genetic algorithm for finetuning the parameters of the model in order to best fit the
problem.
3.2. Methodology
Given the dataset presented in section 3.1 [11] we intend to create a neural network model that
receives a portion (70%) of the dataset as training set, a second portion (20%) for testing and the
remaining part (10%) for validating the final model to avoid overfitting cases. A neural network has a
collation of parameters that can influence their learning process. Example includes the number of
hidden layers, the number of neurons in each layer, the learning rate, batch size, etc. We decided to
use a genetic algorithm approach for finetuning these parameters and present the best model to the
final validation step. Figure 3 presents the idea in a graphical manner: The features of a particular
neural network are encoded in a particular chromosome and the best (most performing) member of a
set of models generates additional members of the population by exchanging the value of the given
features.
Figure 3: The model and its evolution
4. Results and discussion
Following the approach described in section 3.2, we generated a population of 10 neutral networks
with their defining features initialized randomly in a given range and set. The population is then
evolved for several generations until we find one or more neural networks that achieve precision 1 in
the testing data (20% of the dataset). Lastly, the selected model (i.e the neural network that achieve
precision 1) is tested against the remaining 10% of fresh data. We repeated the test varying the
number of samples in the dataset in order to understand how the performance of the model degrades
in relation to the amount of available data. The samples (i.e. raw in the dataset) have been removed
randomly to avoid injecting bias in the dataset.
For this specific case, 3 generations were sufficient for achieving accuracy “1” in the training set.
Given the relatively small number of features and samples of the dataset, the training can occur
almost in real time. Figure 4 presents the experienced ROC curves and related AUC when the selected
model (that achieve accuracy 1 in the testing data) was validated against the 10% of fresh samples
that were randomly removed prior to feeding the data to the model for training.
Figure 4 ROC Curve and AUC value varying the sample size in the dataset
As we can observe, a dataset of 100 samples is not sufficient for training a model that is capable to
generalize from the data as the AUC is significantly lower the other models that use a larger sample.
In addition, as we increase the size of the available data, we obtain better performances as well as a
sounder model. Finally, the accuracy moved from 0.5 for 100 samples to 0.82 for the complete dataset
(1470).
Although the data are clearly specific of a particular use case, it is reasonable to assume that this
method of modelling cloud predicts similar performances if similar datapoints were collected from a
different domain. Thus, confirming that tools that take advantage of a similar AI based model may be
developed for companies where it is possible to identify and monitor a relatively large (500+) number
of employees working at similar jobs. This intrinsic limit of the technology may limit the
competitiveness of SMEs unless they properly organize.
5. Conclusion
In this paper, we present an example of how data can be used for developing a more human-centric
approach and improve the quality of work of employees. We use a machine learning approach for
developing the possibility of predicting attraction in the workforce with reasonably sound
performance. As we are advancing in the 4th and 5th industrial revolution, more tools that take
advantage of a similar approach will be developed and adopted by companies. However, this
paradigm is challenging the interoperability community as the amount of data needed for the
developing of the models may not be available in only one company, in particular if we consider the
case of SMEs. Our experimental results show that is reasonable to assume that we need more than 500
samples if we want to develop an AI based model. This represents a clear limitation if we want to
adopt these solutions in companies that have a limited number of employees.
We could overcome this limitation by developing safe dataspaces where multiple stakeholders can
share their information to achieve a minimum dataset that may enable AI-based solutions.
In conclusion, this work intends to advance the conversation towards the development of
interoperable non-production efficiency AI solutions by:
• Presenting a case where we face data scarcity and the need to jointly overcome this obstacle.
• Providing a working proof of concept coming from the HR domain. Within this specific case,
we define a minimum size threshold for implementing AI based models in the context of
predicting the possibility of an employee experiencing attrition.
• Advocating for the creation of meaningful dataspaces where people and organizations can share
their data in a GDPR compliant manner.
• Outlining the challenge for modern interoperability best practices as these data are from
different and scattered entities.
6. References
[1] K. Lee, Artificial intelligence, automation, and the economy, 2016. URL:
https://obamawhitehouse.archives.gov/blog/2016/12/20/artificial-intelligence-automation-and-
economy
[2] F. Lelli, Interoperability of the Time of Industry 4.0 and the Internet of Things. Future Internet
11 (2019) 36. doi: 10.3390/fi11020036.
[3] B. Sivathanu, R. Pillai, Smart HR 4.0 – how industry 4.0 is disrupting HR, Human Resource
Management International Digest 26 (2018) 7-11.
[4] J. Lee, H. Davari, J. Singh, V. Pandhare, Industrial Artificial Intelligence for industry 4.0-based
manufacturing systems, Manufacturing Letters 18 (2018) 20-23. doi:
10.1016/j.mfglet.2018.09.002.
[5] D. F. Mujtaba, N. R. Mahapatra, Ethical Considerations in AI-Based Recruitment, in: 2019 IEEE
International Symposium on Technology and Society (ISTAS), IEEE, New York, 2019, pp. 1-7,
doi: 10.1109/ISTAS48451.2019.8937920.
[6] S. Berhil, H. Benlahmar, N. Labani, A review paper on artificial intelligence at the service of
human resources management, Indonesian Journal of Electrical Engineering and Computer
Science 18 (2020) 32-40. doi: 10.11591/ijeecs.v18.i1.pp32-40.
[7] F. Fallucchi, M. Coladangelo, R. Giuliano, E. W. De Luca, Predicting Employee Attrition Using
Machine Learning Techniques, Computers 9 (2020) 86. doi: 10.3390/computers9040086.
[8] S. Al-Darraji, D. G. Honi, F. Fallucchi, A. I. Abdulsada, R. Giuliano, H. A. Abdulmalik,
Employee Attrition Prediction Using Deep Neural Networks, Computers 10 (2021) 141. doi:
10.3390/computers10110141.
[9] B. Otto, M. Jarke, Designing a multi-sided data platform: findings from the International Data
Spaces case, Electron Markets 29 (2019) 561–580. doi: 10.1007/s12525-019-00362-x
[10] M. Franklin, A. Halevy, D. Maier, From databases to dataspaces: a new abstraction for
information management, SIGMOD Records 34 (2005) 27-33. doi: 10.1145/1107499.1107502.
[11] IBM HR Analytics Employee, 2017. URL: https://www.kaggle.com/pavansubhasht/ibm-hr-
analytics-attrition-dataset