<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Fast Predictive Maintenance in Industrial Internet of Things (IIoT) with Deep Learning (DL): A Review</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Thomas</forename><surname>Rieger</surname></persName>
							<email>thomas.rieger@plymouth.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="department">School of Computing and Mathematics</orgName>
								<orgName type="institution">Plymouth University</orgName>
								<address>
									<country key="GB">United Kingdom</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Stefanie</forename><surname>Regier</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Karlsruhe University of Applied Sciences</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ingo</forename><surname>Stengel</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Karlsruhe University of Applied Sciences</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Nathan</forename><surname>Clarke</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">School of Computing and Mathematics</orgName>
								<orgName type="institution">Plymouth University</orgName>
								<address>
									<country key="GB">United Kingdom</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Fast Predictive Maintenance in Industrial Internet of Things (IIoT) with Deep Learning (DL): A Review</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">19185A26826D7F4A9B26B874BE8FC2DC</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T02:41+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Predictive Maintenance</term>
					<term>Industrial Internet of Things</term>
					<term>IIoT</term>
					<term>Deep Learning</term>
					<term>Real-time</term>
					<term>Data Streams</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Applying Deep Learning in the field of Industrial Internet of Things is a very active research field. The prediction of failures of machines and equipment in industrial environments before their possible occurrence is also a very popular topic, significantly because of its cost saving potential. Predictive Maintenance (PdM) applications can benefit from DL, especially because of the fact that high complex, non-linear and unlabeled (or partially labeled) data is the normal case. Especially with PdM applications being used in connected smart factories, low latency predictions are essential. Because of this real-time processing becomes more important. The aim of this paper is to provide a narrative review of the most current research covering trends and projects regarding the application of DL methods in IoT environments. Especially papers discussing the area of predictions and realtime processing with DL models are selected because of their potential use for PdM applications. The reviewed papers were selected by the authors based on a qualitative rather than a quantitative level.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>This paper provides an analysis of selected literature applying DL techniques and Artificial Neural Networks (ANN) in the field of industrial IoT (IIoT) to produce fast predictions as required, among others, in maintenance applications. PdM attempts to predict failures before their possible occurrence to avoid unscheduled outages of machines and plants. The aim is to avoid breakdowns by their timely prediction and maximizing the service life at the same time. The predictions are based on data comprising accumulated knowledge and current conditions.</p><p>IIoT environments produce massive amounts of data. The necessity to perform data analytics on such massive data brings the characterizing features of Big Data into play, like the "5V's" volume, variety, velocity, variability, and veracity <ref type="bibr" target="#b0">[1]</ref>. The high volume and the high complexity of data put massive demands on existing data processing techniques. Additionally, evolving data streams and real-time demands intensify the demands even more <ref type="bibr" target="#b1">[2]</ref>. Sensors typically generate continuous streams of data. The term of data streams refers to data continuously generated typically at a high rate <ref type="bibr" target="#b2">[3]</ref>. In fully automated industrial environments, obtaining information in realtime and react immediately becomes indispensable. In IIoT environments Machine to Machine (M2M) communication has high significance <ref type="bibr" target="#b3">[4]</ref>. Intelligent sensors and devices not only sending data but communicating with their environment, anticipate immediate responses. In such IIoT Environments the characteristic of taking a snapshot of the entire data set and performing calculations with unpredictable response time contrasts with the demand for real-time communication and the presence of continuous flowing data streams <ref type="bibr" target="#b4">[5]</ref>. To cope with such demands self-adaptive algorithms continuously learning and improving their models are essential. In addition, such algorithms should provide high performance and real-time behaviour. This is not only true when they are running on powerful cloud systems but also on fog and edge systems or IoT devices <ref type="bibr" target="#b5">[6]</ref>.</p><p>The methodological approach of this paper is a narrative review. The reviewed papers were selected by the authors based on a qualitative rather than a quantitative level. Papers covering the most current research for the topic fast predictions in IIoT with DL were given priority. There are many papers covering the topic of DL in (I)IoT. To the best of our knowledge, there is no paper in literature covering the specific topic of PdM in connection with DL an (I)IoT.</p><p>This review provides a classification of different DL approaches mentioned for use in industry und IoT. It also covers the topics of real-time processing and data streams in regard to the mentioned DL approaches. Techniques intended to improve the realtime and stream processing ability of different approaches mentioned in the reviewed papers are evaluated and classified. Special focus is set on the ability of the mentioned approaches to provide predictions. The paper concludes with a summary and outlook on future developments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Deep Learning Approaches in Industrial Internet of Things</head><p>This section starts with a short introduction into DL and ANN. A classification of different DL methods mentioned for the use in industry und IoT will then be provided. The classification will be done by the theoretic approaches, application areas and strength and weaknesses in regard to the demands of PdM in IIoT environments. The reviewed papers covering the topics of DL methods in Cyber Physical Systems (CPS), IoT, Industry 4.0 (I4.0), as well as the topics of real-time and data stream processing.</p><p>DL can be defined as a subcategory of Machine Learning (ML) whereas ML is a segment in the field of Artificial Intelligence (AI). DL itself is often defined as a class of optimized ANNs comprising numerous layers (hidden layers). The high number of layers and neurons allow the abstraction of more complex problems and support further characteristics like the ability to unsupervised learning or automatic feature extraction <ref type="bibr" target="#b6">[7]</ref>. Examples are Deep Neural Networks (DNN), Deep Belief Networks (DBN) or Recurrent Neural Networks (RNN).</p><p>The basic idea behind an ANN is to imitate the biological neural network in mammalian brains. Components of an ANN are neurons (in ANNs often called nodes) and connections between those nodes. The nodes are organized in layers producing nonlinear output data based on the input data. The connections between the nodes transfer the output of one node to the input of another node. Weights assigned to each connection determine the relevance of the transferred signal. As in biological neural networks the output signal of a neuron (node) is ruled by a threshold function. To set up an ANN all weights have to be set to an initial value (often just simple estimates). By training the network those weights are adjusted in a holistic way following a defined learning rate to achieve a valid and balanced network. This is also often referred to as "connections developing over time with training". ANNs are known for more than 50 years and various ways have been developed since <ref type="bibr" target="#b19">[21]</ref>, <ref type="bibr" target="#b20">[22]</ref>, <ref type="bibr" target="#b21">[23]</ref>.</p><p>In <ref type="bibr" target="#b5">[6]</ref> the following DL models are listed for IoT application: Auto-encoder (AE), RNN, Restricted Boltzmann Machine (RBN), DBN, Long Short-term Memory (LSTM), Convolutional Neural Network (CNN), Variational Auto-encoder (VAE), Generative Adversarial Network (GAN) and Ladder Net. The DL models are categorized in <ref type="bibr" target="#b5">[6]</ref> into the three main groups of generative approaches (AE, RBM, DBN, VAE), discriminative approaches (RNN, LSTM, CNN) and hybrid (GAN, Ladder Net) as a combination of the two approaches mentioned before. This categorisation mainly refer to the underlying learning method whereas generative approaches basically follow the principle of unsupervised learning and discriminative approaches follow the principle supervised learning. Beside the definition of the required number of layers (complexity) the underlying learning method is a decisive factor for the selection of a DL approach. The categorization in generative and discriminative approaches chosen by <ref type="bibr" target="#b5">[6]</ref> can be fundamentally found in many other works. In <ref type="bibr" target="#b5">[6]</ref> different DL models are also categorized by their suitability in IoT applications. The relevant characteristics mentioned in <ref type="bibr" target="#b5">[6]</ref> are the ability to work with (partially) unlabelled data (feature extraction, feature discovery), the magnitude of needed training dataset, dimensionality reduction abilities, the ability to deal with noisy data and time series data and their general performance classification. For the reduction of high dimensional data and to cope with unlabelled data <ref type="bibr" target="#b5">[6]</ref> recommends the combination of RNN with DBN and AE. If the system is meant to make predictions like in PdM systems, DBN and AEs are often used as an upfront layer providing classified data to a subsequent RNN <ref type="bibr" target="#b5">[6]</ref>.</p><p>In case of spatial-temporal data like mobility data, RNNs are recommended because they show good results when data is developing in a sequential way. But if data also comprises long term dependencies, RNNs are not a good choice because RNNs does not memorize previous states and results <ref type="bibr" target="#b7">[8]</ref>. An approach to handle sequential data streams from human mobility and transportation transition models containing long term dependencies (behaviours) is described in <ref type="bibr" target="#b7">[8]</ref>. The described solution is a combination of RNN with LSTM in the form of a specialized RNN architecture. Besides the ability to handle long term dependencies the LMST also adds labelling and predictive functionality to that combination. The combination of RNN with LSTM to cope with data streams or time-series data comprising long-term dependencies (like certain behaviours or wear and tear of machineries) can be found in many other works <ref type="bibr" target="#b7">[8]</ref>, <ref type="bibr" target="#b8">[9]</ref>, <ref type="bibr" target="#b10">[11]</ref>, <ref type="bibr" target="#b17">[18]</ref>.</p><p>The paper "IoT Data Analytics Using Deep Learning" <ref type="bibr" target="#b8">[9]</ref> describes how to select the right ANN to archive predictions from data streams and time-series data. To retrieve trends and predictions and also validate those trends and predictions in parallel by anomaly detection, a combination of LSTM with Naive Bayes models is proposed. The LSTM produces the predictions on data streams whereas the Naive Bayes model is responsible for anomaly detection performed on the results of the LSTM.</p><p>This paper also reflects on the fact that Simple Feedforward ANN (FNN) like Single-layer Perceptron (SLP) and Multi-layer Perceptron (MLP) using standard backpropagation (BP) for training are often not a good choice because they does not perform well in complex situations and on data streams with long-term dependencies. This is especially true when data streams comprise time series data and the aim of the model is to predict future events or trends. Data streams and time-series data usually have dependencies over time. Such dependencies are typical for IoT data and provide relevant insights. In simple ANNs data moves straight through the layers with the assumption that input data is independent from output data. Because of this, there is no way to remember previous input and output states (previous results). This is bad if previous data is linked to current data. Using RNN instead can archive better results in data streams and time-series data. Because the connections between nodes in a RNN are in the form of sequences or loops, it is possible to remember previous states. To avoid gradient explosions normally only a view states are remembered. Therefore only short-term dependencies are recognized. Because of this <ref type="bibr" target="#b8">[9]</ref> recommends the application of LSTM in complex IoT environments to recognize long-term dependencies in the data. LSTM are a variant of RNN introducing memory units. Those memory units are able to remember important previous states and forget the unimportant ones <ref type="bibr" target="#b8">[9]</ref>.</p><p>To predict the behaviour of energy systems in the manner of smart grids <ref type="bibr" target="#b9">[10]</ref> remark that more intelligent systems are necessary to produce accurate predictions on the future energy consumptions. In the paper "Deep learning for estimating building energy consumption" <ref type="bibr" target="#b9">[10]</ref> it is stated that ANN-based prediction methods are a promising approach because of their ability to handle massive and highly non-linear time series data coming from different heterogeneous data sources (e.g. SmartMeter) and containing a lot of uncertainty (unlabelled data). In the paper <ref type="bibr" target="#b9">[10]</ref> they benchmarked two different approaches of the RBN, namely Conditional Restricted Boltzmann Machine (CRBM) and Factored Conditional Restricted Boltzmann Machine (FCRBM), on a synthetic benchmark dataset. Based on this experiment the authors come to the conclusion that FCRBN outperforms in comparison to RNN, Support Vector Machine (SVM), as well as CBRM because of its added factored conditional history layer. A RBM is a stochastic ANN consisting of two layers, a visible layer and a hidden layer. In simple terms, the visible layer of a RBM contains a node for each possible value in the input data whereas the hidden layer defines categories of values. Because in a RBM each visible layer node is connected to any hidden layer node a RBN is good in feature classification, feature extraction and complexity reduction (by identifying the most important features). For DL RBMs can be stacked. In <ref type="bibr" target="#b9">[10]</ref> RBM is extended by a conditional history layer (CRBM) enabling the RBN to detect long-term dependencies in time-series data. Additionally the output of one stacked CRMB layer is factored (FCRBM) to reduce the number of possible compositions.</p><p>Another paper in the field of energy management also emphasizes the very powerful forecasting abilities of DL. In <ref type="bibr" target="#b10">[11]</ref> the application of AE and LSTM is described for predicting the power generation of solar systems. The accuracy reached by a combination of AE and LMST (Auto-LSTM) is compared to other neural networks (namely MLP) as well as to a physical model. The benchmark data is taken from 21 real solar power plants and the benchmark is taken from an experimental setup described in <ref type="bibr" target="#b10">[11]</ref>. The following measurements are taken as benchmarks: average rootmean-square deviation (RMSD), average mean absolute error (MAE), average absolute deviation (Abs. Dev.), average BIAS and average correlation. The measured results show that all ANN-and DL-based models show far better results than the physical model. Among all ANN-and DL-based models Auto-LSTM is the best choice in this specific scenario and specific data set. The capability to extraction features in unlabelled data is mentioned as a decisive factor in making predictions.</p><p>The paper "An enhancement deep feature fusion method for rotating machinery fault diagnosis " <ref type="bibr" target="#b11">[12]</ref> points out the strength of AEs in feature extraction and feature learning. The paper describes how to further improve the feature learning ability with reduced influence of background noises by stacking Deep AE (noise reduction) and contractive AE (enhanced feature recognition), called deep feature fusion method.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Fast Predictions using DL</head><p>In many IoT applications real-time processing is essential. For example in a PdM system high latency could lead to unintentional reactive maintenance because of insufficient lead time to plan the maintenance tasks <ref type="bibr" target="#b4">[5]</ref>. How fast real-time processing needs to be, strongly depends on the application case. According <ref type="bibr" target="#b12">[13]</ref> in micro manufacturing systems, where vast volumes of micro parts are manufactured with high speed, the term real-time means microseconds. <ref type="bibr" target="#b12">[13]</ref> shows that with systems for fault detection and PdM the rejection rate of the manufactured micro parts decrease by increasing processing speed <ref type="bibr" target="#b12">[13]</ref>. In other scenarios, the term of real-time can mean seconds, minutes or hours. For example in PdM Applications for offshore wind turbines the frequency with which the data is available is mostly minutes and hours <ref type="bibr" target="#b13">[14]</ref>.</p><p>The paper "Metro Density Prediction with Recurrent Neural Network on Streaming CDR Data" <ref type="bibr" target="#b14">[15]</ref> describes the implementation of a real-time public transportation crowd prediction system using a weight-sharing recurrent neural network in combination with parallel streaming analytical programming. Fast response time to emergent situations (e.g. entrance records in metro stations combined with telecommunication data) demand real-time analysis. The use of a powerful neural network model with strong learning capability offers a wide range of new insights but contrast with the need for fast response time. The way to meet this goal is described in <ref type="bibr" target="#b14">[15]</ref> with three steps: a) adopting a RNN model to improve its ability to work on data streams, b) implement strategies for parallelization of RNNs and c) the use of parallel streaming analytical algorithms over a cloud-based stream processing platform. In the project described in <ref type="bibr" target="#b14">[15]</ref> each metro station is modelled by an independent RNN. Shared layers are introduced to share weights from stations which are in similar "situations" (e.g. a downtown station during rush hour) across several models dynamically. Weight-sharing also enables co-training in parallel <ref type="bibr" target="#b14">[15]</ref>.</p><p>The application of RNNs and their many variations for fast data analytics is also recommended in <ref type="bibr" target="#b5">[6]</ref>. Especially on typical sensor data like serial data, time-series data and data streams, RNNs can provide better performance than other models. Such sensor data is dominating in most PdM applications <ref type="bibr" target="#b0">[1]</ref>.</p><p>In order to be able to develop and permanently adapt models on massive data comprising the behaviour of people and their spatial and temporal attributes together with transportation capacities, real-time processing and real-time learning capabilities are essential. The paper <ref type="bibr" target="#b7">[8]</ref> describes a multi-task deep LSTM learning architecture. The basic idea of this concept is not to use a joint feature vector but various LSTM tasks separated by their domain (e.g. respectively a separate task for mobility and transportation mode prediction). This architecture performs parallel learning whereas the results are aggregated depending on the intended insights <ref type="bibr" target="#b7">[8]</ref>.</p><p>Assistance systems in cars like traffic sign recognition must deliver accurate results with low latency. The paper <ref type="bibr" target="#b15">[16]</ref> describes how to apply DNN in this field. The model of the system is continuously updated (online learning) and fed only with completely unlabelled data (raw images). A CNN with 9 layers is used for image recognition. To improve the performance of system max-pooling layers are combined with convolutional layers in an alternating way. The convolutional layers perform convolution on 2D input pixel maps. The max-pooling layer works like a pre-processor between two convolutional layers transforming the output of a preceding convolutional layer to the input of a subsequent convolutional layer by eliminating overlapping regions in the pixel maps. This eliminates redundant processing in the complex and time consuming convolutional layers. The approach described in <ref type="bibr" target="#b15">[16]</ref> is referred to as Multi-Column DNN (MCDNN).</p><p>The paper <ref type="bibr" target="#b16">[17]</ref> describes a real-time oriented solution for traffic sign detection and recognition. The primary focus is on the need for parallel processing because of the need to detect diverse traffic sign at the same time. In this approach also CNN is used for image processing in combination with AdaBoost to improve performance and parallel GPU processing.</p><p>Because of its memory cells LSTM models are good if data comprises long-term dependencies. If the data structure allows the separation of single entities with their specific behaviour as well as the formation of groups of entities, it could be then possible to process each entity and every group with its own neural network. This opens up parallel processing possibilities of the single neural networks. Normally each single and parallel processed neural network provides its result to an aggregation layer aggregating all outputs to an overall result. The paper "A Hierarchical Deep Temporal Model for Group Activity Recognition" <ref type="bibr" target="#b17">[18]</ref> describes how to recognize situations in a volleyball match. One LSTM model per player predicts the behaviour of this player, remembering his previous behaviour in the match (long-term dependencies). Each single situation of the match is then modelled as a group of the players. The LSTMs are hierarchically ordered where the LSTM models of all involved players are subordinated to a scene. The scenes and the players behaviour is extracted based on images using CNN <ref type="bibr" target="#b17">[18]</ref>.</p><p>The paper <ref type="bibr" target="#b6">[7]</ref> mentions that because of the demands for real-time processing, the organization of layers and connections have changed. Fully connected networks where each node of a layer is connected to all nodes of the subsequent layer can handle complex problems but also demand a lot computing power. Dropout all connections not really influencing the result is a strategy to reduce the complexity of a DL network, and therefore its computing demand, without affecting accuracy in a relevant manner. Besides dropout <ref type="bibr" target="#b6">[7]</ref> also mention max pooling layers, batch normalization and transfer learning as additional strategies for performance optimization. Despite all the mentioned papers discussing performance enhancements and realtime abilities of DL models, [19] considers that highest accuracy still stands over all in mostly all current DL projects. The paper "An Analysis of Deep Neural Network Models for practical Applications" [19] argues that numerous DL approaches de-scribed in literature are simply not suitable for practical use. This is for example because of their long processing time or excessive power consumption. In his paper he demands to spend more attention to performance issues because they are key factors in practical DL applications. The paper compares 14 different specific DL projects like AlexNet or GoogLeNet by comparing their accuracy, memory footprint, parameters, operations count, inference time, and power consumption. The paper shows that a small increase in accuracy lead to an enormous increase in computational power and computation time. It is recommended to define a maximum energy consumption for each DL project and adjust the accuracy according to it <ref type="bibr">[19]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Conclusions</head><p>In this paper we provided a narrative review of selected literature applying DL techniques in the field of IIoT to produce fast predictions of maintenance issues. The papers have shown that the use of DL in IoT and PdM is a vital topic in industry. Many different applications are in use in practice and are constantly being developed and improved.</p><p>Frequently reported are combinations of different DL models to combine different advantages and strengths in one application. Also, the need for real-time processing of complex data and data streams has been demonstrated in certain application scenarios. This include in particular applications for predictions such as PdM. In order to increase the real-time capability, concepts of parallel DL networks using a final aggregation layer, or intermediate layers for the reduction of complexity are frequently used. Although many activities can be observed in the area of real-time processing of DL models, there are also critical voices criticizing the absolute focus on accuracy and calling for a greater focus on performance and lighter applications suitable for practical use. Almost all reports agree that a lot of research is still needed in this area.  <ref type="table" target="#tab_0">1</ref> gives an overview of the reviewed papers with the DL-Methods mentioned. For each paper the characteristics (or strength and weaknesses) as well as the recommended application areas (like predictions) of the DL-Methods mentioned in the corresponding paper are summarized. Table <ref type="table" target="#tab_0">1</ref> makes no statement regarding the validity of results in a quantitative way. The categorisation of the different DL models is only made in a qualitative way. This is because among all reviewed papers only in <ref type="bibr">[19]</ref> concrete measured values are defined. All other papers solely provide qualitative statements. How to measure and evaluate the validity and quality of results of different DL methods is an open question <ref type="bibr" target="#b18">[20]</ref>. So far, few approaches for measuring, evaluating and benchmarking have been developed. Moreover, those approaches are usually not verifiable as generally valid. For instance, in the case of classifications the use of accuracy estimation techniques, such as the "holdout method" or "n-fold crossvalidation", can be used to evaluate performance, predictive ability and model accuracy <ref type="bibr" target="#b18">[20]</ref>. As such, mentioned techniques divide a training set via varying approaches into data areas for learning and validation. For most models no measuring, evaluating and benchmarking concept has yet been defined. In general, the evaluation is done here by expert opinions <ref type="bibr" target="#b18">[20]</ref>. The paper <ref type="bibr" target="#b18">[20]</ref> points out that there is a demand for improved measuring and benchmark methods. Proven measurement methods to generate representative benchmarks are needed in order to be able to assess DL models.</p><p>The papers <ref type="bibr" target="#b0">[1]</ref> to <ref type="bibr" target="#b4">[5]</ref>, <ref type="bibr" target="#b6">[7]</ref>, <ref type="bibr" target="#b12">[13]</ref>, <ref type="bibr" target="#b13">[14]</ref> and <ref type="bibr">[19]</ref> to <ref type="bibr" target="#b21">[23]</ref> are not part of Table <ref type="table" target="#tab_0">1</ref> because they are used as reference regarding basic statements and explanations made in this paper. These papers were not on the topic of DL methods and techniques.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Summary of reviews papers with the DL-Methods mentioned</figDesc><table><row><cell>Ref.</cell><cell cols="2">DL-Methods Characteristics</cell><cell>Typical applications</cell></row><row><cell>[6]</cell><cell>AE, CNN,</cell><cell>Feature extraction and dimensionality</cell><cell>Fault detection and</cell></row><row><cell>Mohammadi,</cell><cell>DBN, GAN,</cell><cell>reduction of IoT Data with AE, DBN</cell><cell>predictions IoT envi-</cell></row><row><cell>et al., 2018</cell><cell>LSTM, RBM,</cell><cell>CNN for image recognition but needs</cell><cell>ronments</cell></row><row><cell></cell><cell>RNN, VAE,</cell><cell>large training set</cell><cell>Real-time and stream</cell></row><row><cell></cell><cell>Ladder Net</cell><cell>GAN, VAE and Ladder Net suitable for</cell><cell>processing with differ-</cell></row><row><cell></cell><cell></cell><cell>noisy data, used as classification layer for</cell><cell>ent kinds of RNNs</cell></row><row><cell></cell><cell></cell><cell>RNN to enable unsupervised learning</cell><cell></cell></row><row><cell></cell><cell></cell><cell>LSTM provide good performance for</cell><cell></cell></row><row><cell></cell><cell></cell><cell>data with long term dependencies</cell><cell></cell></row><row><cell></cell><cell></cell><cell>RBM for feature extraction, dimensional-</cell><cell></cell></row><row><cell></cell><cell></cell><cell>ity reduction and classification problems</cell><cell></cell></row><row><cell></cell><cell></cell><cell>RNN especial for time-series data</cell><cell></cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_0">Internet of Things, Networks and Security</note>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper was submitted to the Collaborative European Research Conference (CERC 2019) https://www.cerc-conference.eu/</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">Murali</forename><surname>Pusala</surname></persName>
		</author>
		<title level="m">Massive Data Analysis: Tasks, Tools, Applications and Challenges. Big Data Analytics</title>
				<imprint>
			<publisher>Springer Verlag</publisher>
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Sliding Window-Based Fault Detection From High-Dimensional Data Streams</title>
		<author>
			<persName><forename type="first">Liangwei</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS</title>
		<imprint>
			<biblScope unit="volume">47</biblScope>
			<biblScope unit="page">2</biblScope>
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Data stream classification and big data analytics</title>
		<author>
			<persName><forename type="first">Bartosz</forename><surname>Krawczyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michal</forename><surname>Wozniak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neurocomputing</title>
		<imprint>
			<biblScope unit="page">150</biblScope>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Real-time fault detection for advanced maintenance of sustainable technical systems</title>
		<author>
			<persName><surname>Ait-Alla</surname></persName>
		</author>
		<author>
			<persName><surname>Abderrahim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procedia CIRP</title>
		<imprint>
			<biblScope unit="page">41</biblScope>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Movement towards service-orientation and app-orientation in manufacturing IT</title>
		<author>
			<persName><forename type="first">Dennis</forename><surname>Bauer</surname></persName>
		</author>
		<author>
			<persName><surname>Stock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Daniel Und Bauernhansl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">10th CIRP Conference on Intelligent Computation in Manufacturing Engineering -CIRP ICME &apos;16</title>
				<imprint>
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Deep Learning for IoT Big Data and Streaming Analytics: A Survey</title>
		<author>
			<persName><forename type="first">Mehdi</forename><surname>Mohammadi</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1712.04301v2</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE COMMUNICATIONS SURVEYS &amp; TUTORIALS</title>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Risk Perceptions for Wearable Devices</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">N</forename><surname>Lee</surname></persName>
		</author>
		<ptr target="http://arxiv.org/pdf/1504.05694.pdf" />
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
		<respStmt>
			<orgName>Cornell University Library</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">DeepTransport: Prediction and Simulation of Human Mobility and Transportation Mode at a Citywide Level</title>
		<author>
			<persName><forename type="first">Xuan</forename><surname>Song</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
		<respStmt>
			<orgName>Center for Spatial Information Science, The University of Tokyo, Japan Internet of Things, Networks and Security</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">Xiaofeng</forename><surname>Xie</surname></persName>
		</author>
		<title level="m">IoT Data Analytics Using Deep Learning, Key Laboratory for Embedded and Networking Computing of</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
		<respStmt>
			<orgName>Hunan Province, Hunan University</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">Elena</forename><surname>Mocanu</surname></persName>
		</author>
		<title level="m">Deep learning for estimating building energy consumption</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
		<respStmt>
			<orgName>Department of Electrical Engineering, Eindhoven University of Technology, The Netherlands</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Deep Learning for Solar Power Forecasting -An Approach Using Autoencoder and LSTM Neural Networks</title>
		<author>
			<persName><forename type="first">André</forename><surname>Gensler</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Systems, Man, and Cybernetics • SMC 2016</title>
				<meeting><address><addrLine>Budapest, Hungary</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016-10-09">2016. 2016. October 9-12, 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">Haidong</forename><surname>Shao</surname></persName>
		</author>
		<title level="m">An enhancement deep feature fusion method for rotating machinery fault diagnosis</title>
				<meeting><address><addrLine>Xi&apos;an, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page">710072</biblScope>
		</imprint>
		<respStmt>
			<orgName>School of Aeronautics, Northwestern Polytechnical University</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><surname>Rippel</surname></persName>
		</author>
		<author>
			<persName><surname>Daniel</surname></persName>
		</author>
		<author>
			<persName><surname>Lütjen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><surname>Michael Und Freitag</surname></persName>
		</author>
		<title level="m">SIMULATION OF MAINTENANCE ACTIVIES FOR MICRO-MANUFACTURING SYSTEMS BY USE OF PREDICTIVE QUALITY CONTROL CHARTS</title>
				<imprint>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A Concept for the Dynamic Adjustment of Maintenance Intervals by Analysing Hereogenoeous Data</title>
		<author>
			<persName><forename type="first">Michael</forename><surname>Freitag</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Mechanics and Materials</title>
		<imprint>
			<biblScope unit="volume">794</biblScope>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">Victor</forename><forename type="middle">C</forename><surname>Liang</surname></persName>
		</author>
		<idno>-2020-1/16</idno>
		<title level="m">Mercury: Metro Density Prediction with Recurrent Neural Network on Streaming CDR Data</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="978" to="979" />
		</imprint>
	</monogr>
	<note>ICDE 2016 Conference</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">Dan</forename><surname>Ciresan</surname></persName>
		</author>
		<title level="m">Multi-Column Deep Neural Network for Traffic Sign Classification</title>
				<meeting><address><addrLine>Manno -Lugano; Switzerland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2012">2012. 6928</date>
		</imprint>
		<respStmt>
			<orgName>IDSIA -USI -SUPSI | Galleria 2</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Real-time traffic sign recognition based on a general purpose GPU and deep-learning</title>
		<author>
			<persName><forename type="first">Kwangyong</forename><surname>Lim</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">50</biblScope>
		</imprint>
		<respStmt>
			<orgName>Department of Computer Science, Yonsei University</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">A Hierarchical Deep Temporal Model for Group Activity Recognition</title>
		<author>
			<persName><forename type="first">Mostafa</forename><forename type="middle">S</forename><surname>Ibrahim</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1605.07678v4</idno>
	</analytic>
	<monogr>
		<title level="m">AN ANALYSIS OF DEEP NEURAL NETWORK MODELS FOR PRACTICAL APPLICATIONS</title>
				<editor>
			<persName><forename type="first">Alfredo</forename><surname>Canziani</surname></persName>
		</editor>
		<meeting><address><addrLine>Burnaby, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
		<respStmt>
			<orgName>School of Computing Science, Simon Fraser University ; Weldon School of Biomedical Engineering Purdue University, Faculty of Mathematics, Informatics and Mechanics University of Warsaw</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Data stream classification and big data analytics</title>
		<author>
			<persName><forename type="first">Bartosz</forename><surname>Krawczyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michal</forename><surname>Wozniak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neurocomputing</title>
		<imprint>
			<biblScope unit="page">150</biblScope>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Deep Learning Techniques and its Various Algorithms and Techniques</title>
		<author>
			<persName><forename type="first">Nidhi</forename><surname>Bhatia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Engineering Innovation &amp; Research</title>
		<idno type="ISSN">2277 -5668</idno>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">5</biblScope>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">Ken</forename><surname>Chatfield</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1405.3531v4</idno>
		<title level="m">Return of the Devil in the Details: Delving Deep into Convolutional Nets, Visual Geometry Group</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
		<respStmt>
			<orgName>Department of Engineering Science, University of Oxford</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<ptr target="http://www.devcoons.com/literature-review-of-deep-machine-learning-for-feature-extraction/" />
		<title level="m">DEVCOONS Website</title>
				<imprint/>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
