1. Introduction
Applications based on smartphones, sensors and actuators are becoming more and more intelligent over the past decade and facilitate communication between devices and the performance of more complex tasks. The number of network devices exceeded the world population [
1] in 2008 and the figure continues to increase exponentially until today. In the age of the Internet of Things (IoT), smartphones, built-in systems, wireless sensors and most every device are connected by a local network or the internet. The growth in Internet-of-Things (IoT), which includes smartphones [
2], sensor networks [
3], sensors unusual aerial vehicles (UAV) [
4,
5], cognitively smart systems [
6], and so on has created a multitude of new applications across various mobile and remote platforms. The amount of data obtained from such devices often increases with the growing number of devices. New technologies are emerging that evaluate data gathered for practical connections and decision making, progressing to Artificial Intelligence (AI) using Machine Learning(ML) and Deep learning (DL) algorithms.
We generally adopt a work flow model, which consists of collecting data, analysis of data, visualizing data and evaluating data [
7,
8], in order to build successful IoT applications. Data analysis is a crucial, computer-intensive dimension in which historically developed technologies typically combine technical expertise and ML (e.g., logistical regression, vector support, and the random forest) for classification or regression problems such as traffic conditions forecast [
9], tracking vehicles [
10], estimating delivery time [
11], etc. Furthermore, as society enters the “big data” generation, these traditional methods are not strong enough to process huge, volatile, and irregular data from invisible, heterogeneous, IoT-based databases. Nearly all conventional systems focus on fully enclosed features and their efficiency relies heavily on the previous knowledge of particular areas. Most learning techniques used in those systems usually use shallow architectures, whose modeling and representational capacity are very small. As such, it is essential to have a much more effective analytical tool to exploit the maximum potential of the invaluable raw data produced in various IoT operations.
The annual economic effect of IoT in 2025 will range from
$2:7 to
$6:2 trillion based on McKinsey’s study on the IoT global economic consequences [
12]. Healthcare is the largest share of this sector about 41%, followed by industry and oil with 33% and 7%, respectively, of the IoT sector. Additional fields such as transport, irrigation, public infrastructure, security and retailers account for approximately 15% of the entire IoT sector. Such expectations mean the immense and rapid growth of IoT services, their data generation and, therefore, their corresponding demand in the upcoming years. In McKinsey’s report [
12], the economic impact of machine learning is characterized by automated learning; ’the use of computers to carry out tasks that rely on complex assessments, precise evaluations and innovative problem solving’. The study addresses the key proponents of information automation in ML techniques, such as DL and neural networks.
Communication from Machine-To-Machine can be short distance using Wi-Fi, Bluetooth and ZigBee technologies or broad-band mobile networks such as LoRa, M1 CAT, Sigfox, GSM, 4G, LTE and 5G [
13].Since IoT devices are used extensively in all sorts of everyday applications, the cost of IoT devices needs to be kept low. Furthermore, IoT devices should be able to handle fundamental tasks such as collection of data, M2M interaction, etc. IoT is also tightly linked to “big data” as IoT devices collect and exchange vast data continuously. In general, therefore, an IoT infrastructure uses methods to manage, store and evaluate massive data [
14,
15]. In order to facilitate the communication of the M2M with protocols such as AMQP, MQTT, CoAP and HTTP [
16], it has become necessary in IoT infrastructure to use IoT-platforms such as Thingsboard, Thingspeak DeviceHive or Mainflux. It is often necessary for certain data processing to occur on IoT devices instead of other centralized nodes in the “cloud computing” network according to the application. As the processing moves partially to the end network elements, a new data processing model, called “edge computing”, is introduced [
17]. Furthermore, they may not be ideal for heavy workloads due to these devices often being low-end devices. Therefore, an intermediate node with enough resources is required to manage advanced processing tasks which are physically close to the end network components so that the burden caused by massive transmission of all data to a number of the internal cloud nodes is reduced. Here, “Fog nodes” [
18] are introduced to assist big data management on IoT devices through the provision of storage, processing and networking services. Finally, the data is stored in cloud storage, where advanced testing by means of different ML and DL technologies and sharing with other devices leads to the establishment of smart apps with modern value added. DL has this intense publicity because conventional methods of machine learning do not meet the current analytical criteria of IoT systems. Instead, according to the structure of data generation and processing of IoT as shown in
Figure 1, IoT systems require specific conventional data analytic methods, and AI methods. Dl methods used to analyze the big data in IoT cloud and streaming and fast data analysis in the edge or fog computing and data from IoT devices.
While an IoT has been conducted in recent years, the entire field of deep learning in IoT applications remains in infancy. Few researchers [
19,
20,
21,
22,
23,
24], reviewed papers in wireless sensor network(WSN) with ML, implementation of DL methods for healthcare department, mai DL approaches and applicability in IoT applications focusing on Big and streaming data analysis and DL Algorithms with its applications to make smart development, respectively. After taking the survey on existing papers, there is still no survey that explores a wide range of IoT devices thoroughly using DL. We also agree that it is time to review and inspire future study recommendations in current literature. To this end, this paper summarizes current research developments and patterns in using DL techniques to promote IoT applications. We will demonstrate how using DL to enhance IoT applications can be implemented from various perspectives. For example, monitoring of safety, analysis of diseases, indoor locations, artificial control, predicting the traffic, residential robots, drive automation, fault assessment and inspection of manufacturing. The issues, challenges and possible research directions for DL in IoT applications are also discussed to encourage future developments in this promising area and empower them.
The rest of this paper is organized as follows:
Section 2 includes a variety of popular and common DNN architectures. This also offers a concise overview of advances and fast DL architectures along with state-of-the-art DL algorithms.
Section 3 will review IoT applications and challenges in various domains (e.g., education, manufacturing, smart city, healthcare, and Intelligent Transportation Systems(ITS), agriculture) using DL. This article concludes in
Section 4.
2. Deep Learning Techniques
Stakeholders must clearly grasp the meaning, building blocks, potentials and challenges of the IoT and its derivatives big data. IoT and Big Data connect in two ways; an IoT is one of the leading producers and a major target for Big Data research in order to enhance IoT processes and services [
25]. In addition, IoT Big Data Research has shown that it gives value to society. The IoT data is distinct from the overall big data. We need to explore the properties of IoT data [
26] and how they vary from the traditional big data to analyze the requirements of IoT data analytics.
Here, discussing the advantages of DL over conventional methods of ML, which highlight the benefits of DL in IoT applications [
27,
28]. DL has more powerful capacity to generalize the dynamic relation of vast raw data in different IoT applications when compared with normal ML methods. The ability to process data is generally dependent upon the depths and the different architectures of learning models, including convolutional architectures; thus, in big data, DL models can most likely perform better while common learning models can easily be over-touched when dealing with a flood of data. Deep learning is an end-to-end process that is capable of learning how to derive successful features from raw data, without taking time and labor-intensive hand-made applications. In recent years, DL models have been more conscious than the other conventional ML approaches.
Figure 2 shows the search pattern in Google trends, in which DL is increasingly popular amongst other ML algorithms such as random forest, k-means, SVM and decision tree. Moreover, as per the Google trend,
Figure 3 shows that the CNN method became more popular out of all DL methods.
Deep learning is a recently developed multilayer neural network learning algorithm. It has revolutionized the concept of machine learning, propelling artificial intelligence and human–computer interaction forward in leaps and bounds. They have performed the evaluation test for CNN and DBN on the MNIST database and the real-world handwritten character database which gave 99.28% and 98.12% accuracy [
29]. Despite its complex structure and the diversity of registered user data, the researchers [
30] work here assumes the MIA in a semi-white box scenario where system model structures and parameters are available but no user data information is available, and verifies it as a serious threat even for a deep-learning-based face recognition system. The impact of power plants on GEP over their lifetime is studied in this article [
31]. Deep learning-based techniques are also commonly employed for time series forecasting.
DL contains powerful methods for boosting knowledge which allow a large number of unstructured information to be processed [
32]. These techniques are ideal for managing big data and for computer intensive processes such as image pattern recognition, voice acknowledgment and analysis, etc. DL needs strong computing skills and is known to take up time in the model training cycle, which has been one of the biggest challenges in the past. Efficient GPUs are widely used to carry out DL tasks with increased requirements for CPU capacity. Thus, in the period of big data, DL has become a popular form of data processing and modeling [
27]. In the DL methodology, the number of layers is tight with determined characteristics. In DL, the functionality is automatically estimated and feature calculation and extraction are not required before such a method is applied. In addition, the progress of DL introduces a wide range of network structures. The goal of the authors [
33,
34] project is to quantify EEG features in order to better understand task-induced neurological impairments caused by stroke and to assess biomarkers to distinguish between ischemic stroke patients and healthy adults.
In the two phases of training and forecasting, DL models typically offer two major improvements over conventional ML approaches [
21]. Initially, they minimize the need for human training and then remove those features that may not be obvious for the human view [
35]. DL methods also improve the accuracy. DL, like traditional ML, can be partitioned into two cases: non-supervised learning (unlabeled data models) and supervised learning (labeled data models).
2.1. Supervised Learning
The system model for supervised learning is built into a labeled training set. The backpropagation method is the primary approach used in supervised learning [
36].
2.1.1. Recurrent Neural Networks (RNNs)
The RNN is a discriminative categorical method which can process the serial and time series data mainly. In several tasks, the estimation relies on many previous tests in order to evaluate the sequences of inputs, besides the classification of individual tests. A neural feeder network does not apply in such applications because it does not rely on input and output layers. The RNN data contains the present sample as well as the previous observed sample as the input. The output at stage m can be affected by the output at stage m-1. Every neuron has a feedback loop that produces the output to the next step as an input. This step says that each neuron has an internal memory to store the previous level data estimates in RNN. We can not use the original back propagation here, despite the presence of neuronal cycles, because it works depending on loss derivation in relation to the weight in the previous layer even though we do not have a stacked layer model in RNNs. The center of Backpropagation Through Time(BPTT) [
37] is a technique called “unrolling the RNN”, so that we build a feeder network over time.
Figure 4 depicts the structure of an RNN and unrolled concept.
However, due to the diffusion of gradient problems and longer term dependency, RNNs are constrained by looking back only a few steps. New methods are proposed, such as GRU (Gated Recurrent Unit) [
38], LSTM (Long Short-Term Memory) [
39], designing the hidden state simulation to decide what to hold in past and present memory. In order to tackle problems sequentially such as text or speech and time-series data problems in different lengths, RNNs were developed. The RNNs can be applied to processes include detecting drivers’ actions in intelligent cars, defining the patterns of movements of individual persons and estimating household consumption. Consequently, RNN is mainly used in the field of natural language processing (NLP) [
40,
41,
42].
Table 1 shows the where RNN used in IoT fields.
2.1.2. Long Short Term Memory (LSTM)
The LSTM is discriminative method which can work on time-stamp, sequential and long time-dependent data.
Figure 5 shows the model of LSTM. LSTMs are a form of RNN that can learn order-dependence in series estimation. LSTM uses its unit gate definition, each based on its input calculating a value from 0 to 1. Every neuron having four gates to maintain the data such as feedback loop, multiplying forget, read and write gates. These gates control the accessing memory cells and preventing distractions by unrelated inputs. The neuron writes its data into itself when the gate forget is working, otherwise it forgets its last data by sending a 0. The other neurons that are inter connected will write to it and read the content of it, when write and read gate is fixed to 1. By knowing what LSTM data are to be recalled, the stored memory cell computations will not be corrupted over time. The common way to reduce the error by training network is BPTT. LSTM models are stronger than RNN models if the data is characterized by a long time dependence [
38].
Generally, LSTM is the expanding model of RNN. Various LSTM methods are proposed based on the original network [
39,
56]. LSTMs and standard RNNs were implemented for predicting sequence and sequence labeling tasks successfully. These models performed better than RNNs on context-sensitive(CS) and context free(CF) languages [
57]. For connected models with small sizes, LSTMs converge rapidly and provide state-of-the-art machine translation, voice recognition efficiently [
58]. LSTM networks are not well developed for larger networks on one single multi-core computer.
Table 2 shows where LSTM is used in IoT fields.
2.1.3. Convolutional Neural Networks (CNN’s)
The CNN is a discriminatory method which can used more for identifying images and differentiating one from the another. The CNN is made up of an input, an output and some hidden layers. The sub sample layers, pooling layer, convolutional layers, pooling, fully connected (FC) or non-linear layers are hidden layers in the architecture of CNNs. The CNN is the main version of the FCs. All neurons are connected from one layer to the each and every neuron in the another layer. So, FCs makes them overwrite the data.
Figure 6 depicts the structure of a CNN.
DNNs with a dense relation between the layers are difficult to train and do not test well for vision-based tasks due to the characterization of the translation-invariance [
77]. These problem can be solved by the CNN with the support of the above said property. A CNN will obtain a 2D input and will extract high quality characters from a range of hidden layers (e.g., image or speech). The convolution layer is the heart of a CNN which contains filters with the same input shape but small size. In order to streamline the underling computing, complex networks can involve global or local pooling layers which decrease the data dimensions by integrating neuron cluster outputs into one neuron in one layer in the next. Usually, a RELU layer is the activation function [
78,
79], accompanied by additional convolutions such as pooling layers, FC layers and hidden layers since the activating feature and the final convolution cover their inputs and outputs.
Table 3 shows where the CNN is used in IoT fields.
2.1.4. Transformer-Based Deep Neural Networks
In the deep learning context, the transformer denotes a sequence-to-sequence architecture of neural networks that depend on the self-attention process to capture global dependencies [
112]. It attracted the thoughtfulness of several researchers from the field of natural language processing (NLP) due to the transformer being designed to take sequence data as input. One of the most successful transformer-based models which achieved state-of-the-art performance in many NLP tasks is Bidirectional Encoder Representations from Transformers (BERT) [
113]. Recently, the transformer is also charming progressively standard in the computer vision community. Image classification with the transformer that takes patches of images as input was proposed by Dosovitskiy et al. [
114]. One of the successful project works carried out for an end-to-end object detection framework based on the transformer named it as a detection transformer (DETR) [
115]. DETR simplifies the object detection pipeline by dropping multiple hand-designed components that encode prior knowledge, such as spatial anchors and non-maximal suppression. So, the transformer-based deep neural network is also a promising mechanism to handle the artificial intelligence tasks such as NLP and computer vision related areas.
2.2. Unsupervised Learning
Unsupervised learning must be used as a complement to traditional learning methods to deal with massive unlabeled data. Training can be performed using stacked restricted Boltzmann machines (RBMs) or stacked auto-encoders to initialize, replicate back and modify globally.
2.2.1. Autoencoder (AE)
The AE is a generative method which can be suitable for extracting the features and reducing the size with same number of input and output units. These input and output layers are connected with one or more hidden layers. A neuronal network configured to copy its input to its output is an auto-encoder [
116]. The layer of the code is private (hidden) to show the input. The layer is made up of two major parts: an encryption encoder which maps the code input and a decoder which maps the code in order to decrypt the original input. The auto-encoder is equipped by reducing input-output errors. The AEs are used mainly for diagnostics and fault identification due to their action in creating the input at the output layer. It will shows many application in IoT. Sparse auto-encoders [
117], denoising auto-encoders [
118] and contractive auto-encoders are included in the AE variants.
Figure 7 shows a brief architecture of an auto-encoder and a concrete example.
Table 4 shows where the AE is used in IoT fields.
2.2.2. Restricted Boltzmann Machines (RBMs)
The RBM is a generative method which can work on various kinds of data and its suitable for classifying data, reducing the dimensionality, extracting features, etc. RBMs [
128] are probabilistic graphic models which can be viewed as deep stochastic networks. A Boltzmann version of the RBMs is a constraint on the fact that their neurons can form a bipartite chart; there can be a symmetrical relation between a pair of nodes in both visible and hidden groups. However, there is no connection among the nodes in the same group. In addition, all visible and secret (hidden) neurons are linked to the bias device. It may be necessary to stock RBMs to make DNNs. They are also the backbone of the networks of DBNs. DBNs can be built in particular by stacking RBMs and alternatively fine-tuning the associated deep gradient descent and backpropagation networks. The goal of RBM training is to optimize the product for the visible units in all probabilities. RBM has a similar feature to AEs, which is used to measure the latent parameters, which are used in turn to reconstruct the data input with the backward stream.
Figure 8 shows the structure of an RBM.
Table 5 shows the where RBM used in IoT fields.
2.2.3. Deep Belief Networks (DBNs)
The DBN is a generative method which can work on various types of data. DBNs can be seen as a combination of basic, unsupervised networks (e.g., RBMs and AEs), where a hidden layer of each sub-network is used as a visible layer for the next one. Such a network has connections between the layers, but not inside the layer. DBNs can also be trained greedily layer by layer. This composition leads to a rapid and unregulated training process, which is carried out by the “lowest” layers, where contrasting divergence is applied in turn for each sub-network. The DBN training is performed layer by layer to view each layer as an RBM trained above the previously trained layer. Hence, DBN can be fast and efficient in DL methods. The first is intended for learning about data processing with unlabeled data, and the second attempts to achieve an optimum solution through the harmonization’s of DBN with marked data [
135]. DBN combines unsupervised training and supervised methods to create model designs.
Figure 9 shows the structure of a DBN model.
Table 6 shows the where the DBN is used in IoT fields.
3. IoT Applications and Challenges
The data analysis leads significantly to IoT as discussed in the previous section. In this section, we first review the IoT data features and its applications. Then, we review several issues (challenges) important for the implementation and development of IoT analysis from the point of view of DL.
3.1. Data Features of IoT
As data is the basis for the extraction of knowledge, high-quality information is important. The IoT has many features and is a complex program. The features vary from domain to domain. Here, some features are discussed.
Connectivity allows the Internet of Things to bring ordinary objects together. Such objects are important for their communication since simple interactions of the object level lead to IoT collective intelligence. With this connection, the networking of smart devices and applications will build new business opportunities for the IoT. The key operation of the IoT is the gathering of data from the world, accomplished through the complex changes across the devices. Dynamically, the state of such devices varies, such as sleep and wake up, connection and/or disconnection and the contexts of devices such as temperature, position and speed and also number of machines can change with time, place and person.
IoT is nothing without
sensors which can detect or quantify any changes in the environment in order to produce data that can report on their status or even communicate with the environment. IoT sensors and machine learning techniques combined have taken a major role in health informative systems such as discovering heart failure, lung infections, brain movement and many more [
142,
143]. The sensing data give a rich view to the dynamic world, even though it is basically equivalent to the input from the physical world. Sensors can be used in many applications including our daily life activities. For example, an Automatic Aircraft Control System is made up of multiple sensors that are used for a variety of activities such as speed control, height monitoring, position tracking, door status, avoid obstacles, fuel level, navigating, and more. A computer analyzes the data from all of these sensors by comparing it to predetermined values. IoT is
smart because of its group of smart computing methods, software and hardware. Despite smart technology’s widespread popularity, IoT intelligence is only concerned with the interaction between devices while traditional input methods and visual user interface ensure the user and system interaction. Securing endpoints, networks and data that are passed over all of this means developing a
security framework is critical.
Several papers defined the overall characteristics of big data in terms of volume, speed and variety from different aspects [
144,
145,
146]. However, to characterize the IoT big data by the following 6V features:
Volume: In IoT, a billion devices will generate the huge data.
Velocity: How the IoT data can be accessed quickly and efficiently in real time?
Variety: Basically, IoT data is text, video, audio, sensor data, etc. It may be structured or unstructured data.
Veracity: Refers to the accuracy, consistency and trust of data, which leads to precise analytics in effect.
Variability: Basically, Data flow rate depends on IoT applications, generating data components, time and space.
Value: To transform IoT big data into useful information and insights that offer many advantages to organizations.
3.2. Deep Learning Using IoT Devices
The availability of the latest IoT frameworks and their open source libraries for continuous monitoring, real-time edge-level processing and encrypted storage of generated data such as text, tabular data, audio and video leads to enormous increasing rate of IoT datasets [
147]. Such data are produced by diverse hardware systems working in outdoor and indoor ground-works which includes smart city sensors, smart organization fields, AR/VR practice centers, etc. In order to do the training of such large scale high-quality IoT datasets that have been collected over a period of time within a reasonable amount of time, we need a distributed training system which is scalable and efficiently utilizes the hardware resources of millions of IoT devices. Specifically, such a system environment should consider the current network connectivity among these devices, and be able to work together during training to generate the final deep learning (DL) models at very high speeds for real-time problem-solving [
148].
The authors in [
149] proposed distributed training on multiple IoT devices instead of following the traditional approach that loads such large scale datasets to train and build a model locally within a data center or GPU cluster. In this method, instead of using a GPU cluster available within a data center, the DL model trains and builds on the hardware of millions of medium-sized IoT devices across the context of the infrastructure. They addressed model convergence of the subsequently generated model and scalability of the system. The key issues during involvement of the all IoT devices for training are privacy to the data, time consuming dataset loading I/O, the slow exchange of model gradients while training, and high computational operations. These are the some of the challenges yet to address elaborately in order to train and build a DL model using global infrastructure.
3.3. Applications of IoT
The IoT application is classified according to its basic attributes and characteristics. Some problems should be taken into consideration for the effective operation of IoT data analysis. Some of the IoT applications shown in
Figure 10. The IoT applications may be categorized in the following ways:
Smart Home: Probably, the first application of the IoT is smart home. As per the IoT analytics, more than 70,000 people are searching about the ‘smart home’ every month. Many big companies funding the IoT startup for smart home projects. The smart home appliances include washing machines, refrigerators, bulbs, fans, televisions, smart doors which can built and communicate online each other with approves users to provide better monitoring and managing the appliances and also optimizing the energy consumption.
Smart City: The hypothesis of the optimized traffic system I mentioned earlier, is one of the many aspects that constitute a smart city. This category is most specific to the cities. Mostly, the problems are common in all cities. However, sometimes, they may vary city to city. Global problems are also emerging in numerous cities, including safe drinking water, declining air quality and rising urban density. The IoT applications in city areas are water management, waste management, security, climate monitor, traffic management, etc. We can reduce the noise, pollution, accidents, parking problems, street light problems and public transport because of the smart transportation in cities.
Health care: Relevant real-world knowledge is missing in the tools of modern medical science. It uses the remaining data, managed environments and medical examination volunteers mainly. By research, real-time field data, and testing, IoT opens the door to a sea of useful data. To improve the health of a patient, new technologies have been developed using the IoT in the medical field [
150,
151]. The sensors can monitor a wound’s state, blood pressure, heart rate, sugar and oxygen levels, body temperature, etc., without the presence of the doctors and medical practitioners. In the article [
152], physiological signals are instantaneous and sensitive to neurological changes caused by the cognitive load imposed by diverse driving conditions, and are used to assess the relationship between neurological results and driving environments.
Security: IoT can improve security everywhere in the world using smart cameras. Smart security systems can identify criminals or avoid dangerous situations by means of real-time digital image recognition. Security is the biggest challenge in the IoT field.
Smart Retail: It is one of the biggest applications in the IoT field. Solutions for tracking goods while they are on the road, or getting suppliers to exchange inventory information have been on the market for years. However, it is also limited. The use of intelligent GPS and RFID technologies makes it easy to track the product between the output and the store and greatly reduce costs and time. The applications of IoT in retail are tracking location, inventory management, equipment maintenance, analyzing mall traffic, etc.
Agriculture: Many researchers have already worked in this emerging application of IoT [
153,
154]. Through the growing use of the IoT, connected devices have penetrated everything from health and well-being to home automation, car and logistics to intelligent cities, security, retail and industrial IoT. However, since farming operations are remote and there are many resources that the IoT can monitored, the way farmers operate can be completely changed. Here, the major problem is to change farmers to smart farming. They can be benefited in many ways such as checking soil quality, weather conditions, cost management, reducing wastage, managing crop etc.
Wearables: Now a days, we can see wearables with anyone which can monitor heart rate, sugar and oxygen levels, blood pressure, temperature, sleeping status, walk distance, etc. Wearable technology is an excellent aspect in IoT applications and is undoubtedly one of the first industries to use IoT.
Industrial Automation: Remote access and control are enabled with industrial IoT networking, but more significantly data extraction, processing, sharing and analysis by various data sources. This has tremendous productivity and performance improvement potential. Their low cost and rapid development characterize the IIoT solutions. In order to achieve a better result in cost and customer service, IoT Applications can also re-engineer devices and their packages with IoT automation easily. Some applications are product flow monitoring, digitization, quality control, safety and security, package optimization, logistics and supply chain optimization.
3.4. Challenges
Data sources are a foundation for the success of DL methods. To apply DL to IoT is having a problem with the lack of big datasets, to make DL more accurate, we need more data. Another difficulty in IoT applications is to generate raw data in a suitable form to be fed into DL models. Many DL methods need preprocessing data to get more accurate results. For IoT applications, preprocessing is more complex since the system deals with data from different sources that may have various formats and distributions while showing missing data. The way data collection systems are applied is really a vital research topic. The number of sensors working and the way the sensors are deployed have an influence on data quality. Even if the model architecture is well built, you must build a data collection module for the entire IoT system layout. It should be more reliable, cost effective and trustworthy model.
Security is the biggest challenge in the IoT filed as we collect data from many sources. In many IoT applications, maintaining data protection and confidentiality is a major concern as IoT large data is distributed for review through the Internet, making this accessible worldwide. In several applications, anonymization can be used, but these methods can be exploited and re-identified as anonymized data. DL models are learning the characteristics of raw data and thus can benefit from any invalid data stream. Here, DL models must be updated using certain methods for finding irregular or invalid data.
For IoT system designers, designing DL is a great challenge to meet the needs of managing DNNs on resource-restricted devices. This is expected to increase as the dimensions of the datasets expand daily and new algorithms are included in DL’s IoT solutions. DL also has many limitations. The authors [
155] published on DDN’s false faith in human-recognizable images. The other drawback is that DL models concentrate on classification, while other IoT applications have a sort of regression in their analysis core. Few researchers attempted to introduce regression capabilities to DNNs, for example, in [
156] proposing the DBN and Support Vector Regression (SVR) ensemble.
Off-road vehicles’ digital monitoring is hampered by their sophisticated and pricey IoT sensor technologies. In remote off-network locations, the high reliance on cloud/fog compute, network availability, and expert knowledge make it a handicap. The answer that has yet to be commercialized is the use of edge devices, such as smartphones, with computation capability. The researchers offer a hybridized computational intelligence technique for developing an edge-device-enabled AI system for off-road vehicle health monitoring and diagnosis (HM&D) using super-cheap microphones as sensors [
157]. The authors have shown their results as the taxonomy clearly demonstrates how an EC method may be utilized to improve and optimize DL. Furthermore, this survey addresses potential research avenues that could lead to the development of EDL in the future [
158].
Deep learning is a strong tool to process IoT big data and thus requires high-level hardware requirements. The design of a DL model of an embedded system with limitations on resources remains a challenge. We may get network failure, data disclosure while collecting and transferring data to the servers and analyzed. A movement to develop a cloud-based learning framework that includes leading devices and the cloud is emerging. A cloud-based device will use the edge to reduce delay, maximize safety and protection and use smart techniques for data retention [
159]. It can also use the cloud to exchange data around cutting edges and to train high-quality computational models [
160].
4. Conclusions
In this paper, a review has been presented on the DL and IoT techniques exploited in various domains such as smart home, smart city, smart transport, energy, localization, health sector, security, agriculture, etc. In recent years, DL and IoT have attracted the attention of researchers and business units, both of which have shown their positive impact on our life, cities, and the earth. Many IoT applications are obviously supported by DL resources. DL models are effective methods for solving large-scale problems with data analysis. We addressed the issue of training and building the DL model using large scale datasets which are being produced at ever increasing rates due to the availability of the latest IoT frameworks and open source libraries to collect the same. The literature suggested that using distributed IoT devices themselves for training a model is better than centralized cluster-like infrastructure. However, distributed approach needs to address the challenges such as data privacy, time consuming for IO operations and high complex operations. We reviewed the latest research on how supervised (RNN, LSTM, Transformer-based deep neural networks and CNN) and unsupervised (AE, RBM and DBN) can create a profound DL model for IoT applications. Deep learning ensures that attempts to construct specific characteristics are futile. In addition, the major advances in various domains have been made with IoT and DL, while a further development is expected in the next few years. Moreover, how to design a highly accurate and resource-efficient architecture remains a challenge, and exploration of this area has not yet ended.