Next Article in Journal
Speckle Filtering of GF-3 Polarimetric SAR Data with Joint Restriction Principle
Next Article in Special Issue
Design and Implementation of a Wireless Sensor and Actuator Network to Support the Intelligent Control of Efficient Energy Usage
Previous Article in Journal
Vibro-Perception of Optical Bio-Inspired Fiber-Skin
Previous Article in Special Issue
How to Stop Disagreeing and Start Cooperatingin the Presence of Asymmetric Packet Loss
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Combining Fog Computing with Sensor Mote Machine Learning for Industrial IoT

Department of Information Systems and Technology, Mid Sweden University, 851 70 Sundsvall, Sweden
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(5), 1532; https://doi.org/10.3390/s18051532
Submission received: 10 March 2018 / Revised: 27 April 2018 / Accepted: 9 May 2018 / Published: 12 May 2018
(This article belongs to the Special Issue Dependable Monitoring in Wireless Sensor Networks)

Abstract

:
Digitalization is a global trend becoming ever more important to our connected and sustainable society. This trend also affects industry where the Industrial Internet of Things is an important part, and there is a need to conserve spectrum as well as energy when communicating data to a fog or cloud back-end system. In this paper we investigate the benefits of fog computing by proposing a novel distributed learning model on the sensor device and simulating the data stream in the fog, instead of transmitting all raw sensor values to the cloud back-end. To save energy and to communicate as few packets as possible, the updated parameters of the learned model at the sensor device are communicated in longer time intervals to a fog computing system. The proposed framework is implemented and tested in a real world testbed in order to make quantitative measurements and evaluate the system. Our results show that the proposed model can achieve a 98% decrease in the number of packets sent over the wireless link, and the fog node can still simulate the data stream with an acceptable accuracy of 97%. We also observe an end-to-end delay of 180 ms in our proposed three-layer framework. Hence, the framework shows that a combination of fog and cloud computing with a distributed data modeling at the sensor device for wireless sensor networks can be beneficial for Industrial Internet of Things applications.

1. Introduction

A current global trend is the digitalization of all aspects of our world. An example is applications that can utilize information from sensors attached to things in order to provide automatized operation and imitate intelligent behavior. This concept is commonly referred to as the Internet-of-Things (IoT) [1], and its application areas have spread to include industrial applications [2]. Most IoT solutions today are based on traditional client-server architectures with cloud back-ends, attached data analysis, visualization, and end user data access on the cloud side. This approach does, however, have some limitations, such as the cloud servers becoming single points of failure and an excessive delay when sending data to a cloud server far away on the Internet. A number of workarounds are used to avoid and address these limitations. One of the most recent workarounds is adding a fog layer to the cloud system [3]. When the fog servers act as cloud systems closer to the edges of the network, it can offload computationally heavy tasks of the IoT devices.
In the industry, wireless sensor networks (WSNs) have been a reality for some time, and research is being carried out on topics such as energy awareness, ad hoc routing, and reliable communication. WSNs are implemented through deployment of motes in usually harsh and/or hardly accessible environments. The sensor motes are mainly battery-powered and they are expected to remain active for a long period of time without human supervision and maintenance. Thus, it is preferable to expand a network’s lifetime by optimizing the energy consumption at mote level. Radio is the main source of energy consumption in motes, while idle listening, which requires as much energy as receiving, and overhearing are the main sources causing unnecessary draining of the battery. A common way to avoid unnecessary energy consumption is by using duty cycle management with Medium Access Control (MAC) protocols [4,5,6,7,8], where the battery life is extended by keeping the mote in sleep mode for most of the time. This solution, although effective for energy consumption, introduces a new problem: with a lower duty cycle the requirements of real-time applications cannot be met. Some research has proposed energy aware routing methods to solve this issue [4,9,10]. While recent solutions try to balance the trade-off between energy consumption and delay, the aim is to move away from duty cycle based MAC protocols towards wake-up radio systems [11]. Another challenge of WSNs concerns scalability. That is, how to reduce the communication for hundreds of sensor motes residing in the same network, and to better utilize the limited available bandwidth. This problem becomes more relevant in industrial scenarios. In coexistence with other wireless technologies such as 802.11 b/g/n/ac, WSN is prone to communication failure and packet loss, which results in retransmissions.
In recent years, many studies have been conducted to solve the aforementioned problems utilizing various methods of data mining and machine learning on large volumes of data generated by numerous motes in WSNs. However, while the availability of data in WSNs is promising for mining algorithms, the motes’ limited resources and volatility of generated data streams in WSNs pose great challenges for the traditional data mining algorithms. To overcome these challenges, many studies modified traditional data mining algorithms, to make them more suitable for deployment in WSNs, or suggested innovative approaches that consider the prerequisites of WSNs in the design phase [12]. There have been many important contributions in this field, but an important problem still exists: in literature, experiments are usually conducted with the assumption of access to a prior knowledge of the system, ground truth, and availability of data from all the resources (sensor devices) at any chosen time. These assumptions limit deployment of suggested models in real industrial environments, where sparsity of the collected data and dynamic system behavior are the norm.
This article will explore how to solve some of the aforementioned problems. More specifically, how to save spectrum for the most urgent communications, and save energy in the motes by preventing unnecessary uplink transmission. We propose a distributed modeling method in a three-layer framework for Industrial IoT (IIoT) systems, consisting of a cloud back-end, a fog middle layer, and a lower WSN layer, see Figure 1. On the sensor layer we try to reduce energy consumption and unnecessary uplink transmissions. We achieve this by first creating a model on the sensor devices by mining the collected data that approximates the data stream behavior, and then transmitting only the updated model parameters to a fog computing system. To further enhance this, we will also utilize a distributed fog system to reduce the amount of data required to be sent to a large cloud system. We also investigate how introducing an IIoT can be beneficial to industrial monitoring systems. Hence, the research questions we aim to address in this work are:
  • To what extent can we save energy and reduce the number of uplink transmissions at the sensor motes, by introducing a model learned from data stream?
  • Can this model be practically implemented and utilized in a fog computing architecture?
  • What are the expected end-to-end delay and information query times of such a system, and can it be used in industrial scenarios?
The remainder of this article is structured as follows: Section 2 presents an overview of our approach and related work. Section 3 gives the details of our approach, including modeling the data stream at sensor devices and the distributed modeling process in the fog. Section 4 describes the details of the implemented testbed system. Section 5 presents the used hardware and measurement methods. Section 6 presents the results and evaluation of the proposed model, and the resulting system and their implications. Finally, Section 7 presents the conclusions and future work.

2. Data Mining and Machine Learning in WSNs

Tools and algorithms in data mining and machine learning have been used in many recent studies to create predictive models in sensor networks [13]. Kernel linear regression was used in an in-network approach where the sensor network tries to optimally fit a global function to each of the sensors’ local measurements [14]. The learned linear model takes the form of a weighted sum of local basis functions given by users, which provides an expressive yet tractable class of models for sensor network data.
Extended Kalman filters (EKFs) [15] have also been widely used in related works to predict the sensor value. To be able to use Kalman filters, the user needs to provide the next state function, the measurement function and noise term. In [16], a family of algorithms for training the noise parameters of an EKF in a sensor were proposed. The algorithms adjust the covariances of the EKF in a way that maximizes their prediction accuracy. To make use of the spatial information of the sensor nodes, distributed Kalman filtering estimates the system state based on the information not only from itself but also from its neighboring sensors according to the networks topology. In [17], an event-based distributed filter has been proposed in wireless sensor networks to reduce the sensor data transmission rate and the energy consumption. Each node makes decisions independently about when local messages need to be broadcast. While the presented results are promising, the paper does not provide any suggestions for further communications with cloud or fog nodes. Hence, filtering is only used to reduce number of transmissions and no mechanism is suggested to simulate the data stream in the central monitoring unit.
Dividing the sequences into meaningful subsequences and representing them by using real-valued functions is one of the most common and fundamental components of time series data mining [18]. In [19], trend change is detected by using Bayesian inference; the patterns are learned offline; finding the most accurate trend changing point is an optimization problem. In [20], evolutionary computation was suggested for time series segmentation of stock price data. A series segmentation was proposed by [21], which is achieved by a polynomial approximation of the time series using a basis of orthogonal polynomials. These methods require repetition and a relatively large storage, which makes deployment in WSNs infeasible.
A hierarchical three-layer deployment for energy efficient IIoT architecture is proposed in [22]. The high level architecture consists of a cloud server, a Representational State Transfer (REST) service, as well as sensing entities. The study mainly focuses on the sensing entities to efficiently reduce the energy consumption by optimizing their sensing, processing and communications. For this purpose, a three-layer framework is proposed, which divides nodes by functionality and extends network life by balancing traffic load and predicting sleep intervals. Although the framework in [22] carries similarities with our proposed framework, in terms of architecture and energy efficiency, the methodological aspect differs vastly. In [22], semi-centralized smart scheduling is deployed to achieve energy efficiency, whereas our approach utilizes a distributed adaptive data-driven modeling and event-based communication at the sensor motes to achieve the same goal.
Some recent articles focus on the communication and distribution problems that appear between sensor and fog devices. Li et al. [23] highlights scheduling problems in deep learning IoT edge devices, including performance evaluation of their proposed scheduling. Dautov et al. [24] also highlights distribution and communication problems when performing stream analytics on edge devices. They also investigate the differences of having traditional vertical device and a cloud structure, compared to a more horizontal device, fog, and cloud system. To the best of our knowledge there has so far not been any studies deploying a distributed data model utilizing fog computing for monitoring systems in IIoT.

Approach

We aim to create a three layer framework for an industrial monitoring system to investigate and answer the previously presented research questions. An overview of the approach is found in Figure 1. The bottom layer consist of resource constraint sensor devices collecting field data, and the top layer is a cloud back-end with a view over the whole system. The fog layer connects the distributed view of the WSN layer to the overall view of the cloud layer. At the lower layer, a sensor device starts by learning parameters that describe the data stream in the best way to create an initial model. Then it continues to monitor the data stream and update the model parameters. When the sensor device detects a trend change or a necessary model parameter update, the new information will be sent to the associated fog node. The fog node will simulate the data streams based on the model with the received updated parameters. Through collection of updated model parameters from the associated sensor devices, a synthesis directed probability graph can be created in the fog node. This graph provides the fog node with a view of the operational state on the associated monitoring area, making it possible to detect local anomalies, trends and sensor devices’ faulty behavior. The global model of the directed graph is constructed in the cloud by collecting all local graphs from the fog nodes. Anomalies, changes in trend and faults in the system can be detected in the cloud by monitoring the parameter changes in the global graph.
In the next section we present the data streams learning and monitoring model for the sensor network layer and fog computing layer. With respect to the research questions stated in Section 1, the construction of the global graph in the cloud computing layer, and a more sophisticated fault detection system are out of the scope of the current study, and we leave these tasks for future work.

3. Proposed Data Streams Learning and Monitoring Model

This approach aims to reduce the communication between sensor motes and fog nodes to save energy and spectrum. Therefore, we divide the problem into learning and monitoring in the sensor device, and simulation of the data streams in the fog node. A sensor device learns an initial model from the data streams and when it detects a trend change or a necessary model parameter update, the new information will be sent to the associated fog node. The fog node will simulate the data streams based on the model with the received updated parameters. In the sensor motes residing in the WSN layer, we aim to utilize a novel energy saving and packet reducing technique, to address the first research problem. This will be done in order to make the motes last longer on battery and to send as little data as possible to open up for higher priority data transmission. Instead of sending a large number of small data packets, the sensor devices will calculate parameters of a mathematical model from the raw values and send the updated parameters of the model to the fog. This means sending new data only when changes are detected and the model needs to be updated. In the fog layer an approximation of the original data stream is simulated by using the updated model parameters for a duration of one segment, which is the interval between the two last packets, received from sensor devices in the WSN layer. Furthermore, the information from the fog layer is sent to the cloud to provide a view of the whole system and other services required by end users. A flow chart of the learning process in the sensor device and the simulation process in the fog node can be seen in Figure 2.

3.1. Learning and Monitoring in the Sensor Device

The operation of a sensor device consists of two phases, which can also been seen in the flow chart. Namely: The initialization phase, and the monitoring phase.

3.1.1. Initialization Phase

In the initialization phase, the sensor device collects data stream values for a certain number of data points m, the maximum segment length, to set the parameters of an initial linear regression model. In this phase we assume that the system behavior is normal, that there are no anomalies or errors in the collected data. A sensor device normalizes the collected values and calculates a reference value for a set of statistical information, ζ , which will be required to detect changes in the data stream and for the fog node to simulate the segments of the data stream. For each segment, the sensor device extracts the following statistical information.
The reference slope
A r e f = ( x m 1 x 0 ) / m ;
The sum of the values:
s u m X = i = 0 m 1 x i ;
The second moment (surprise number [18]):
s u r p r i s e X = i = 0 m 1 x i 2 ;
And the step calculated as:
s t e p X = i = 0 m 1 i x i .
By the end of the initial phase, the sensor device, sends the initial model parameters, A r e f , the first value of the first segment x r e f and starting time point t 0 to the fog node. The process of the sensor device in the initialization phase is presented in Algorithm 1.
Algorithm 1: Initialization Phase.
Sensors 18 01532 i001

3.1.2. Monitoring Phase

In this phase a sensor device learns the dynamics of the data stream, updates the model parameters, detects changes, and communicates this information to the fog node. Algorithm 2 represents the process of the monitoring phase. The sensor device carries out the task by reading the data stream one time point at the time, for maximum of m time points, which is the maximum length of each segment. The maximum length of each segment is the time interval between two consecutive packet transmissions. This is to overcome the issue of limited memory on the device, and to meet the requirements of minimum number of required communications in WSNs. At each time point, the sensor device reads the sensor value, updates the statistical information and predicts the behavior of the data stream in the next time point. When reading the next sensor value, the device calculates the root square error ( R S E ) of its prediction. If the prediction error is negligible and does not reach the specific prediction error threshold θ , the data stream shows a stable behavior, meaning that no change in trend has been detected. The sensor device then updates the model parameters and transmits a packet at the end of the current segment consisting of the statistical information ζ i ( A i , x 0 i , n i , s u m X i , s u r p r i s e X i , s t e p X i ) to the fog node. When the prediction error for any time point is larger than θ , it is an indication of a possible change in system behavior. The sensor distinguishes a trend from a state change, by checking the error value of the predicted trend for the current segment, and the error value of the predicted trend with respect to the reference trend, A r e f . If the evidence indicates that a trend has occurred, the sensor device updates the model parameters and transmits them in the scheduled transmission, at the end of the current segment. In cases where the detected change indicates a new state, the sensor device ends the current segment, transmits ζ to the fog node and starts the initialization phase to identify the new model parameters.
Algorithm 2: Monitoring Phase.
Sensors 18 01532 i002
Having given an overview of the proposed learning and monitoring model, we formulate the model of the sensor device. Let X = { x 0 , x 1 , x 2 , x T } be a data stream sampled at uniform and ordered time points t = 0 , , T . Then any time point k, 0 < k < t , of a polynomial temporally correlated data stream, with respect to a starting point t 0 , can be represented as:
f ( t ) = i = 0 k a i ( t i t 0 ) + d ( t 0 ) + ϵ ;
where a i and d are constant coefficients describing the characteristics of the stream, slope and intercept respectively, and ϵ is a small random value with standard normal distribution. According to the assumption of uniform and ordered time points we have t 1 t 0 = t i t i 1 , for i = 1 , , n . We also know that according to Equation (5) for any segment with length l = 1 , , k and k < n :
f 0 ( t n ) = f ( t n ) ,
f l ( t n ) = f l 1 ( t n ) f l 1 ( t n 1 ) .
If a i ( t i t 0 ) = A r e f , where A r e f is the reference value of the slope (trend), to model the data stream behavior and estimate the next step, it is reasonable to assume that in the absence of a trend change in a data stream, Equation (5) for the nth time point, can be simplified to:
f ( t n ) = A r e f + ϵ ,
and, prediction of the next time point, n + 1 , for any chosen point in the data stream segment, f k , can be approximated by:
f k ( t n + 1 ) A r e f .
So the next value of the lth segment can be approximated as:
f l ( t n + 1 ) f l 1 ( t n + 1 ) + f l 1 ( t n ) .
It should be mentioned that we do not impose any restriction on the interval value between two consecutive time points, as long as the intervals are the same for all the points in the data stream. Indeed, in Section 6, we show that our proposed model performs well on various data streams with different sampling rates. Another task of a sensor device is to predict the trend for each segment. For a segment f l ( t n ) in any given data stream, a trend can be calculated as:
A l = E [ f l ( t n ) ] ;
where E [ . ] is the average function, and A l is the segment’s trend. The problem arises when considering the existing random noise in collected sensor values, which poses negative effect on the approximation accuracy. This negative effect makes it more difficult to predict trends or detect new states, especially when the standard deviation of a data stream, V A R ( f ( t n ) ) = σ ), is a large value, or a value close to the prediction error threshold θ . However, since the random noise is independent, the standard deviation function can be rewritten as:
V A R ( i = 1 m ( f l ( t n i + 1 ) ) = m ( V A R ( f l ( t n ) ) ) = m σ .
Hence, we can rewrite the estimation of trend as:
A l = E [ i = 1 m f l ( t n i + 1 ) m ] ,
with σ m as the standard error of the estimation.
At the end of each segment the sensor device sends statistical information of the normalized segment, f l ( t i ) , to the fog node. As mentioned previously, one of the main challenges for a sensor device is the limited resources, namely memory and processing. To overcome this problem, a stepwise process updates the model parameters. In other words, instead of iterating over all values in the segment to update the required parameters, a sensor keeps one value for each of the pieces of statistical information of a segment and updates the values as follows:
s u m X = s u m X + x n ,
s u r p r i s e X = s u r p r i s e X + x n 2 ,
s t e p X = s t e p X + n x n .

3.2. Simulation in the Fog Node

The main task of a fog node is to collect the summarized information from the associated sensor devices, and further merge these small segments into a longer trend segment and different states. The complex system of multi-nonlinear data streams can be simply described as a succession of sequential multi-linear streams, which can be described by a probability graph network model. The model is defined as a directed graph G = ( V , E ) , where V = s i is a set of vertices or nodes with state s i . Each state s i , represents multiple data streams in which the temporal correlation of each of the streams has not changed in a specific time interval. A state can be defined as:
s i = { D i , j ( T ) | j { i 1 , , i k } : D i , k ( [ t s , t e ] ) F D i , j ; T [ t s t a r t , t e n d ] } ,
where D i , j ( T ) is a finite subset of the data stream D i , j for a time interval T ( t s t a r t < t e n d ), and F D i , j is a linear regression model that describes one sub data stream. An edge of the graph, E = { < s i , s j , p i j > } , is a set of links that represent a switch from one node to another with some probability p i j :
p i j = P ( D [ t 3 , t 4 ] s j | D [ t 1 , t 2 ] s i t 3 : t 3 t 2 > δ D [ t 1 , t 3 ] s j ) .
Each state, s i , provides summary information about the nature of a subset of multiple data streams. After receiving statistical information from several sensor devices, a fog node constructs the graph by merging several short consecutive segments with the same trend, and adding a new node to the graph, that is if the node does not exist in the graph. If the node already exists, the fog node only updates the summary information of the state. To distinguish between a new state and an already existing state, a fog node compares the mean square error of the new trend A i with an acceptable error threshold δ . Given the summary information of the segment ζ i ( A i , x 0 i , n i , s u m X i , s u r p r i s e X i , s t e p X i ) , the difference will be:
s u r p r i s e X n i 2 x 0 i s u m X n i 2 A i n i s t e p X + x 0 i 2 + ( n i + 1 ) A i + A i 2 ( n i 2 3 + n i 2 + 1 6 ) ,
where A i is calculated as:
A i = 2 n i s t e p X i ( n i + 1 ) 2 ( n i 2 3 + n i 2 + 1 6 ) .
So the segments i s t a r t to i e n d can be merged if the following condition is true:
s u r p r i s e X n i 2 x 0 s u m X n i + x 0 i 2 ( 2 n i s t e p X ( n i + 1 ) ) 2 4 ( n i 2 3 + n i 2 + 1 6 ) < δ .
Although the process of the cloud node is out of the scope of this work, it is easily conceivable that a process similar to the fog node can be carried out in the cloud to construct the meta graph and provide a complete overview of the system.

4. The Testbed System

We have chosen to implement and test our approach as a real world implementation in a testbed system. Figure 3 shows an overview of this testbed system. The testbed system is split into three layers: the sensor network layer, the fog computing layer, and the cloud computing layer.
The sensor layer at the bottom is a wireless sensor network consisting of motes with both sensor devices and a gateway. It is in the sensor devices that the raw sensor data is collected for the proposed system, but these sensor devices could have very limited resources and power. The raw sensor values are made into the sensor data model by the sensor device, in accordance with the algorithm explained in Section 3. The model and updates to the model are then sent over IEEE 802.11.15.4 industrial wireless communication to a gateway node connected via USB to a fog computing device in order to push the information further up the system. The second layer is the fog computing layer, which has been implemented using small resource constrained computers. The purpose of the fog computing layer is to reduce the amount of data sent to the cloud, provide close to edge computational capabilities, and reduce the communication delay to end users, since they do not have to go through the cloud to get to the sensor values. The data model is received from the gateway mote, from which sensor values are generated and pushed upwards to the cloud at regular intervals. The final and top layer is the cloud computing layer. This cloud serves as a persistent storage point with scalable storage and computational power. This layer was implemented by using an IoT cloud platform to store the data in a database and display the sensor values in a more user-friendly manners.

4.1. Sensor Network Layer

Two types of motes are programmed: gateway and sensor. The gateway works as a connection between the sensor network layer and the fog computing layer; hence its role is limited to collecting data from the payload of packets received from sensors. The sensor motes are programmed to imitate the algorithm presented in Section 3. Correspondingly, sensor devices’ functionality is divided into three states, implemented utilizing the multi-threading module of Contiki [25] to switch between threads, according to the sensor device’s current state in the algorithm. Upon joining the network, sensor devices continuously collect sensor values, such as temperature and illumination, and save them in an internal flash memory for a user-defined sampling duration, w i n i t . The sensor device sends the first unicast message to the gateway when the sampling duration ends. The payload contains the minimum ( m i n i n i t ) and maximum ( m a x i n i t ) values observed during this time. When the sampling process is complete, the sensor device maintains a fixed length list, containing the most recent collected sensor values, in an internal flash memory. Motes are programmed to sample data periodically. The sampling frequency is set to one sample per second, which is a common choice for industrial monitoring systems. Each sampling event triggers a chain of mote-level calculations and evaluations to update the pre-defined relevant model parameters, explained in Section 3, and the change point detection. Unicast messages are sent to the gateway with updated values of model parameters, either if the current prediction error exceeds the hard-coded error threshold δ , or a sudden change within the accepted error range is detected (that is, the error has not exceeded δ value, but the change is significant enough to indicate a new period.)

4.2. Fog Computing Layer

The fog computing layer was implemented using small resource-constrained computers, namely Raspberry Pi devices. The purpose of the fog device is to interpret the model from the sensor motes, generate sensor values to send to the cloud, and act as a direct access point for the sensor values. A Java program was written to run on the fog devices for this purpose. To be specific, the program is split into four parts, which are implemented as concurrent threads. The sensor mote reader thread, the model interpreter thread, the cloud publisher thread and the REST interface thread. The sensor mote reader part reads from the gateway mote, which is connected via USB to the Raspberry Pi. The gateway mote created a virtual serial port, which is then used to communicate between the fog and gateway. When the gateway mote receives new data from the sensor mote, it sends the data model parameters to the gateway to be further interpreted. These model parameters include a device ID and the values for s u r p r i s e X , s u m X , s t e p X , X m i n , X m a x , x i 0 and A i . The model interpreter uses these values to generate sensor values from the normalized model values using the formula Y = X m i n + x ( X m a x X m i n ) in accordance with the established model parameters. This in turn is handled by the cloud publisher, which publishes sensor values to the cloud via the MQTT protocol on a fixed interval set to every two seconds. Finally, the REST interface is a thread that listens for an incoming HTTP GET connection on port 9999 in order to return the latest sensor value as a JSON object from the model in a HTTP response.

4.3. Cloud Computing Layer

The top layer is the cloud computing layer, which is implemented using the ThingsBoard IoT cloud platform running on a desktop computer as a server system. The purpose of this layer is to act as persistence storage of sensor values in a database and to show the sensor values in more user-friendly graphs and tables. The ThingsBoard IoT cloud system has many of these functions built in, including an MQTT broker and an MQTT client to listen and handle MQTT messages from the fog devices. Hence, the cloud system simply acts as a back-end storage system and sensor value visualizer, based on the sensor values sent from the fog layer.

5. Materials and Methods

This section will give a detailed description of the hardware and software used to implement the system as a testbed. It will also explain the details of setting up the measurements inside the testbed system.

5.1. Hardware and Software Materials

The hardware we used in the wireless sensor network layer was TelosB [26] motes with an IEEE 802.15.4 complaint transceiver, CC2420 [27] and CSMA/CA medium access control protocol, running the Contiki [25] operating system. The fog computing layer was created using Raspberry Pi model B+ hardware, running the Raspbian operating system version 9. The program running on the Raspberry Pi was written in Java 7, and when the getaway mote was connected to the Raspberry Pi, it created a virtual serial port at 115,200 baudrate, which is used by the Java program to communicate between the gateway mote and the Raspberry Pi. The Raspberry Pi was then connected to a local gigabit Ethernet network created by a Linksys WRT1200AC network router. The cloud computing layer was realized using regular desktop computer hardware, also connected to the Linksys WRT1200AC router. Hence, a Dell Optiplex 780 with a 2.93 GHz Intel Core 2 Duo processor and 4 GB of RAM was used as the cloud machine, running Ubuntu version 16.04 LTS and ThingsBoard 1.3.1. A laptop was also used for some of the measurements. This laptop was a HP Elitebook 1040 G3, with an Intel I7 2.5 GHz processor and 8 GB RAM, connected to the same Ethernet network provided by the Linksys WRT1200AC router.

5.2. Evaluation Methods and Setup

To verify that our testbed system and model are working as expected, a series of measurements and evaluations were performed on the system. Specifically, we decided to measure and evaluate the reduction in the number of packet transmission attempts by sensor devices due to implementation of the proposed model, the end-to-end delay of the proposed framework, the query times of fog and cloud, the scalability of fog, and finally the cost imposed on the sensor devices’ awake time in terms of computation of model parameters.
The reduced number of packet transmissions was measured and evaluated using a mathematical analysis of our model implemented in Matlab. We used three different data streams, collected from sensor devices monitoring various aspects of an industrial environment. All of the data are noisy, and each of them have different sampling rates and show diverse behaviors, see Figure 4. To describe average model-performance, the difference between the simulated data stream at the fog node and the original data stream was measured by Mean Square Error (MSE); this is when the fog node was only provided with the statistical information received form the sensor device. Furthermore, we compared the performance of the proposed model to a moving average model, by comparing their relative errors on simulating a data stream and similar parameters setting.
The total end-to-end delay was evaluated and measured in our testbed system, measuring from sensor node to end user application, running either towards the cloud or the fog system. This delay was measured in three steps. First between the sensor mote to the fog node, secondly between the fog node and the cloud node, and lastly between the cloud and end user application. The total delay time can also be represented by Equation (17), where d t o t a l is the total delay, d s e n s o r is the sensor delay, d s e r i a l is the delay of the reading of the serial communication from the mote, d f o g is the fog delay for sending MQTT messages, and d c o u d is the cloud REST interface delay.
d t o t a l = d s e n s o r + d s e r i a l + d f o g + d c l o u d
The sensor delay was measured on the TelosB motes. The end-to-end delay at the sensor network layer can be measured by considering four elements: transmission delay, propagation delay, processing delay and queuing delay. To compare the effect of added computational overhead at sensor level, we compared the end-to-end delay of a unicast process with the delay of the proposed model. The payload size was the same in both settings. The delay was measured and the performance was evaluated using simulation scenarios, with the same setting as the implementation on TelosB motes, defined in Cooja [28]. The fog to cloud communication was measured on the fog devices with the Java program that published sensor values to the cloud, and finally the cloud system was evaluated using an end user Java application accessing the REST interface of the cloud. The query time of both the fog and cloud systems were also evaluated in the testbed. To measure this we used the measurement laptop running a Java program that performs and measures the query response times of the respective REST interfaces. Finally, we evaluated and measured the scalability of the fog nodes. Since we expect each fog device to handle multiple sensor motes and their models, we had to determine the maximum amount of sensor values that our fog device (Raspberry Pi) could handle. All these measurements were made on the local network, which means that they add up to an indication of what the best case minimum response time of the system can be.

6. Results and Discussion

We present and evaluate the results of the model and the testbed separately. The model results present measurements and evaluations of the sensor model itself, and its effectiveness in terms of reduced number of packet transmissions and reproducibility of the approximated sensor values based on the model parameters at the fog node. The testbed results present the features and measurements on how well the sensor network layer, fog computing layer and cloud computing layer collaborate and perform as one system. The implications of these results are discussed at the end of this section.

6.1. Sensor Data Experiments

The average model performance and the reduced number of packet transmissions were measured and evaluated by implementing the proposed model in Matlab. Forty thousand continuous data points from each of the data streams were sampled.
According to the model explained in Section 3, two parameters of trend threshold θ and maximum number of data points in each segment, as well as segment length m, have the highest effect on the model performance. The reason is that they indicate the level of trade off between energy efficiency, with respect to transmission reduction, and sensitivity of the model in trend and change detection, which has a direct effect on the accuracy of the simulated data stream at the fog node. Several experiments with various values of θ and m for each of the data streams were conducted, to investigate these effects on simulation accuracy at the fog node in terms of mean square error, and to choose the most appropriate set of parameters for the model. Figure 5 presents the results. As can be seen from the graphs, the mean error of the simulated data streams at the fog node has the sharpest decline for all sets of parameters, when the simulations were based on statistical information from two data points, m = 2 . From this point on the results of each of the data streams show different patterns; nonetheless, the best parameter set for all of the data streams, with respect to the MSE, was achieved by { θ = 0.005 , m = 10 } , see Figure 5a–c. At this point for almost all sets of parameters, the trend reversed and has raised steadily. The complete results are summarized in Table 1.
It should be noted that, when choosing the maximum length of the segment, although a greater m means fewer transmissions, hence achieving more energy efficiency, it also means more delay on detecting trend changes and state switches. It also indicates decreased accuracy, since the summary information would be an approximation of a larger number of data points. So although smaller values of m might seem encouraging, for improving the energy efficiency, it is better for them to be excluded to prevent accuracy loss. Looking at Figure 5, it can be seen that the MSE shows a fairly stable trend when segment length increases from 10 to 20, with θ = 0.02 for stream A and θ = 0.01 for streams B and C. For this reason, the parameter set { θ = 0.01 , m = 20 } for streams B and C, and { θ = 0.02 , m = 20 } for A, were chosen to test the performance of the proposed model. After setting the model parameters, several experiments were carried out for all three data streams. The red lines on Figure 6 show the simulated data streams at the fog node, and the blue lines illustrate the original data streams. On the left column of Figure 6a,c,e, the results of the model for the duration of the experiment are presented. To illustrate the results in more detail, Figure 6b,d,f show zoomed in versions of the same data streams. It is clearly observable that the model captures the behavior of the data streams with good accuracy and low delay.
For this experiment, 40,000 data points from each of the data streams were used. The number of packets sent to a fog node after deploying the model reduced to 812, 898 and 926 for A, B and C, respectively; that is a communication reduction ratio of 49 to 1, 45 to 1 and 43 to 1. In total, for all the data streams, only 2636 of the data points (from the original 120,000 data points) were transmitted. Hence, approximately 2.2% of the packets were sent from sensor devices to a fog node. To compare the performance of the proposed method, the MSE measure of the base model was obtained, with respect to the segment length, see Table 2. As mentioned before, the best parameters set for the proposed method was { θ = 0.005 , m = 10 } , with respect to MSE of the simulated data streams at fog node; in addition, the MSE measure was fairly stable with segment length m = 10 to m = 20 . Choosing the parameter set { θ = 0.005 , m = 20 } , we can compare the simulation accuracy of the two models. The MSE measure of the data streams A, B, and C were 0.036, 0.022, and 0.017, respectively, when the proposed model was used to simulate the data streams at the fog node; these values for the base model were A = 0.43, B = 0.025, and C = 0.019. It is clear that the proposed method was able to simulate the data stream with a higher accuracy. In terms of transmission reduction, the proposed model simulated the data streams A, B and C with MSE of 0.038, 0.024 and 0.02, respectively, when transmitting one packet every 20 data points, with the most appropriate parameter set. For the base model to achieve the same level of accuracy, the maximum segment length cannot be more than 17 for stream A, 19 for stream B, and 23 for stream C. This means that approximately 15.5% of all the packets need to be sent, which makes the communication cost of the base model 7.5 times more than the proposed model.
Considering the presented results, it is reasonable to conclude that the proposed model can successfully reduce the number of packet transmissions, while keeping the error of the simulated data stream, using only statistical information, within an acceptable range.

6.2. Testbed Evaluation

The proposed cloud, fog and sensor system works as expected. The sensor mote generates a sensor model from the raw sensor values, which is sent to the sensor gateway to reduce the required wireless communication. The fog system retrieves the model parameters from the sensor network layer, and regenerates the sensor values from the model parameters. The fog system both provides the sensor values directly to the end-user applications via a REST interface and sends the values to the cloud system via the MQTT protocol. On the cloud system, the values are persistently stored and the end user applications can access the sensor values in a more user-friendly manner. Figure 7a shows the sensor motes and fog node with attached sensor gateway. A screenshot of the cloud dashboard with sensor values from the sensor motes via the model and fog device can be seen in Figure 7c. The figure shows three different views of the same sensor value, as a card with exact value, as a digital gauge, and as an animated graph. Figure 7b shows the output from the REST interface running on the fog devices. The output is a simple JSON object with the sensor name and its value.
As previously stated, the end-to-end delay was evaluated and measured in the testbed system. It was measured by adding together each of the steps in the proposed system. Firstly, from sensor node to end user application, secondly between the sensor mote to the fog node, thirdly between the fog node and the cloud node, and lastly between the cloud and end user application.
Each measurement was made 1000 times and the results of the evaluations can be seen in Table 3, where μ denotes the mean and σ denotes the standard deviation of the measurements. Hence, we can see that the total delay is on average 180 ms with a standard deviation of 37 ms. It is, however, important to note that this is made in an optimal network situation where all devices are on the same network.
The query time of both the fog and cloud systems were also evaluated in the testbed. To measure query time we used the measurement laptop running a Java program that performs and measures the query response times of the respective REST interfaces on the fog and cloud. Results can be seen in Table 4.
The experiments on the sensor layer, on both hardware implementation and Cooja simulation, show that the computational overhead has a negligible effect on delay with the chosen precision in the testbed, which is in the order of milliseconds, and the process still fits within the required sampling rate of 1 sample per second. The computational cost adds to processing time within the allocated time slot in magnitude of less than 2 ms. Finally, the scalability of the fog nodes was evaluated. Since we expect each fog device to handle multiple sensor motes and their models, we had to determine the maximum number of sensor values the fog device (Raspberry Pi) could handle from each sensor. According to measurement, the Raspberry Pi used on average 3.4 ms with a standard deviation of 1.8 ms for the serial communication to handle each sensor model update, which translates to about 290 sensor values per second. This means that each fog node can scale up to 290 sensor model updates per second. The realization of the proposed model in the form of the testbed system shows acceptable performance. Considering the presented results it is conceivable that the proposed model can meet the required latency of monitoring systems in industrial scenarios, as it can keep the performance above the required accuracy threshold.

6.3. Discussion

Considering the presented results, the distributed modeling at the WSN layer implemented on sensor devices, clearly reduces communication, hence saving spectrum. The simulation results of the fog layer clearly shows how approximations of the model parameter can regenerate the data stream at the fog node with acceptable accuracy. This is possible by providing the sensor nodes with transmission opportunities when detecting change points, instead of limiting them to a scheduled-based packet transmission. Furthermore, the prediction accuracy is enhanced by using more relevant statistical information about data streams. In addition to the presented results, an advantage of the proposed approach in comparison to other studies in this field, is that the model reduces some of the negative effects of wireless sensor communication on the performance of the learning algorithm, namely, synchronization of motes and high rate of dropped packets that adds additional delays to the learning process, and introduces the missing values problem to the learning algorithm.
It is worth mentioning that the presented model and testbed system are still under development. We consider the presented study a preliminary investigation of benefits of edge computing and its possible contribution to enhance network performance in IIoT. Moreover, we showed that it is possible to design cross-layer frameworks and to develop decentralized system models to implement collaborative systems by facilitating different communication standards. In the presented work, there are aspects that deserve more in-depth investigation. In the WSN layer, the complexity of the model can be reduced to make it more suitable for resource constraint devices. In addition, incorporating a more sophisticated medium access control protocol with prioritized packet handling mechanisms can be used to improve the reliability of the proposed model and expand its applicability over scenarios with various characteristics. Since the fog layer provides a localized view over the network, a fault detection mechanism can be deployed as a step toward decentralized fault detection and possible performance prediction. These are topics for future work.

7. Conclusions

The goal of this research was to answer the three research questions in Section 1. The first question involved investigating to what extent we can save energy and reduce the number of packet transmissions from sensor devices by introducing a data stream model learned at the WSN layer and approximated at the fog layer. We found that by deploying the proposed model, the number of packets sent over the wireless link can be reduced by 98%, with negligible accuracy loss. The second research question was to investigate if this model can be practically implemented and utilized in a fog computing layer. This was shown in an operational testbed system with real-world hardware. Using TelosB motes as sensor devices, Raspberry Pi devices as fog nodes, and ThingsBoard as an IoT cloud. The final research question aimed at investigating the expected end-to-end delay and information query times of such a system, and find out whether it can be used in industrial monitoring scenarios. To test this, measurements were made on the testbed system. The measurements show that the fog devices can handle 290 sensor model updates per second, and have an average of 180 ms delay from end to end. The measurements also provide an information query time of 5.3 ms for the fog, and 8.9 ms for the cloud under ideal network conditions. Hence, the proposed system shows that a combination of fog and cloud computing including a distributed data stream modeling for wireless sensor networks shows great promise for Industrial Internet of Things applications.

Author Contributions

The idea and the article was a collaboration between all authors. However, T.Z. verified the mathematical model; S.F. lead the work; M.L., S.F. and U.J. realized different parts of the design, produced the implementation, and performed the measurements.

Funding

This research was supported by grant 20150363, 20140321, and 20140319 of the Swedish Knowledge Foundation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Atzori, L.; Iera, A.; Morabito, G. The internet of things: A survey. Comput. Netw. 2010, 54, 2787–2805. [Google Scholar] [CrossRef]
  2. Da Xu, L.; He, W.; Li, S. Internet of things in industries: A survey. IEEE Trans. Ind. Inform. 2014, 10, 2233–2243. [Google Scholar]
  3. Bonomi, F.; Milito, R.; Zhu, J.; Addepalli, S. Fog computing and its role in the internet of things. In Proceedings of the First Edition of the MCC Workshop on Mobile Cloud Computing, Helsinki, Finland, 13–17 August 2012; pp. 13–16. [Google Scholar]
  4. Han, G.; Dong, Y.; Guo, H.; Shu, L.; Wu, D. Cross-layer optimized routing in wireless sensor networks with duty cycle and energy harvesting. Wirel. Commun. Mob. Comput. 2015, 15, 1957–1981. [Google Scholar] [CrossRef]
  5. Rasouli, H.; Kavian, Y.S.; Rashvand, H.F. ADCA: Adaptive duty cycle algorithm for energy efficient IEEE 802.15.4 beacon-enabled wireless sensor networks. IEEE Sens. J. 2014, 14, 3893–3902. [Google Scholar] [CrossRef]
  6. Xie, R.; Liu, A.; Gao, J. A residual energy aware schedule scheme for WSNs employing adjustable awake/sleep duty cycle. Wirel. Pers. Commun. 2016, 90, 1859–1887. [Google Scholar] [CrossRef]
  7. Choudhury, N.; Matam, R.; Mukherjee, M.; Shu, L. Adaptive Duty Cycling in IEEE 802.15. 4 Cluster Tree Networks Using MAC Parameters. In Proceedings of the 18th ACM International Symposium on Mobile Ad Hoc Networking and Computing, Chennai, India, 10–14 July 2017; p. 37. [Google Scholar]
  8. Huang, M.; Liu, A.; Wang, T.; Huang, C. Green data gathering under delay differentiated services constraint for internet of things. Wirel. Commun. Mob. Comput. 2018, 2018. [Google Scholar] [CrossRef]
  9. Azharuddin, M.; Kuila, P.; Jana, P.K. Energy efficient fault tolerant clustering and routing algorithms for wireless sensor networks. Comput. Electr. Eng. 2015, 41, 177–190. [Google Scholar] [CrossRef]
  10. Naranjo, P.G.V.; Shojafar, M.; Mostafaei, H.; Pooranian, Z.; Baccarelli, E. P-SEP: A prolong stable election routing algorithm for energy-limited heterogeneous fog-supported wireless sensor networks. J. Supercomput. 2017, 73, 733–755. [Google Scholar] [CrossRef]
  11. Oller, J.; Demirkol, I.; Casademont, J.; Paradells, J.; Gamm, G.U.; Reindl, L. Has time come to switch from duty-cycled MAC protocols to wake-up radio for wireless sensor networks? IEEE/ACM Trans. Netw. 2016, 24, 674–687. [Google Scholar] [CrossRef]
  12. Mahmood, A.; Shi, K.; Khatoon, S.; Xiao, M. Data mining techniques for wireless sensor networks: A survey. Int. J. Distrib. Sens. Netw. 2013, 9, 406316. [Google Scholar] [CrossRef]
  13. Alsheikh, M.A.; Lin, S.; Niyato, D.; Tan, H.P. Machine learning in wireless sensor networks: Algorithms, strategies, and applications. IEEE Commun. Surv. Tutor. 2014, 16, 1996–2018. [Google Scholar] [CrossRef]
  14. Guestrin, C.; Bodik, P.; Thibaux, R.; Paskin, M.; Madden, S. Distributed regression: an efficient framework for modeling sensor network data. In Proceedings of the 3rd International Symposium on Information Processing in Sensor Networks, Berkeley, CA, USA, 26–27 April 2004; pp. 1–10. [Google Scholar]
  15. Gelb, A. Applied Optimal Estimation; MIT Press: Cambridge, MA, USA, 1974. [Google Scholar]
  16. Abbeel, P.; Coates, A.; Montemerlo, M.; Ng, A.Y.; Thrun, S. Discriminative Training of Kalman Filters. In Proceedings of the Robotics: Science and Systems, Cambridge, MA, USA, 8–11 June 2005; Volume 2, p. 1. [Google Scholar]
  17. Liu, Q.; Wang, Z.; He, X.; Zhou, D. Event-based recursive distributed filtering over wireless sensor networks. IEEE Trans. Autom. Control 2015, 60, 2470–2475. [Google Scholar] [CrossRef]
  18. Leskovec, J.; Rajaraman, A.; Ullman, J.D. Mining of Massive Datasets; Cambridge University Press: New York, NY, USA, 2014. [Google Scholar]
  19. Schütz, N.; Holschneider, M. Detection of trend changes in time series using Bayesian inference. Phys. Rev. E 2011, 84, 021120. [Google Scholar] [CrossRef] [PubMed]
  20. Chung, F.L.; Fu, T.C.; Ng, V.; Luk, R.W. An evolutionary approach to pattern-based time series segmentation. IEEE Trans. Evolut. Comput. 2004, 8, 471–489. [Google Scholar] [CrossRef]
  21. Fuchs, E.; Gruber, T.; Nitschke, J.; Sick, B. Online segmentation of time series based on polynomial least-squares approximations. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 2232–2245. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, K.; Wang, Y.; Sun, Y.; Guo, S.; Wu, J. Green industrial internet of things architecture: An energy-efficient perspective. IEEE Commun. Mag. 2016, 54, 48–54. [Google Scholar] [CrossRef]
  23. Li, H.; Ota, K.; Dong, M. Learning IoT in Edge: Deep Learning for the Internet of Things with Edge Computing. IEEE Netw. 2018, 32, 96–101. [Google Scholar] [CrossRef]
  24. Dautov, R.; Distefano, S.; Bruneo, D.; Longo, F.; Merlino, G.; Puliafito, A. Pushing Intelligence to the Edge with a Stream Processing Architecture. In Proceedings of the 2017 IEEE International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), Devon, UK, 21–23 June 2017; pp. 792–799. [Google Scholar]
  25. Dunkels, A.; Gronvall, B.; Voigt, T. Contiki-a lightweight and flexible operating system for tiny networked sensors. In Proceedings of the 2004 29th Annual IEEE International Conference on Local Computer Networks, Tampa, FL, USA, 16–18 November 2004; pp. 455–462. [Google Scholar]
  26. Crossbow TelosB Mote Plattform, Datasheet. Available online: http://www.willow.co.uk/TelosB_Datasheet.pdf (accessed on 4 March 2018).
  27. Single-Chip 2.4 GHz IEEE 802.15.4 Compliant and ZigBee™ Ready RF Transceiver. Available online: http://www.ti.com/product/CC2420 (accessed on 4 March 2018).
  28. Osterlind, F.; Dunkels, A.; Eriksson, J.; Finne, N.; Voigt, T. Cross-level sensor network simulation with Cooja. In Proceedings of the 2006 31st IEEE Conference on Local Computer Networks, Tampa, FL, USA, 14–16 November 2006; pp. 641–648. [Google Scholar]
Figure 1. Proposed three layer framework for IIoT monitoring systems.
Figure 1. Proposed three layer framework for IIoT monitoring systems.
Sensors 18 01532 g001
Figure 2. The learning process in sensor devices and simulation process in fog node.
Figure 2. The learning process in sensor devices and simulation process in fog node.
Sensors 18 01532 g002
Figure 3. The testbed system consisting of cloud server, Raspberry Pi fog nodes, and TelosB sensor motes.
Figure 3. The testbed system consisting of cloud server, Raspberry Pi fog nodes, and TelosB sensor motes.
Sensors 18 01532 g003
Figure 4. The original sensor data streams. The sampling rate varies for each of the data streams; (a) Stream A 100 ms, (b) Stream B 500 ms, and (c) Stream C 200 ms.
Figure 4. The original sensor data streams. The sampling rate varies for each of the data streams; (a) Stream A 100 ms, (b) Stream B 500 ms, and (c) Stream C 200 ms.
Sensors 18 01532 g004
Figure 5. The model MSE measure comparison with respect to the segment length and the trend threshold.
Figure 5. The model MSE measure comparison with respect to the segment length and the trend threshold.
Sensors 18 01532 g005
Figure 6. Comparisons between normalized sensor data (blue) and the simulated sensor stream (red). The left column (a,c,e), shows the performance of the proposed method for the duration of the experiment (40,000 samples) on different streams. To illustrate the results in more detail, the right column (b,d,f), shows a zoomed in view of a shorter interval (1200 samples on the same streams).
Figure 6. Comparisons between normalized sensor data (blue) and the simulated sensor stream (red). The left column (a,c,e), shows the performance of the proposed method for the duration of the experiment (40,000 samples) on different streams. To illustrate the results in more detail, the right column (b,d,f), shows a zoomed in view of a shorter interval (1200 samples on the same streams).
Sensors 18 01532 g006
Figure 7. The testbed experimental setup: (a) Sensor motes and fog node with attached sensor gateway; (b) REST interface, running on fog node, in a regular web browser; (c) Cloud dashboard with regenerated sensor values by the proposed model.
Figure 7. The testbed experimental setup: (a) Sensor motes and fog node with attached sensor gateway; (b) REST interface, running on fog node, in a regular web browser; (c) Cloud dashboard with regenerated sensor values by the proposed model.
Sensors 18 01532 g007
Table 1. The mean square error of different settings for segment length (m) and trend threshold ( θ ).
Table 1. The mean square error of different settings for segment length (m) and trend threshold ( θ ).
Stream AStream BStream C
θ\m121020301210203012102030
0.0050.080.030.030.030.040.150.050.020.020.020.060.020.010.010.02
0.010.070.040.040.040.380.140.050.020.020.020.050.020.020.020.02
0.020.060.040.030.030.430.100.040.020.020.020.040.020.010.020.02
0.030.080.050.040.050.050.070.040.030.030.030.030.030.030.030.03
0.040.050.050.050.050.050.050.040.040.050.040.040.040.040.040.04
0.050.060.060.060.060.060.050.050.050.060.050.050.050.050.050.05
Table 2. The mean square error of sending the average value with different segment length.
Table 2. The mean square error of sending the average value with different segment length.
m161718192021222324252627
Stream A0.0380.0390.040.040.0430.0430.0450.0460.0460.0490.050.05
Stream B0.0240.0240.0240.0240.0250.0250.0250.0250.0250.0250.0260.026
Stream C0.0170.0180.0180.0190.0190.020.0200.0200.0210.0220.0220.023
Table 3. Delay measurements split into each step.
Table 3. Delay measurements split into each step.
Delay Measurement μ σ
d s e n s o r 140 ms14 ms
d s e r i a l 3.4 ms1.8 ms
d f o g 32 ms34 ms
d c o u d 8.9 ms7.1 ms
d t o t a l 180 ms37 ms
Table 4. Query time of the REST interfaces.
Table 4. Query time of the REST interfaces.
Query Measurement μ σ
Fog5.3 ms9.0 ms
Cloud8.9 ms7.1 ms

Share and Cite

MDPI and ACS Style

Lavassani, M.; Forsström, S.; Jennehag, U.; Zhang, T. Combining Fog Computing with Sensor Mote Machine Learning for Industrial IoT. Sensors 2018, 18, 1532. https://doi.org/10.3390/s18051532

AMA Style

Lavassani M, Forsström S, Jennehag U, Zhang T. Combining Fog Computing with Sensor Mote Machine Learning for Industrial IoT. Sensors. 2018; 18(5):1532. https://doi.org/10.3390/s18051532

Chicago/Turabian Style

Lavassani, Mehrzad, Stefan Forsström, Ulf Jennehag, and Tingting Zhang. 2018. "Combining Fog Computing with Sensor Mote Machine Learning for Industrial IoT" Sensors 18, no. 5: 1532. https://doi.org/10.3390/s18051532

APA Style

Lavassani, M., Forsström, S., Jennehag, U., & Zhang, T. (2018). Combining Fog Computing with Sensor Mote Machine Learning for Industrial IoT. Sensors, 18(5), 1532. https://doi.org/10.3390/s18051532

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop