Next Article in Journal
Embedded Machine Learning System for Muscle Patterns Detection in a Patient with Shoulder Disarticulation
Previous Article in Journal
Energy-Efficient and Highly Reliable Geographic Routing Based on Link Detection and Node Collaborative Scheduling in WSN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Domain Knowledge Transfer for Sustainable Heterogeneous Industrial Internet-of-Things Networks

1
National Engineering Laboratory for Mobile Network Technologies, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
Data61, Commonwealth Science and Industrial Research Organization (CSIRO), Marsfield, NSW 2122, Australia
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(11), 3265; https://doi.org/10.3390/s24113265
Submission received: 7 April 2024 / Revised: 29 April 2024 / Accepted: 18 May 2024 / Published: 21 May 2024

Abstract

:
In this article, a novel cross-domain knowledge transfer method is implemented to optimize the tradeoff between energy consumption and information freshness for all pieces of equipment powered by heterogeneous energy sources within smart factory. Three distinct groups of use cases are considered, each utilizing a different energy source: grid power, green energy source, and mixed energy sources. Differing from mainstream algorithms that require consistency among groups, the proposed method enables knowledge transfer even across varying state and/or action spaces. With the advantage of multiple layers of knowledge extraction, a lightweight knowledge transfer is achieved without the need for neural networks. This facilitates broader applications in self-sustainable wireless networks. Simulation results reveal a notable improvement in the ’warm start’ policy for each equipment, manifesting as a 51.32 % increase in initial reward compared to a random policy approach.

1. Introduction

The sustainability of communication networks is a critical goal for next-generation wireless systems (e.g., 6G and beyond [1]). Network sustainability is defined as an approach that successfully integrates and balances environmental responsibility, economic viability, and social equity. Despite the growing attention and hype surrounding the sustainability of 6G, there is a lack of a rigorous and practical definition to guide its implementation in networks. Sustainability has been mainly linked to green networking to achieve the United Nations’ Sustainable Development Goals (SDGs) [2]. In practice, this is particularly related to energy efficiency of the versatile network elements. In particular, smart factories constitute a significant component in Industrial Internet-of-Things (IIoT) [3] and Industry 4.0 [4], playing a key role in enabling cyber-physical systems to function autonomously. IIoT applications typically requires the automation of a large number of devices in manufacturing with limited hardware capabilities and energy resources, usually with small batteries [5]. Industrial 4.0 [6] encompasses emerging technologies, such as artificial intelligence (AI), edge computing, and digital twin (DT) and so on. In particular, the work in [7] comprehensively investigated the intelligence maintenance in various aspects of maintenance. Specifically, it focused on the human-in-the-loop-based maintenance and its role in enhancing physical resilience in smart manufacturing. This paradigm requires increased flexibility, agility and resilience through the lifespan of the IIoT devices. Consequently, in the realm of IIoT, the smart factories are expected to integrate advanced autonomous capabilities along with enhanced energy-efficient functionality. Nevertheless, the robots, sensors and actuators in the factories are empowered with different sources of energy. Such sources include power grids [8], renewable technologies [9] (e.g., solar), and other energy harvesting techniques [10] (e.g., radio frequency (RF) energy). Subsequently, ensuring the energy efficiency of each individual equipment necessitates adopting a unique mode of operation that is specifically tailored to the varying availability and abundance of their respective energy sources. This can have a direct implication on other critical performance metrics of operation in smart factory. Chief among these metrics is the age of information (AoI) [11] that represents the degree of freshness of the data acquired from the monitored autonomous physical systems [12]. With a focus on both energy efficiency and information freshness, the sustainability of each individual equipment can be significantly enhanced. However, assuring the sustainability of the IIoT as a whole requires looking beyond the individual equipment. In fact, the overall performance and environmental impact of the IIoT will crucially depend not only on the performance of single piece of equipment but also on long-term environmental friendliness of its solution. This encompasses considerations of the system’s overall energy consumption and its ability to sustain prolonged operation without causing harmful impacts on the environment, by considering the associated complexity and energy efficiency of the solution.
The minimization of hybrid energy sources in smart factories has been extensively investigated in various scenarios [13,14]. For instance, the works in [13,14] study the minimization of grid energy consumption in a mixed energy supply scenario. Nonetheless, these works leverages reinforcement learning (RL) solutions [15] that assume a homogeneous model across equipment having heterogeneous energy utilities. In fact, these studies often assume uniformity of state and/or action spaces between heterogenous scenarios, which can barely hold true with the unique operation associated to each equipment [16]. Therefore, in practical real-world scenarios, a robust RL approach is needed to effectively address the heterogeneous nature of the cyber-physical system, while ensuring the sustainability of the solution. Notably, one should consider an RL solution that generalizes across multiple tasks. For instance, the works in [17,18] employ multiple experts to optimize the aggregated performance across different groups. However, the use of multiple agents hinders knowledge sharing among these groups and leads to increased costs as the number of groups grows. The work in [19] considers a federated imitation learning method for cross-domain knowledge sharing framewor. However, the utilization of neural networks slows down the learning process. Furthermore, the application of gradient descent (GD) [20] in such operations incurs additional energy costs as it requires a significant amount of resources to converge. Consequently, to ensure network sustainability, encompassing both the energy efficiency of individual equipment and the computational efficiency of the entire network, a more universally applicable and generalizable solution is essential for heterogeneous Internet of Things (IoT).
The main contributions of this paper is the development of a globally generalizable RL solution, designed to enhance the overall sustainability of cyber-physical systems comprising heterogeneous energy sources. In particular, we tackle the sustaibality issues at both the equipment and system levels by introducing a lightweight, cross-domain knowledge sharing solution. This innovative approach leverages a three-layered knowledge repository structure to facilitate efficient knowledge storage and transfer across the system. Numerical simulations demonstrate that the proposed method consistently outperforms other baseline methods in computational complexity while maintain a comparable performance for smart factories.
The rest of this paper is organized as follows. The system models and problem formulation are provided in Section 2. The proposed cross-domain knowledge sharing framework and the corresponding solutions are presented in Section 3. Simulation results are given in Section 4. Finally, conclusions and future works are drawn in Section 5.

2. System Models

Consider a set N of N smart factory equipment having heterogeneous energy resources in a smart factory. These pieces of equipment are distributed for various manufacturing purposes such as supply chain integration, pre-production setup, production, quality control and inspection, packaging and storage, delivery and so on. Each equipment collects sensory data from its surrounding environment and subsequently executes actions that are tailored to the information gathered. As illustrated in Figure 1, these pieces of equipment are clustered into three distinct groups according to their energy sources. We use x X to index the three groups such that x = 1 , 2 , , X , whereby each group includes a set N x of N x equipment. In particular, three sources of energy supply are considered: (i) grid power (GP), (ii) green sources (GS), and (iii) mixed sources (MS). Specifically, MSs encompasses both the grid and harvested energy resources. In addition, cyber-physical equipment within each group collects data packets from their respective surrounding environments and abstract useful information using their processing capabilities. As illustrated in Figure 2, the abstracted information is subsequently transmitted to a nearby base station (BS).
We consider a time-slotted system where each timeslot has a uniform length denoted as τ . These timeslots are indexed sequentially as t = 1 , 2 , , T . A Rayleigh fading channel is considered for the uplink communication between smart factory equipment and BS. The data transmission rate ϕ ( c y ( t ) ) for each equipment y N at time slot t can be obtained as below:
ϕ ( c y ( t ) ) = B log 2 1 + g p y ( t ) I + B N 0 , y N ,
where c y ( t ) is the number of bits to be processed, ϕ ( c y ( t ) ) is the number of bits to be transmitted after processing, p y ( t ) [ 0 , p y , max ] (in dBm) is the transmitter power used to upload the abstracted information, B is the channel bandwidth, I is the interference from other pieces of equipment in corresponding group, N 0 is noise power spectral density, g is the channel response, which is related to the distance between each equipment y and the BS, i.e., l y . Next, we present the energy models of each group based on their energy sources:
  • GP Source: GP typically refers to power that is supplied through an electrical grid. Hence, GP-powered equipment does not have energy limitations. For instance, the robots and actuators in production line are connected to grid energy supply. The energy consumption e i ( t ) of each equipment i N 1 can be divided into two categories: (a) transmission energy e i T ( t ) = τ p i ( t ) consumed to transmit abstracted information to the BS and (b) computing energy e i C ( t ) = ς κ i ϑ 2 c i , t used to process the collected data packets:
    e i ( t ) = e i T ( t ) + e i C ( t ) = τ p i ( t ) + ς κ i ϑ 2 c i , t ,
    where ς is the energy consumption coefficient depending on the chip of each IIoT equipment, κ i is the number of central processing unit (CPU) cycles required for processing per bit data, assumed to be equal for all pieces of equipment and ϑ is the frequency of the CPU clock of each equipment [21].
  • GS Source: Renewable energy sources, such as wind power, solar power, thermal power and RF are used to enable the establishment of a self-sustainable green network. For example, drones and robots utilized for quality inspection and automated delivery systems are predominantly powered by battery technology. This reduces dependence on conventional grid energy and, consequently, enhances the mobilities while offering greater flexibility and efficiency in operational processes. These energy harvesting methods consistently capture energy from natural environments, converting it into electrical power and storing the collected energy in rechargeable batteries. We define E max as the maximum amount of energy that can be stored in a battery. When the battery reaches its full capacity, any additional harvested energy will be discarded. Consider an ideal rechargeable battery with no energy loss during storage or retrieval processes. At each time slot, the harvested energy e j h ( t ) 0 by equipment j N 2 follows follows a Bernoulli distribution with probability σ [ 0 , 1 ] , such that:
    e j h ( t ) = p solar × ϵ 0 × ϵ 1 × ϵ 2 ,
    where p solar is the density of solar power to the equipment [22]. We consider a typical solar-powered equipment that is equipped with a photovoltaic panel with size ϵ 0 and the energy transfer efficiency ϵ 1 . Considering the heterogeneity in solar power density, a uniformly distributed random variable, ϵ 2 is taken into account. Consequently, the energy level of the battery e j b ( t ) will be given by:
    e j b ( t + 1 ) = min { E max , e j b ( t ) + e j h ( t ) e j T ( t ) e j C ( t ) } ,
    where e j T ( t ) and e j C ( t ) are transmission energy and computing energy for equipment j. Moreover, the following constraint stands:
    0 e j b ( t ) E max
    e j T ( t ) + e j C ( t ) = e j ( t ) e j b ( t ) ,
    where e j ( t ) is the energy consumption at time slot t for equipment in group 2. (5a) implies the battery limitation of equipment j N 2 . (5b) implies that the available energy, which can be used for processing and transmitting energy at the beginning of each time slot, must not exceed the energy level of the battery.
  • MS Source: The third group of cyber-physical equipment is powered by hybrid energy sources comprising both the grid and renewable energy sources. For example, the industrial sensors are strategically deployed to monitor a range of environmental parameters as well as the status of products. This design aims to reduce energy consumption from the grid power while mitigating the randomness and intermittency associated with green energy. Accordingly, for an equipment k N 3 , the consumed energy at time slot t comprises two sources: grid energy e k G ( t ) and battery energy e k B ( t ) . We assume the same energy harvesting model as previously defined, such that e k h ( t ) is updated as in (3). Different from GS, the battery level is updated as:
    e k b ( t + 1 ) = min { E max , e k b ( t ) + e k h ( t ) e k B ( t ) }
    where e k b ( t ) is the battery level at each time slot. Furthermore, the following constraints are held:
    e k T ( t ) + e k C ( t ) e k B ( t ) + e k G ( t ) ,
    e k B ( t ) e k b ( t ) ,
    e k ( t ) = e k G ( t )
    0 e k b ( t ) E max
    where e k T ( t ) and e k C ( t ) are the transmission and computing energy separately. Moreover, (7a) implies that the consumed energy must not exceed the total energy provided by both the battery and the grid. (7b) implies that the permissible battery energy must not exceed the available battery capacity. Since our optimization objective is to minimize grid energy consumption, we set the energy optimization variable e k ( t ) equal to e k G ( t ) in (7c). (7d) indicates the battery limitation of each equipment k N 3 .

AoI Model for Heterogeneous Scenarios

At each time slot, sensing data packets arrive at the equipment with a probability λ y . The size of the data packet, denoted as a y ( t ) , follows a Gaussian distribution. Data packets, once collected, are placed in a waiting queue. The system processes and transmits these packets employing a first-come-first-served (FCFS) approach. Consequently, the queue size b y ( t ) can be updated as follows:
b y ( t + 1 ) = b y ( t ) + a y ( t ) c y ( t ) .
The AoI at time slot t is defined as the timestamp of the most recently processed and successfully received packet at the receiver. The entire process encompasses both data processing time and transmission time. Formally, the update of AoI Δ y ( t ) is as follows:
Δ y ( t + 1 ) = Δ y ( t ) + 1 , if ϖ y ( t ) = 1 min { ( t + 1 ) U ( t ) , Δ max } , otherwise .
where, U ( t ) represents the generation timestamp of the most recent packet and Δ max is the maximum AoI value and ϖ y ( t ) = 1 indicates that the processing of a packet is finished. This limit is imposed to constrain the impact of AoI on performance after a certain level of staleness is reached.
With the aforementioned models, we can now proceed to our optimization objective. For each equipment, we define a cost function w y ( t ) = η 1 Δ y ( t ) + ( 1 η 1 ) e y ( t ) , where η 1 is a tradeoff factor to balance AoI and energy cost and e y ( t ) represents the energy cost of any equipment of three groups. Our objective is to minimize the averaged cost w y ( t ) of all pieces of equipment throughout all the time, which can be written as:
min { p y ( t ) } , { e y G ( t ) } , { e y B ( t ) } x = 1 3 1 N x y N x w y ( t )
s . t . 0 p y ( t ) p y , max
0 Δ y ( t ) Δ max
( 2 ) , y N 1
( 3 ) , ( 4 ) , ( 5 ) , y N 2
( 6 ) , ( 7 ) , y N 3 ,
where (10a) indicates that the transmission power must not surpass the maximum power of each equipment, (10b) indicates the limitation requirements of AoI. (10c) implies the conditions of equipment i N 1 while (10d) and (10e) are constraints for all pieces of equipment in group GS and MS, separately.

3. Cross-Domain for Heterogeneous Scenarios

Problem (10) is NP complete, making it inherently computationally expensive. Furthermore, the unique constraints specified in (10c), (10d) and (10e) add to the complexity of this problem. Additionally, this problem is compounded by the absence of any presupposed knowledge regarding the distribution patterns of data and energy arrivals. To tackle these challenges, we adopt RL, a method that does not require prior knowledge of the underlying distribution patterns. First, three distinct RL models are presented for each group. Then, the overall minimization problem in (10) is considered as a cross-domain knowledge sharing problem.

3.1. Markov Decision Processes (MDPs) Models

Initially, the manufacturing related equipment is partitioned into three distinct groups. As such, the objective of each group is to minimize the averaged cost for all pieces of equipment. Without interfering the overall objective in (10), we model the objectives of each group using MDPs:
  • GP Source: The MDP tuple of the first group can be presented as ( S i , A i , R i ) , where S i is the state space and A i is the action space and R i is the reward function separately. Particularly, S i = { s i , t } = { Δ i , t , b i , t | Δ i , t [ 0 , Δ max ] , b i , t N } . The action space is the set of all possible transmitting powers such that A i = { a i , t } = { p i , t | p i , t [ 0 , p i , max ] } . The reward function can be defined as r i ( s i , t , a i , t ) = w i ( t ) . The parameterized policies can be defined as π θ i ( a i , t | s i , t ) = Pr { a i , t | s i , t , θ i } , where θ i R d 1 , with d 1 = 2 .
  • GS Source: The MDP tuple of the second group can be presented as ( S j , A j , R j ) . Particularly, S j = { s j , t } = { Δ j , t , b j , t , e j , t b | Δ j , t [ 0 , Δ max ] , b j , t N , e j , t b { 0 , E max } } . The action space is the the same as group 1, such that A j = { a j , t } = { p j , t | p j , t [ 0 , p j , max ] } . Similarly, the reward function and the parameterized policies can be defined as r j ( s j , t , a j , t ) = w j ( t ) and π θ j ( a j , t | s j , t ) = Pr { a j , t | s j , t , θ j } , where θ j R d 2 , with d 2 = 3 .
  • MS Source: The MDP tuple of the third group can be presented as ( S k , A k , R k ) . Particularly, S k = { s k , t } = { Δ k , t , b k , t , e k , t b | Δ k , t [ 0 , Δ max ] , b k , t N , e k , t b { 0 , E max } } . The action space is A k = { a k , t } = { p k , t , e k , t B , e k , t G | p j , t [ 0 , p j , max ] , e k , t B [ 0 , E max ] , e k , t G N } . With the similar reward function r k ( s k , t , a k , t ) = w k ( t ) , the parameterized policies can be defined as π θ k ( a k , t | s k , t ) = Pr { a k , t | s k , t , θ j } , where θ k R d 3 , with d 3 = 9 .
So far, we have successfully formulated the cost tradeoff between energy consumption and AoI as a series of MDPs, each corresponding to an individual equipment. For each equipment y, the optimization target can be writen as min θ y J ( θ y ) = E [ 1 / T t = 1 T r y , t ( s y , t , a y , t ) ] . While RL demonstrates the capability to learn and optimize for each equipment individually, the scalability of this approach becomes a concern in large-scale factories due to the potentially large number of learning agents involved. More importantly, this approach can be both time-consuming and energy-intensive. Consequently, there is a pressing need for a more efficient method that can collectively optimize across all pieces of equipment in the three distinct groups. Such a method is crucial not only for the energy efficiency of these equipment but also for the overall sustainability of the cyber-physical system.

3.2. Cross-Domain Knowledge Sharing

To facilitate learning and knowledge sharing across multiple groups, techniques like multi-task learning (MTL) and meta-learning are employed. These methods are adept at managing the simultaneous learning of multiple tasks. However, a significant limitation of these methods is their inherent assumption of model consistency across groups. This assumption poses a challenge when optimizing groups across heterogeneous groups, especially when there is a variation in the state and action spaces of these groups. Consequently, given the substantial resource costs associated with the requirement for numerous learning agents, the need for an efficient cross-domain knowledge transfer method becomes increasingly apparent.
To facilitate knowledge sharing among groups and within each group, a three-layered knowledge base is designed as in Figure 3. A global knowledge base, L R d × m , serves as the shared knowledge among groups. Three group-based knowledge matrices, denoted as G x R d x × d , where x { 1 , 2 , 3 } , are also utilized to store the knowledge specific to each group. These matrices serve as a bridge between the global knowledge base and equipment-specific mapping vectors, represented by s y R m , where y N . As demonstrated by the MDP models of each group, the state and action spaces of each group can vary. In other words, the dimensions of the MDP policies, i.e., θ i , θ j , and θ k , have different dimensions d x , as illustrated in Figure 3. However, due to the existence of varied group knowledge bases G x , the variations of the policy vectors are mapped to the same space. This enables the achievement of a global knowledge base L that can be shared across different domains. As a result, the policy parameters of each equipment can be obtained as:
θ y = G x L s y .
Accordingly, our objective in (10) with the three-layered knowledge system can be represented as the minimization problem:
g ( L , G x ) = x = 1 X 1 N x y N x min s y J ( θ y ) + μ 1 s y 1 + μ 2 G x F 2 + μ 3 L F 2
where L 1 -norm approximates the vector sparsity and L F = ( t r ( L L ) ) 1 / 2 is the Frobenius norm of matrix L . The parameter μ 1 controls the balance between the policy’s fit and the feature’s fit. Also, μ 2 and μ 3 are two regularization parameters, where μ 2 controls the sparsity of s y . The penalty on the Frobenius norm of G and L regularizes the predictor weights to have low L 2 -norm and avoids overfitting.
The above objective can be approximated by performing a second-order Taylor expansion towards J ( θ y ) around the optimla policy α y , which can be obtained using regular RL methods, such as policy gradient: α y = arg min θ y J ( θ y ) . By operating first derivative and second derivative to J ( θ y ) , the above equation can be rewriten as:
g ^ ( L , G x ) = x = 1 X 1 N x y N x min s y α y G x L s y Γ y 2 + μ 1 s y 1 + μ 2 G x F 2 + μ 3 L F 2
where Γ y is the Hessian matrix and α y G x L s y Γ y 2 = ( α y G x L s y ) Γ y ( α y G x L s y ) . The constant term was ignored because it has no effect on the minimization. The linear term was ignored because the α y is the estimated optimal policy.
In further, we can split the above equation by all the equipment. Such that, we only optimize the equipment specific θ y while fix the value of θ y for all other equipment. The improvement of L and G x can be reflected to other equipment. As such, we can obtain the update function of L and G x :
Δ L ( k ) = β 1 x = 1 3 1 | Z x ( k ) | z Z x ( k ) G x Γ z α z s z + G x Γ z G x L s z s z + μ 3 L ,
Δ G x ( k ) = β x 1 | Z x ( k ) | z Z x ( k ) Γ z α z ( L s z ) + Γ z G x ( L s z ) ( L s z ) + μ 2 G x ,
where, β 1 and β x , x { 1 , 2 , 3 } , are the learning rates for L and G x , separately. z Z x is the set of observed equipment for each group. Such that, L ( k + 1 ) = L ( k ) + Δ L ( k ) and G x ( k + 1 ) = G x ( k ) + Δ G x ( k ) , where k means k-th update step. With the updated global knowledge base and group base, s y can be obtained by solving a Lasso:
s y ( k + 1 ) arg min s y ( k ) G x L , s y ( k ) , α y , Γ y ,
where G x L , s y , α y , Γ y = α y G x L s y Γ y 2 + μ 1 s y 1 . Consequently, the full algorithm can be organized as in Algorithm 1: (1) Initialize the L , G x and α y for all equipment. (2) Estimate α y for all equipment. (3) Randomly choose a piece of equipment y and update L and corresponding G x using (12) and (13). (4) Compute s y according to (14). (5) Repeat steps (3) and (4) until the time period comes to an end. It is worth noting that, at each step, we update only the global knowledge base and the corresponding group knowledge base. The performance improvement of equipment in other groups can benefit from the updating of the global knowledge base. In this case, we have N = Z 1 Z 2 Z 3 and θ y = G x L s y .
Algorithm 1 Overview of the Proposed Algorithm
Require:  T 0 , L zeros d , m , G x zeros d x , d , x { 1 , 2 , 3 }
Require:  α y , s y for all pieces of equipment
    while  t T  do
        Randomly choose a piece of equipment
        Identify the group of the chosen equipment as x { 1 , 2 , 3 }
        Obtain interaction history and compute Γ y
        Update L , G x using (12) and (13)
        Update s y for device i using (14)
         t t + 1
    end while

3.3. Computing Complexity

Each update begins with the computing of θ y and Γ y for each individual equipment. We adopt a base-learner, specifically the episodic Natural Actor Critic (eNAC), characterized by a computational complexity of O ( ξ ( d x , n t ) ) for each step. Here, n t represents the number of trajectories obtained for a piece of equipment during the current iteration. The update of L includes multiplication of matrix and vectors, which yields O ( d x 3 + d x d m + m 2 ) for each step. Similarly, the update of G x has a complexity O ( d x d m + m 2 + d x 2 ) . The update of s y requires solving an instance of Lasso, which typically would be O ( d 3 h 2 + m d x 2 + d x m 2 ) . Therefore, the overall complexity of each update for an individual equipment is O ( d x 3 + d x d m + m d x 2 + d x m 2 + ξ ( d x , n t ) ) .

4. Simulation Results

4.1. Simulation Settings

For our simulations, we consider a circular network area with a radius of 500 m and one BS at its center serving three groups of equipment. Each group is distributed in a circle with a radius of 250 m . Within each group, we consider N x = 10 pieces of equipment uniformly distributed. For each group, we have the following simulation parameters:
  • GP: This group of equipment relies solely on the grid power as the energy source. Therefore, there is no limit on the amount of energy could be utilized, i.e., e i ( t ) [ 0 , ] . For this group, we consider the state vector dimension as d 1 = 2 .
  • GS: This group of equipment is equipped with grean energy harvesting capabilities. For solar energy collection, we consider solar panels with the following parameters: p solar = 300 W / m 2 , ϵ 0 = 3.8 cm × 9 cm , ϵ 1 = 50 % , and ϵ 2 [ 0.5 , 1.5 ] . The collected energy is stored in batteries with maximum capacity E max = 10 J . The dimension of the state vector d 2 = 3 is considered in this group.
  • MS: The equipment in this group relies on both the grid energy and the green energy source. For the grid energy source, there’s no limit on the amount of energy could be utilized, i.e., e k B ( t ) [ 0 , ] . For the green energy harvesting, we consider solar energy collection as in group GS. Similarly, the same solar panel parameters are considered here. Moreover, the collected energy is stored in the batteries with the maximum capacity E max = 10 J . We consider a state vector dimension of the d 3 = 9 for this group.
Moreover, for our three layered knowledge model, the values of d and m are obtained through validation experiment. In addition to the above parameters, the parameters shared by all parties pieces are listed below. We consider a bandwidth B = 180 kHz and noise power spectral density N 0 = 174 dBm / Hz . In addition, the loss of the channel is g = 128.1 + 37.6 log 10 ( l y ) , where l y (in km) and the standard deviation of shadow fading is 8 dB . With regard to computing energy consumption, we utilize associated values such as: ς = 10 27 , κ y = 40 and ϑ = 10 9 . We consider the harvested energy arrival probability σ = 0.7 . For each piece of equipment, we assume the average size of the arrived data packets for each group is randomly generated from the range [ 20 , 60 ] . Within each group, the a y , t follows a Gaussian distribution specifically. For all the equipment, we assume p y , max = 0.01 W and Δ max = 30 . The simulations in the article were conducted using a MacBook Pro with an M1 chip. The code was executed on Matlab R2024a, and the MacOS system version used was Ventura 13.2.1.

4.2. Results and Analysis

For comparison purposes, two benchmark algorithms are compared with our proposed algorithm. The first is a Random strategy, which employs a randomly initialized policy and regular Policy Gradient (PG) updating method. To be specific, any PG methods capable of estimating policy gradient can be utilized, such as REINFORCE [23] and Natural Actor Critic (NAC) [24]. In our simulation, we adopt the NAC method as the base learning method. Additionally, we compare our proposed algorithm with the policy gradient efficient lifelong learning (PGELLA) algorithm [25]. PGELLA facilitates learning and knowledge sharing within each individual group, which is different from our cross-domain approach.
In Figure 4, we examine the initial performance improvement of the two algorithms, which we refer to as the warm start policy, compared to the random initial policy. Figure 4 shows the improvement of warm start policies of our proposed method and PGELLA over random initial policies. Both methods surpass the random initial policies. To be specific, the overall averaged results show that our algorithm can achieves 51.32 % warm start policy improvement over random policy while PGELLA achieves 28.91 % in general. As shown in Figure 4, for the devices in group GP, the proposed algorithm can provide a slight better performance compared to PGELLA. This is because the group GP, with lower complexity, requires less knowledge from other domains. Therefore, simpler intra-cluster knowledge sharing and migration models can provide satisfactory performance compared to cross-domain methods. Furthermore, for both the group GS and group MS, the proposed algorithm obviously outperforms PGELLA. Particularly for the MS group, the algorithm proposed can achieve a performance improvement of 90.04 % compared to the random initial policy, while a 23.32 % performance improvement can be achieved by PGELLA. The differing performance of the algorithm proposed and PGELLA across different groups can be attributed to the varying model complexities of these groups. The proposed algorithm enables cross-domain knowledge sharing, leading to a higher degree of exploration for complex models and the ability to migrate learned knowledge to new tasks. The enhanced performance of our method is attributed to its capacity to retain a broader spectrum of knowledge, while PGELLA is limited to providing insights specific to individual groups.
In Figure 5, a comprehensive overview of the learning trajectory is depicted across 1000 iterations for three groups. The implementation of more effective warm start policies can significantly reduce the convergence time for each group. This underscores the impact of initial policy selection on the efficiency of the learning process. In Figure 5a, it is evident that for group GP, both PGELLA and the proposed algorithm demonstrate improved warm start policies compared to the random initial policy. This enhancement effectively improves the performance of the initial policy. Additionally, both methods exhibit similar convergence speeds and final convergence performance. This similarity may be attributed to the simpler Markov models in group GP. These simpler models require less demanding algorithms to achieve satisfactory performance. In other words, PGELLA suffices for simpler models, despite lacking extensive knowledge collection and migration capabilities. Figure 5b illustrates that the enhanced capacity for broader exploration facilitates attaining globally optimal solutions. In Figure 5b, for group GS, our proposed algorithm significantly enhances the performance of the warm start policy and outperforms both PGELLA and random PG algorithm in terms of convergence speed. This improvement stems from the ability of the proposed algorithm to learn cross-domain knowledge and migrate accumulated knowledge from other domains to the current one, potentially achieving global optimality or sub-optimality and breaking out of local optima. It is worth noting that PGELLA also achieves superior results by enabling knowledge sharing within group devices, which helps overcome local optimality limitations. In Figure 5c, it can be observed that both PGELLA and the proposed algorithm achieve notable improvements in warm start policies compared to the random PG algorithm. Particularly, the proposed algorithm exhibits a significant enhancement, consistent with the performance of warm start depicted in Figure 4. Additionally, although PGELLA also contributes to the improvement in convergence rate, the algorithm proposed outperforms the other two algorithms significantly in terms of convergence speed. This is attributed to our proposed algorithm’s capability in cross-domain knowledge transfer, enabling it to achieve greater performance improvements on complex models than on simpler ones. In Figure 5d, the average performance of different algorithms across three groups is presented. Based on the averaged results, it can be observed that overall, both in terms of warm start policies performance and convergence rate, our proposed algorithm outperforms PGELLA. In a nutshell, while the performance in each single group can vary, our algorithm shows better overall performance in respect to warm start policy and convergence speed.
Figure 6 considers the default groups and environmental settings as specified in Section 4.1. In this scenario, mixed energy sources, including grid power energy and green harvested energy, are considered to minimize grid energy consumption. In a nutshell, our proposed method can achieve a better balance between grid energy consumption, battery energy consumption and AoI. By contrast, the other methods, such as PGELLA and Random methods, fail to optimize the overall performance. This proves the sustainability improvement of our proposed method. To be specific, Figure 6 shows the performance attributes for the MS group, encompassing average AoI, queue length, grid energy consumption and battery energy consumption. This group is selected due to its model complexity, allowing for a more comprehensive comparison of the algorithms. Three algorithms are considered: random PG with a random initial policy, PGELLA with intra-group knowledge sharing capability, and the proposed algorithm, the cross-domain knowledge migration algorithm. Figure 6a demonstrates the significant impact of the proposed algorithm on reducing AoI as the number of learning steps increases. In contrast, algorithms with random initial policies or PGELLA exhibit inferior performance in this regard, albeit the latter showing initial superiority compared to random initial policy. Figure 6b reveals a similar trend in average queue length reduction across all the devices in the group MS with the algorithms proposed, indicating enhanced packet processing efficiency. Notably, queue length is closely correlated with AoI. A notable difference in grid energy consumption among the three algorithms is evident in Figure 6c. While both the random initial policy and PGELLA show a decrease in grid energy consumption with increasing learning steps, our proposed algorithms exhibits an increase. However, it’s essential to highlight that this aligns with our goal of minimizing balanced cost. Our method achieves a better balance between AoI and grid energy consumption. Figure 6d indicates relatively consistent performance among the three algorithms in terms of battery energy consumption. The stability in battery energy consumption observed with the proposed algorithms is attributed to the significant performance enhancement with minimal energy consumption in the grid power network. Conversely, the battery energy consumption of the other two algorithms decreases with increasing learning steps, through with limited performance enhancement. Additionally, the PGELLA exhibits lower battery energy consumption than the random initial strategy.
As depicted in Figure 7, the performance of group GS is evaluated in terms of average AoI and average battery energy consumption. From Figure 7a, it’s evident that the performance of all three algorithms improve as the number of learning steps increases. Particularly, the initial AoI of the random PG is lower than the other two algorithms, and this trend persists as the number of learning steps increases. In Figure 7b, the battery energy consumption of all three algorithms decreases with the increase in the number of learning steps. This reduction in energy consumption is accompanied by an increase in average AoI, indicating the optimization of multiple parameters rather than a single objective. Notably, the algorithms proposed demonstrate superior optimization for average energy consumption compared to the other two algorithms. While all algorithms exhibit a significant decrease in average energy consumption with increasing learning steps, the proposed algorithms perform the best in terms of overall performance improvement across multiple parameters, as demonstrated in Figure 5b.

4.3. Influence of the Number of Groups

As depicted in Figure 8, the impact of the number of groups on the proposed algorithm is illustrated. It can be observed that as the number of device groups increases, the proposed algorithm offers improved initial policies. This enhancement is attributed to the increased richness of knowledge in the knowledge base with each type of group, resulting in a more diverse global knowledge base. Furthermore, since the group knowledge base relies on the existence of the global knowledge base, optimizing the global knowledge base further enhances the performance of each group. Specifically, the warm start policy performance of the proposed algorithm improves from 29.73 % for a single group to 51.32 % for three groups compared to a random initial policy. Similarly, as the number of groups increases, the PGELLA algorithm maintains relatively stable performance, with the warm start policy performance ranging from 28.28 % for one group to 29.91 % for three groups. This is because, PGELLA, as an intra-cluster knowledge learning algorithm, does not perform knowledge sharing among groups, and its warm start policy performance variation is due to the differentiated performance of different groups. Nevertheless, our proposed algorithm still achieves a better warm start policy improvement than PGELLA, demonstrating its effectiveness in handling differentiated cluster data. Furthermore, when the number of groups is insufficient, the cross-domain knowledge sharing framework has less knowledge to abstract and share, leading to a degradation in performance compared to scenarios with a larger number of groups. As a result, as the number of groups increases, the advantages of the three-layer knowledge base framework proposed become apparent, significantly outperforming the performance of PGELLA. This highlights the advantages of the three-layer knowledge base framework in handling cross-domain knowledge transfer. Therefore, the proposed algorithm exhibits more potential application scenarios and advantages when the number of groups is high.
As depicted in Table 1, the table provides a comparison of the running time and the running time difference between two algorithms: PGELLA and our proposed algorithm, for varying numbers of groups. From the table, it’s evident that as the number of groups increases, the running time of both algorithms also increases approximately linearly. With PGELLA, each additional group necessitates a complete repetition of the algorithm’s process. On the other hand, our proposed algorithm requires visiting a higher number of devices with each additional group, resulting in a longer time to visit all devices compared to PGELLA. Hence, with a single group, the runtime of our proposed algorithm ( 3.9225 s) is less than that of PGELLA. Yet, as the number of groups rises, PGELLA’s runtime progressively diminishes compared to our proposed algorithm. In particular, when there are three groups, PGELLA’s runtime surpasses that of the proposed algorithm by 0.5931 s. Combining the findings from Figure 8, we observe that our proposed algorithm achieves approximately a 15 % performance enhancement, with a mere 4.94 % increase in runtime. This suggests a perfect balance between runtime efficiency and performance improvement.

5. Conclusions

The article introduces a lightweight cross-domain knowledge sharing model leveraging diverse energy supply methods. It employs a three-layered knowledge base, incorporating global, group-specific, and individual policy vectors. By integrating grid, harvested, and mixed energy sources, significant improvements in warm start policy performance are demonstrated compared to random initial policies. Moreover, the collaborative nature of the global knowledge base contributes to enhanced sustainability, surpassing that of two-layered models. Considering the significant energy savings and AoI optimization achieved, our approach can facilitate the sustainability of IIoT and Industrial 4.0 initiatives. To advance in the field of cross-domain knowledge sharing, several potential research directions can be considered. These include investigating the impact of mobility, addressing privacy and security concerns, and exploring the integration of edge computing.

Author Contributions

Conceptualization, Z.G. and Q.C.; methodology, Z.G. and W.N.; software, Q.C.; validation, Z.G., Q.C. and W.N.; investigation, W.N.; resources, Q.C.; data curation, Z.G.; writing—original draft preparation, Z.G.; writing—review and editing, W.N.; visualization, Z.G.; supervision, Q.C.; project administration, W.N.; funding acquisition, Q.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Regional Innovation and Development of the National Natural Science Foundation of China (No. U21A20449).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IIoTIndustrial Internet-of-Things
IoTInternet of Things
AoIAge of Information
SDGsSustainable Development Goals
RLReinforcement Learning
GDGradient Descent
GPGraid Power
GSGreen Source
MSMixed Sources
BSBase Station
CPUCentral Processing Unit
FCFSFirsy-Come-Firsy-Served
MDPMarkov Decision Processes
MTLMulti-Task Learning
eNACepisodic Natural Actor Critic
PGELLAPolicy Gradient Efficient Lifelong Learning Algorithm
PGPolicy Gradient
CDCross Domain

References

  1. Saad, W.; Bennis, M.; Chen, M. A vision of 6G wireless systems: Applications, trends, technologies, and open research problems. IEEE Netw. 2019, 34, 134–142. [Google Scholar] [CrossRef]
  2. The Sustainable Development Goals Report. Available online: https://unstats.un.org/sdgs/report/2019/The-Sustainable-Development-Goals-Report-2019.pdf (accessed on 6 April 2024).
  3. Li, X.; Cui, Q.; Feng, D.; Gong, Z.; Tao, X. Deep Reinforcement Learning-Based Solution for Minimizing the Alterable Urgency of Information in UAV-Enabled IIoT System. In Proceedings of the GLOBECOM 2023—2023 IEEE Global Communications Conference, Kuala Lumpur, Malaysia, 4–8 December 2023; IEEE: New York, NY, USA, 2023; pp. 437–442. [Google Scholar]
  4. Malik, P.K.; Sharma, R.; Singh, R.; Gehlot, A.; Satapathy, S.C.; Alnumay, W.S.; Pelusi, D.; Ghosh, U.; Nayak, J. Industrial Internet of Things and its applications in industry 4.0: State of the art. Comput. Commun. 2021, 166, 125–139. [Google Scholar] [CrossRef]
  5. Moloudian, G.; Hosseinifard, M.; Kumar, S.; Simorangkir, R.B.; Buckley, J.L.; Song, C.; Fantoni, G.; O’Flynn, B. RF Energy Harvesting Techniques for Battery-less Wireless Sensing, Industry 4.0 and Internet of Things: A Review. IEEE Sens. J. 2024, 24, 5732–5745. [Google Scholar] [CrossRef]
  6. Folgado, F.J.; Calderón, D.; González, I.; Calderón, A.J. Review of Industry 4.0 from the Perspective of Automation and Supervision Systems: Definitions, Architectures and Recent Trends. Electronics 2024, 13, 782. [Google Scholar] [CrossRef]
  7. Cortés-Leal, A.; Cárdenas, C.; Del-Valle-Soto, C. Maintenance 5.0: Towards a Worker-in-the-Loop Framework for Resilient Smart Manufacturing. Appl. Sci. 2022, 12, 11330. [Google Scholar] [CrossRef]
  8. Dileep, G. A survey on smart grid technologies and applications. Renew. Energy 2020, 146, 2589–2625. [Google Scholar] [CrossRef]
  9. Zhang, H.; Lu, Y.; Han, W.; Zhu, J.; Zhang, Y.; Huang, W. Solar energy conversion and utilization: Towards the emerging photo-electrochemical devices based on perovskite photovoltaics. Chem. Eng. J. 2020, 393, 124766. [Google Scholar] [CrossRef]
  10. Ibrahim, H.H.; Singh, M.J.; Al-Bawri, S.S.; Ibrahim, S.K.; Islam, M.T.; Alzamil, A.; Islam, M.S. Radio frequency energy harvesting technologies: A comprehensive review on designing, methodologies, and potential applications. Sensors 2022, 22, 4144. [Google Scholar] [CrossRef] [PubMed]
  11. Kaul, S.; Yates, R.; Gruteser, M. Real-time status: How often should one update? In Proceedings of the 2012 Proceedings IEEE INFOCOM, Orlando, FL, USA, 25–30 March 2012; IEEE: New York, NY, USA, 2012; pp. 2731–2735. [Google Scholar]
  12. Dang, Q.; Cui, Q.; Gong, Z.; Zhang, X.; Huang, X.; Tao, X. AoI oriented UAV trajectory planning in wireless powered IoT networks. In Proceedings of the 2022 IEEE Wireless Communications and Networking Conference (WCNC), Austin, TX, USA, 10–13 April 2022; IEEE: New York, NY, USA, 2022; pp. 884–889. [Google Scholar]
  13. Hamdi, R.; Baccour, E.; Erbad, A.; Qaraqe, M.; Hamdi, M. LoRa-RL: Deep reinforcement learning for resource management in hybrid energy LoRa wireless networks. IEEE Internet Things J. 2021, 9, 6458–6476. [Google Scholar] [CrossRef]
  14. Ye, J.; Gharavi, H. Deep reinforcement learning-assisted energy harvesting wireless networks. IEEE Trans. Green Commun. Netw. 2020, 5, 990–1002. [Google Scholar] [CrossRef] [PubMed]
  15. Gong, Z.; Cui, Q.; Chaccour, C.; Zhou, B.; Chen, M.; Saad, W. Lifelong learning for minimizing age of information in Internet of Things networks. In Proceedings of the ICC 2021-IEEE International Conference on Communications, Montreal, QC, Canada, 14–23 June 2021; IEEE: New York, NY, USA, 2021; pp. 1–6. [Google Scholar]
  16. Gong, Z.; Hashash, O.; Wang, Y.; Cui, Q.; Ni, W.; Saad, W.; Sakaguchi, K. UAV-aided lifelong learning for AoI and energy optimization in non-stationary IoT networks. arXiv 2023, arXiv:2312.00334. [Google Scholar]
  17. Kolter, J.Z.; Maloof, M.A. Dynamic weighted majority: An ensemble method for drifting concepts. J. Mach. Learn. Res. 2007, 8, 2755–2790. [Google Scholar]
  18. Wu, Q.; Iyer, N.; Wang, H. Learning contextual bandits in a non-stationary environment. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 495–504. [Google Scholar]
  19. Yu, A.; Yang, Q.; Dou, L.; Cheriet, M. Federated imitation learning: A cross-domain knowledge sharing framework for traffic scheduling in 6G ubiquitous IoT. IEEE Netw. 2021, 35, 136–142. [Google Scholar] [CrossRef]
  20. Cui, Q.; Gong, Z.; Ni, W.; Hou, Y.; Chen, X.; Tao, X.; Zhang, P. Stochastic online learning for mobile edge computing: Learning from changes. IEEE Commun. Mag. 2019, 57, 63–69. [Google Scholar] [CrossRef]
  21. Pan, Y.; Pan, C.; Yang, Z.; Chen, M. Resource allocation for D2D communications underlaying a NOMA-based cellular network. IEEE Wirel. Commun. Lett. 2017, 7, 130–133. [Google Scholar] [CrossRef]
  22. Lu, X.; Wang, P.; Niyato, D.; Kim, D.I.; Han, Z. Wireless networks with RF energy harvesting: A contemporary survey. IEEE Commun. Surv. Tutorials 2014, 17, 757–789. [Google Scholar] [CrossRef]
  23. Williams, R.J. Simple Statistical Gradient-following Algorithms for Connectionist Reinforcement Learning. Mach. Learn. 1992, 8, 229–256. [Google Scholar] [CrossRef]
  24. Peters, J.; Schaal, S. Natural Actor-critic. Neurocomputing 2008, 71, 1180–1190. [Google Scholar] [CrossRef]
  25. Ammar, H.B.; Eaton, E.; Ruvolo, P.; Taylor, M. Online multi-task learning for policy gradient methods. In Proceedings of the International Conference on Machine Learning. PMLR, Beijing, China, 21–26 June 2014; pp. 1206–1214. [Google Scholar]
Figure 1. System model of three groups of smart factory equipment with diverse energy sources. Group 1 supplied by grid power wire, group 2 supplied by battery energy and group 3 supplied by grid power and green power source.
Figure 1. System model of three groups of smart factory equipment with diverse energy sources. Group 1 supplied by grid power wire, group 2 supplied by battery energy and group 3 supplied by grid power and green power source.
Sensors 24 03265 g001
Figure 2. Illustrative figure of the system model representing an smart factory.
Figure 2. Illustrative figure of the system model representing an smart factory.
Sensors 24 03265 g002
Figure 3. Illustration of three layers knowledge framework with varied state and/or action spaces.
Figure 3. Illustration of three layers knowledge framework with varied state and/or action spaces.
Sensors 24 03265 g003
Figure 4. The warm start policy improvement for all the groups.
Figure 4. The warm start policy improvement for all the groups.
Sensors 24 03265 g004
Figure 5. The learning methods comparison over 1000 iterations: (a) Group GP. (b) Group GS. (c) Group MS. (d) All groups.
Figure 5. The learning methods comparison over 1000 iterations: (a) Group GP. (b) Group GS. (c) Group MS. (d) All groups.
Sensors 24 03265 g005
Figure 6. AoI and energy related performance comparison for group MS: (a) Average AoI. (b) Average queue length. (c) Average grid energy consumption. (d) Average battery energy consumption.
Figure 6. AoI and energy related performance comparison for group MS: (a) Average AoI. (b) Average queue length. (c) Average grid energy consumption. (d) Average battery energy consumption.
Sensors 24 03265 g006
Figure 7. AoI and energy consumption comparison for group GS: (a) Average AoI. (b) Average total battery energy consumption.
Figure 7. AoI and energy consumption comparison for group GS: (a) Average AoI. (b) Average total battery energy consumption.
Sensors 24 03265 g007
Figure 8. Warm start policy improvement when the number of groups increases.
Figure 8. Warm start policy improvement when the number of groups increases.
Sensors 24 03265 g008
Table 1. The Running Time Comparison.
Table 1. The Running Time Comparison.
Running Time (Seconds)One GroupTwo GroupsThree Groups
CD3.92259.862012.5984
PG4.00149.457512.0053
Gap−0.07930.40440.5931
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gong, Z.; Cui, Q.; Ni, W. Cross-Domain Knowledge Transfer for Sustainable Heterogeneous Industrial Internet-of-Things Networks. Sensors 2024, 24, 3265. https://doi.org/10.3390/s24113265

AMA Style

Gong Z, Cui Q, Ni W. Cross-Domain Knowledge Transfer for Sustainable Heterogeneous Industrial Internet-of-Things Networks. Sensors. 2024; 24(11):3265. https://doi.org/10.3390/s24113265

Chicago/Turabian Style

Gong, Zhenzhen, Qimei Cui, and Wei Ni. 2024. "Cross-Domain Knowledge Transfer for Sustainable Heterogeneous Industrial Internet-of-Things Networks" Sensors 24, no. 11: 3265. https://doi.org/10.3390/s24113265

APA Style

Gong, Z., Cui, Q., & Ni, W. (2024). Cross-Domain Knowledge Transfer for Sustainable Heterogeneous Industrial Internet-of-Things Networks. Sensors, 24(11), 3265. https://doi.org/10.3390/s24113265

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop