1. Introduction
In order to protect user’s privacy and meet law regulations, Google proposes Federated Learning in 2016. Currently, Federated Learning is widely used in health care [
1,
2], smart city [
3], smart phone [
4] and other fields. Although Federated Learning improves the privacy of local training data by exchanging model updates between clients and server, such as local gradients or updated parameters, and each client’s raw data is stored locally [
4,
5,
6,
7,
8], some studies have shown that sharing even the local gradients update still have privacy risks. Adversary can make use of the updated gradients and weights to reconstruct the local training data [
9,
10,
11]. Therefore, it is a worthy issue to study the influence of gradient on user’s private training image data disclosure.
Exchanging gradient is a widely used method in modern multinode machine learning system (e.g., distributed training, Federated Learning). It is meaningful to research the gradient’s safety. According to the research history of gradient leakage, we review relevant work, which is listed in
Figure 1.
In 2017, Phong et al. [
12] first show that user’s private training data can be recovered from gradient in theoretical. They make use of gradient of loss function about model weight parameters
and biases
. From the Formula
, user’s private training data
is completely leaked. This suggests a single-sample recovery is possible. Because of the use of biases, this type of attacks are called Bias Attacks.
Based on the conclude in [
12] that gradient can compromise the privacy of user’s private training, Wang et al. [
13] propose mGAN-AI attack against the federated learning for reconstructing private data of a specific victim. They first utilize an optimization approach to minimize the distance between gradients. This approach is adopted as a submodule in their Multitask GAN model. Their work first constructs a representation of the input image, then improves with a GAN. The optimization approach is the origin of the later Optimization-Based Attacks.
Salem et al. [
14] aim at inferring different information of the updating set
in the output of a black-box Machine Learning model.
denotes the posterior difference of both outputs. They propose a single-sample reconstruction attack
to reconstruct the updating set itself. Extensive experiments show that full reconstruction in challenging conditions. However, the threat model in most distributed learning paradigms is not common.
Melis et al. [
15] demonstrate with both CNNs and RNNs that periodical gradient updates during training can leak features as well as class memberships. Further, they also show that possible defences, such as selective gradient sharing, reducing dimensionality, and dropout are proved to be ineffective or have a negative impact on the quality of the collaboratively trained model.
Chai et al. [
16] show server can deduce user’s rating data. However, there need knowing gradients of a user uploaded in two continuous steps. In the end, they enhance the distributed matrix factorization framework with homomorphic encryption.
Subsequently [
13], Zhu et al. [
9] officially propose Optimization-Based Attacks. They first randomly initialize a dummy input
and label input
, then, feed these dummy data into models and get dummy gradient, at last, there need optimize the dummy gradient close as to original, which also makes the dummy data close to the real training data. They use L-BFGS [
17] to perform the optimization. Experiments show the recovery from gradient is pixelwise accurate for images and tokenwise matching for texts, even if missing label information. Zhao et al. [
10] extend the algorithm of [
9] and show that label information can be computed analytically from the gradients of the last fully connected layer. They reconstruct the one-hot label of multiclass classification in terms of a single input.
Jonas Geiping et al. [
11] adopt Optimization-Based Attacks, and follow the conclude of Bo Zhao et al. [
10] who find the label information can be reconstructed analytically for classification tasks. Thus, they consider label information to be known. Experiments show that it is possible to faithfully reconstruct a image from parameter gradients. Further, They discuss several images does not protect the user’s privacy in federated learning applications.
Wei et al. [
18] continue to work on Optimization-Based Attacks. They show the algorithm [
9] is sensitive to initialization, and the same class image is an optimal initialization.
Fan et al. [
19] analyze the Bias Attacks as a system of linear equations, and propose a method of perturbing the gradients to protect personal privacy. They propose a Secret Polarization Network (SPN). SPN consists of a public and a private network based on a backbone network. Fully connected polarization layers are kept private with its parameters not shared during the distributed learning process. Further, they first perform a rank analysis about network’s security.
Milad Nasr et al. [
20] investigate the impact of gradients of different layers on attack accuracy, and show that the gradients of the later layers leak more membership information. It belongs to white-box inference attack. In [
20], they argue that parameter aggregation in federated learning scenarios will have a negative impact on the accuracy of membership inference attacks. Furthermore, they propose to isolate target participant and improve the accuracy of attacker.
Pan et al. [
21] show theory-oriented deep leakage from gradients via linear equation solver. The concept of Linear Equation Solver Attacks is first proposed. They analysize neural networks with ReLU activation function and use Optimization-Based Attacks technique to form a sparse linear equation system about gradients. In the process of solving the gradient equation, they made use of the tricky properties of ReLU. They focus their attacks on fully connected neural networks.
Zhu et al. [
22] break the gradient attack down to a recursive process of solving systems of linear equations and propose recursive procedure to recover data from gradients in deep neural networks, which is named Recursive Gradient Attack. It belongs to Linear Equation Solver Attacks. They achieve an analytic gradient attack from fully connected layers to CNNs for the first time. Furthermore, they also propose a novel rank analysis to estimate the feasibility of performing gradient based privacy attacks.
Yin et al. [
23] extend single image label restoration [
10], and formulate label restoration algorithm for batch size K. They assume nonrepeating labels in batch. Furthermore, they adopt Optimization-Based Attacks technique to realize a full recovery of detailed individual images from batch averaged gradients in deep networks such as ResNet-50.
Scheliga et al. [
24] introduce a PRivacy EnhanCing mODulE (PRECODE) that can be used as generic extension for arbitrary model architectures. It can be used for privacy leakage defense mechanisms.
In summary, the research history of gradient mainly focus on privacy attacks and privacy defense strategies. Existing works about privacy attacks can be roughly categorized into three groups: (i) Bias Attacks, bias reconstruction attacks mean to use gradients about bias term and weights to realize attack [
12,
19]. This attack can be solved by removing the bias term. (ii) Optimization-Based Attacks, the core idea of Optimization-Based gradient attacks are to minimize the distance between gradients. It first appears in [
13], subsequently [
9,
10,
11,
18] refine the method. (iii) Linear Equation Solver Attacks, the main idea is to form gradient equation or weight equation, further, relizes reconstruct attacks [
19,
21,
22]. In [
19], they use a bias term and weights to form a gradient equation to realize attacks. It only studys a single neuron. While in [
21], they rely on Optimization-Based gradient technique to form a gradient equation. In [
22], they break the gradient attacks from fully layer to convolution layer, furthermore, make use of gradient constraints and weight constraints to form a matrix, solving the matrix can derive training image data.
Usually, there is a paradox between privacy reconstruction and privacy protection. Privacy-Preserving Deep Learning aims to collaboratively train and share a deep neural network model among multiple participants, without exposing their private training data. There are three common methods as privacy defense strategies: Differential-Privacy, Homomorphic Encryption, and Secure Multiparty Computation. In [
25], differential privacy can be used for solving the privacy problem of crowdsourced. In [
26], they focus on privacy leakage issue of publishing well-trained deep neural network models, differential privacy can be used for solving the problem. In [
16], they enhance the distributed matrix factorization framework with homomorphic encryption. In [
27], They provide formal data privacy guarantees using both differential privacy and secure multiparty computation frameworks. In order to proptct privacy, there are also some researchers to study Machine Learning Approaches for Malware Detection [
28,
29,
30]. In this paper, our work mainly focus on gradient inversion attack algorithms.
In this article, our work mainly studies the most representative gradient inversion algorithms. DLG algorithm [
9] belongs to Optimization-Based Attacks. It can recover private training data pixelwise accurate for images without label. However, this algorithm only can reconstruct image at the fully connected layer. Inverting Gradients [
11] also belongs to Optimization-Based Attacks. It can faithfully reconstruct images at high resolution to fully connected layer. However, this algorithm can work under the premise of label information is known. R-GAP [
22] belongs to Linear Equation Solver Attacks. It works as well as or even better than Optimization-Based Attacks. This algorithm extends attack from the fully connected layers to CNNs, However, this algorithm works under the premise of label information is known. Therefore, according to the characteristics of these algorithms, we propose one privacy attack system, i.e., Single-Sample Reconstruction Attack System (SSRAS). This system combines the advantages of most representative attack algorithms, which can realize image reconstruction regardless of whether the label can be determined. It can extends gradient inversion attack from a fully connected layer with bias terms to attack a fully connected layer and convolutional neural network with or without bias terms. Further, we propose Improved R-GAP Alogrithm, which can utlize DLG algorithm to derive ground truth. In short, compared with the most representative attack algorithms, this reconstruction attack system can recover user’s private training image with high fidelity and attack success rate. Experimental results also show the superiority of the attack system over some other state-of-the-art attack algorithm.
Our main contributions are as follows:
We propose one privacy attack system, i.e., Single-Sample Reconstruction Attack System (SSRAS). This system combines the advantages of most representative attack algorithms, which can realize image reconstruction regardless of whether the label can be determined. It can extends gradient inversion attack from a fully connected layer with bias terms to attack a fully connected layer and convolutional neural network with or without bias terms.
R-GAP works only if the label information is known. In this section, we propose Improved R-GAP Alogrithm, which can utlize DLG algorithm to derive ground truth. Further, it can extends the attack from the fully connected layer to the convolutional layer.
We introduce Rank Analysis Index (RA-I) to measure the possible of whether the user’s raw image data can be reconstructed. This rank analysis derive virtual constraints
from weights. This Rank Analysis Index(RA-I) is superior to [
19].
In order to guide attack from gradient towards natural images, we adopt to the loss function to steer reconstructed image away from unrealistic image. Improved R-GAP Alogrithm can make use of the difference of the reconstructed image, and adopt smoothed version.
Simulation experiments and analysis of the optimization scheme verify that gradients encode a large amount of information, and this reconstruction attack system can recover user’s private training image with high fidelity and attack success rate, regardless of whether the label can be determined.
The rest of the paper is organized as follows:
Section 2 describe the fundamental milestone framework of gradient leakage.
Section 3 propose one privacy attack system, i.e., Single-Sample Reconstruction Attack System (SSRAS).
Section 4 propose Improved R-GAP Alogrithm.
Section 5 the experimental results are shown.
Section 6 conclude the paper and give the further work.
2. Related Work
In this section, we review existing works about privacy attacks, and explain three types of attacks, for example Bias Attacks, Optimization-Based Attacks, and Linear Equation Solver Attacks.
2.1. Bias Attacks
In [
12], Phong et al. first show that recover user’s private training data from gradient is possible. It is called Bias attacks.
is the input data,
,
y is a corresponding truth label,
is the weight parameter to be learned,
,
b is the bias,
f is an activation function. The loss function is defined as the distance between the predicted value and the truth value. The predicted value is
, the truth value is
y. The loss function
Gradient on the training sample is exactly the partial derivative of the loss function w.r.t.the model weight parameter and the bias.
According to Formulas (2) and (3), we can have
Therefore, from Formula (4), we can draw a conclude is completely leaked if the gradients are shared to server. In theory, this suggests a single-sample recovery is possible. However, we can simply disable this attack by removing the bias term. Besides, because of dimension mismatch, this way can not work on convolutional neural networks.
Fan et al. [
19] aim at sloving the Bias Attacks. They propose a Secret Polarization Network(SPN), which is a method of perturbing gradients. The architecture of the SPN contains backbone network, and fully connected polarization layers. Thereof, fully connected polarization layers are devided into public and a private network. The parameters of private network are not shared in distributed learning. The loss about gradients
can be shown as follows,
and
denote hyperparameters,
,
denotes cross-entropy loss,
denotes polarization loss. The function of
is to introduce interference,
controls the protection levels of training data. Fan et al. [
19] also believe convolutional networks and fully connected networks are equivalent. However, they do not take into account that gradients are aggregated in a convolutional network. Further, they also first propose a rank analysis to estimate the security of network.
2.2. Optimization-Based Attacks
Wang et al. [
13] propose mGAN-AI attack against the federated learning for reconstructing private data of a specific victim. They first utilize an optimization approach to minimize the distance between gradients. This approach is adopted as a submodule in their Multitask GAN model. Their work first constructs a representation of the input image, then improved with a GAN.
Subsequently [
13], Zhu et al. [
9] propose that even the absence of label information, the recovery from gradient is pixelwise accurate for image. Zhao et al. [
10] extend the algorithm of [
9] and show that label information can be computed analytically from the gradients of the last fully connected layer. They reconstruct the one-hot label of multiclass classification in terms of a single input.
The central recovery mechanism discussed in [
9,
10,
23] is to recover the data from gradients. Zhu et al. [
9] first randomly initialize a dummy input
and label input
, then feed these dummy data into models and get dummy gradient, at last, they adopt L-BFGS [
17] to perform the optimization. Zhao et al. [
10] make use of the shared gradients of fully connected layer to extract the ground-truth label, then they can extract the data more effectively based on correct label. Yin et al. [
23] extend single image label restoration [
10], and formulate label restoration algorithm for batch size K. They assume nonrepeating labels in batch. Furthermore, they adopt Optimization-Based Attacks technique to realize a full recovery of detailed individual images from batch averaged gradients in deep networks, such as ResNet-50.
In [
9,
10], they all optimize the dummy gradients close to original gradients, which also make the dummy data close to the real training data. Their optimization adopts euclidean.
The cost function is minimized to recover the original input image , from transmitted gradient . Note that, this optimization requires 2nd order derivatives.
Zhao et al. [
10] firstly propose the ground truth label information can be derived from gradients of the last fully connected layer. They adopt a classification scenario, the loss function is defined as follows,
x denotes input data, c denotes corresponding ground-truth label,
denotes the predicting score of the
class. The loss about output
partial can derived gradients according to Formulas (9) and (10),
The gradient vector
is the weight
connected to the
logit in the output layer. Usually, the gradients of model weights are shared. Combining with Formula (10), there are
where the network has L layers,
is the output of
L layer,
is the bias parameter of
L layer,
.
Combining Formulas (10) and (11), the ground-truth label
c can be predicted as follows,
The conclude of ground-truth c can be derived according to the sign of gradient. The signs of and are the same. The negative gradient can be the index of the ground-truth label. Note that, all the conclude need the assumption that there is non-negative activation function.
Jonas et al. [
11] propose that it is no matter to architecture of training deep networks or trained deep networks, any input to a fully connected layer can be reconstructed. They consider that label information is known, this assumption base on the conclusion of Zhao et al. [
10]. They adopt cosine similarity loss function, and add
to control the image prior to the overall problem. Furthermore, they replace L-BFGS with Adam for optimization about networks, such as ReLU or LeakReLU. The specific form is defined as Formula (13),
In [
11], they also discuss that federated average algorithm can average gradients over several iterations or several images in a batch, but it does not protect the user’s privacy in federated learning applications. They use a ConvNet architecture, which is roughly similar to AlexNet [
31].
In [
18,
23], they propose to use
loss function to recover the original input image. Wei et al. [
18] show that the algorithm of Zhu et al. [
9] is sensitive to initialization, the optimal initialization is the same class image. The specific form is defined as Formula (14),
is the gradient of local training on private training data, is attack seed, the gradient of attack seed is , is reconstructed training data, is regularizer ratio.
Pan et al. [
21] show theory-oriented privacy analysis in neural networks with ReLU for data reconstruction attacks. They rely on Optimization-Based gradient technique to form gradient equations, The specific form is shown as follows,
are variables in this equation. In solving the gradient equation, they take advantage of ReLU’s properties.
2.3. Linear Equation Solver Attacks
In [
19], they use a bias term and weights to form a gradient equation to realize attack. While, in [
21], they rely on Optimization-Based gradient technique to form a gradient equation. In [
22], they break the gradient attacks from fully layer to convolution layer, furthermore, make use of gradient constraints and weight constraints to form a matrix, solving the matrix can derive training image data. They also propose to estimate the feasibility of performing gradient attacks by rank analysis. The most representative attack is R-GAP [
22]. The concrete form of gradient constraints can be described as Formula (16),
denotes the input in the ith layer, denotes coefficient matrix containing all gradient constraints in the ith layer.
The concrete form of weight constraints can be described as Formula (17). It needs assumption that they know the input of the subsequent layer.
represents convolutional kernel, is weight constraints.
There is research about property inference attack in [
32]. Researchers [
32] explore to infer properties of training data using the model parameters. It belongs to retrieving input attributes from local updates, and is a shallow leak. Property inference can be derived directly in the reconstruction phase or by classifying the reconstructed data.
Some researchers propose model inversion attack. They show that the attacker attempts to obtain information of the training dataset from the trained model [
13,
33,
34]. They utilize data representation to infer attribute values of data samples. With the development of Generative Adversarial Network, GAN-based reconstruction attack appear in [
13]. The participant utilizes GAN structure to construct sensitive information about the victim. They can infer general image composition or dominating colors. Model inversion attack relies on a given layer and only reconstructs similar image. Fredrikson et al. [
33] demonstrate that model inversion attack can recover similar image from facial recognition system. Pan et al. [
34] utilize the intermediate data representation to infer sensitive attribute values of data samples. Model inversion attack generally is challenging for deeper neural network architectures if no additional information is provided.
Model extraction attack refers to an attacker trying to steal the parameters and hyperparameters of the model, Further, it can break model confidentiality or infer user data sets and model characteristics. In [
35], they use model interpretation to reconstruct significant parts of training set.
3. SSRAS: Single-Sample Reconstruction Attack System
According to the research of related work, we have the pros and cons of the main gradients inversion algorithms. In this section, we propose one privacy attack system, i.e., Single-Sample Reconstruction Attack System (SSRAS). We first show threat model and attack objective. Then, we explain the key components of the SSRAS.
3.1. Threat Model and Attack Objective
Federated Learning improves the privacy of local training data by exchanging model updates, such as local gradients or updated parameters. Howerever, some attack algorithms have shown that the adversary can utilize gradients to obtain user’s private training image data. There are pros and cons of existing privacy attack algorithms. The most representative attack algorithms are listed in
Table 1. The first attack Algorithm 1 (DLG) [
9] and Algorithm 2 (Inverting Gradients) [
11] belong to Optimization-Based Attacks from Gradients. Algorithm 3 (R-GAP) [
22] belongs to Linear Equation Solver Attacks. DLG algorithm [
9] belongs to Optimization-Based Attacks. It can recover private training data pixelwise accurate for images without label. However, this algorithm only can reconstruct image to fully connected layer. Inverting Gradients [
11] also belongs to Optimization-Based Attacks. It can faithfully reconstruct images at high resolution to fully connected layer. However, this algorithm can work under the premise of label information is known. R-GAP [
22] belongs to Linear Equation Solver Attacks.
Federated Learning trains a shared global model by a server and client working together. This security threat may come from: honest-but-curious server or malicious client. During training, server and client konw architecture of neural network, weights, gradients and other relevant information. Weight and gradient are transmitted in plaintext [
4], instead of using privacy-preserving deep learning techniques [
25,
26,
27,
36], such as Differential-Privacy, Homomorphic Encryption, and Secure Multiparty Computation. The main purpose is to investigate whether user’s privacy security can be ensured by transmitting only gradients and weights in federated learning. Furthermore, it is necessary to study how to use gradients and weights to recover the private training image data of users.
Ihe goals of the experiment is to recover user’s private training image data, we first focus on the reconstruction of a single input image and the label from the gradients and model weights. In order to measure the effectiveness of data reconstruction attack, we measure Attack Success Rate, Attack iteration and the reconstruction error between each reconstructed image and it’s ground truth.
3.2. Overview of Single-Sample Reconstruction Attack System (SSRAS)
In this paper, we propose a Single-Sample Reconstruction Attack System. The overall framework of Single-Sample Reconstruction Attack System is illustrated in
Figure 2.
Figure 2 shows the overview of Single-Sample Reconstruction Attack System. In Federated Learning, this security threat may come from: honest-but-curious server or malicious client. They can get gradients according to user’s private training image data, the architecture of neural network and the associated weight information. Based on the existing information, the server and client can reconstruct user’s private training image data. Let us consider a classification scenario.
Step 1, Label Restoration. In Federated Learning, the malicious server or client can make use of gradients according to the last fully connected layer to restore grouth truth label. It needs base condition that the previous layer before fully connected layer, has used non-negative activation functions, such as ReLU or sigmoid, which plays a key position.
Step 2, The way of Attacks. If the grouth truth label has been determined, we can use algorithm 2 or 3 to reconstruct the image. If the label cannot be determined, we also can use algorithm 1 to reconstruct the label, and then utilize R-GAP to recover user’s private training image data. In this way, it can extends the attack from the fully connected layer to the convolutional layer. Reconstructed image is better.
Step 3, Fidelity Regularization. In order to guide attack from gradients towards natural images, we adopt to the loss function to steer reconstructed image away from unrealistic image.
Step 4, Rank Analysis Index. Rank analysis Index provides an overall estimate of whether the data can be reconstructed.
In short, by using step 1 and 2, we can get a reconstructed image with some noise points. In order to make reconstructed image closer to the natural images, we can use step 3. The purpose of step 4 is to judge the security of the neural network according to the existing gradient and weight information.
3.2.1. Label Restoration
Zhao et al. [
10] extract the ground-truth from the shared gradient based on the last fully connected layer. It is single-sample batch as given gradient.
The negative gradient can be the index of the ground truth.
Multi-sample gradients are averaged over
k images in a batch [
23], Yin et al. [
23] propose an algorithm for target class label recovery from given gradients.
There can utilize column minimum values to formulate the final label restoration algorithm for batch size
k:
is the ground truth of batch size k. is the label corresponding to .
Yin et al. [
23] extend label Restoration algorithm from single-sample [
10] to multisample in a batch. Note that, whether the label can be recovered from the gradient depends on the activation function of the last fully connected layer. It needs base condition that the previous layer before fully connected layer, has used non-negative activation function, such as ReLU or sigmoid, which plays a key position.
3.2.2. The Way of Attacks
The relevant technologies used in this privacy attack system are Algorithm 1 (DLG) and Algorithm 2 (Inverting Gradients) belong to Optimization-Based Attack from Gradients. Algorithm 3 (R-GAP) belongs to Linear Equation Solver Attack.
DLG [
9] adopts Formula (21) to recover the original input image
,
from a transmitted gradient
. The algorithm is described in Algorithm 1 in detail.
Algorithm 1 DLG |
Input: : Model; : weights; : gradients of training data; Result: , private training data- 1:
/*Initialize dummy inputs and labels. */ - 2:
- 3:
for to n do - 4:
/*Compute dummy gradients. */ - 5:
- 6:
- 7:
/*Update data to match gradients.*/ - 8:
- 9:
end for - 10:
return , - 11:
end procedure
|
Inverting Gradients [
11] proposes to use cosine distance. The specific form is shown in Formula (22). The algorithm is described in Algorithm 2 in detail.
Algorithm 2 Inverting Gradients |
Input: : Model; : weights; : gradients of training data; Result: , private training data /*Initialize dummy inputs and labels. */- 2:
for to n do - 4:
/*Compute dummy gradients. */ - 6:
/*Update data to match gradients.*/ - 8:
end for - 10:
return , end procedure
|
R-GAP [
22] breaks the gradient attack down to a recursive process of solving linear equations. They propose recursive procedure to recover data from gradient in deep neural networks. However, the gradient is aggregation gradient in CNNs. In order to effective analytic gradient attack for CNNs, there need peel off padding entries, and the stride should be appropriate, it need equal the size of convolutional kernel. Gradient constraints can be described as follows,
denotes the input in the ith layer and is a coefficient matrix containing all gradient constraints in the ith layer.
Weight constraints is shown in Formula (24), at the same time, it needs assumption that they know input of the subsequent layer.
is corresponding circulant matrix representing convolutional kernel, is weight constraints. The Formulas (23) and (24) can form matix A and B. The reconstructed user’s private training image data can be transformed into a matrix solution, . The condition of restoring user’s private training image is that the corresponding coefficient matrix A is equal to the number of entries.
3.2.3. Fidelity Regularization
In order to guide attack from gradients towards natural images, we adopt
to the loss function to steer reconstructed image away from unrealistic image [
23].
and
denote standard image priors, with scaling factors
and
. The cost function of DLG is shown in Formula (25)
The cost function of Inverting Gradients is shown in Formula (26)
R-GAP can make use of difference between reconstructed image, and adopt smooth version. The purpose of smooth version is to eliminate noise. There are two ways to do it in spatial domain or in frequency domain. Low—Pass filtering can be used to remove noise in spatial domain. While, the frequency domain can remove noise by removing high frequency components.
3.2.4. Rank Analysis Index
Rank Analysis Index (RA-I) is used to measure the possible of whether the user’s raw image data can be reconstructed.
Fan et al. [
19] analyze the Bias Attacks as a system of linear equations, and perform a rank analysis about network’s security. The matrix expression is as follows,
denotes the partial derivative of the loss function w.r.t.the model bias, denotes the partial derivative of the loss function w.r.t.the model weight parameter.
The condition of restoring user’s private training image is that the corresponding coefficient matrix is equal to the number of entries.
According to R-GAP algorithm, weight constraints can derived a new constraint, which is named virtual constraints
. The virtual constraints
can be derived from the weight constraints of d-1 layer.
Split W, O into two parts,
From Formula (28), we can get Formula (29) and Formula (30).
Because of the activation function is the identity function, there is
. We can draw a conclusion as follows.
From Formula (33), virtual constrains can be derived from weights. Further, the virtual constrains of each layer can be derived, which based on the assumption that we know the output of the current layer.
Rank analysis includes gradient constraints
, weight constraints
and virtual constraints
. It is named Rank Analysis Index (RA-I). All conclusions are based on assumption that the activation function is ReLU or LeakyRelu. The specific form of RA-I can be described by Formula (34).
represents the number of input entries in i-th layer, denotes gradient constraints in the ith layer, represents weight constraints in the ith layer, represents the number of virtual constraints in the ith layer.
, indicates that complete reconstruction cannot be performed. The larger the index, the worse the reconstruction quality. , indicates the ability to fully recover the input.