The key point of this paper: the capture of structural information is extended in DEC's single-view depth clustering model, and it is captured by using GCN structure. Compared with the structure of GAE, the part about GCN does not use the reconstruction of adjacency matrix to supervise, but uses the target distribution information of clustering to construct another structure distribution to quantify the supervision of structure information.
Legend: and are input data and reconstruction data respectively. And are the outputs of the first layer DNN and GCN modules respectively. Different colors represent different expressions learned from DNN. The blue solid line indicates that the target distribution is calculated by distribution, and the two red dashed lines indicate the dual self-monitoring mechanism. Target distribution guides the update of DNN module and GCN module at the same time.
Summary: First, construct a KNN diagram according to the original data. Then input the original data and KNN map into AE and GCN respectively. The author connects each layer of AE with the corresponding GCN layer, so that AE-specific representation can be integrated into structure-aware representation through transfer operators. At the same time, a dual self-monitoring mechanism is proposed to monitor the training process of AE and GCN.
B: Generally speaking, the number of layers mentioned when introducing AE structure refers to the number of layers from the first hidden layer to the coding layer except the input layer and the reconstruction layer.
DNN module adopts layered basic AE structure, which will not be described here.
Step-:Obtain the convolution operation output result of the first layer.
The convolution operation of the results of each layer is consistent with that of the graphic nerve, but in the output structure, the author connects the expression matrix of the corresponding layer of DNN module (which will be propagated through the normalized adjacency matrix), and selects the balance factor to combine the information from DNN and GCN.
Step-: But for the output of the first layer, only the original output is kept.
Step-:In the construction of structured information distribution, multi-classification softmax layer is adopted to obtain it.
The results show that the probability sample belongs to the cluster center, which can be regarded as probability distribution.
Advantages of the objective function:
(1) Compared with the traditional multi-classification loss function, KL divergence updates the whole model (soft label) in a more "gentle" way to prevent the data representation from being seriously disturbed;
(2) GCN and DNN modules are unified in the same optimization goal, so that their results tend to be consistent in the training process.
Because the goal of DNN module and GCN module is approximate target distribution, and there is a strong connection between the two modules, it is called double self-monitoring mechanism.
In this paper, the expression of GCN+ AE in different layers is used to increase the structure. The experimental results verify the effectiveness of the model and provide theoretical support. The whole experiment is finished. For the structure of the relationship between samples, a new idea is given, and the overall sample structure is further explored.
I haven't seen this piece about theoretical support, and I'm not going to continue. .