Adversarial deep embedded clustering: on a better trade-off between feature randomness and feature drift
N Mrabah, M Bouguessa… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
N Mrabah, M Bouguessa, R Ksantini
IEEE Transactions on Knowledge and Data Engineering, 2020•ieeexplore.ieee.orgTo overcome the absence of concrete supervisory signals, deep clustering models construct
their own labels based on self-supervision and pseudo-supervision. However, applying
these techniques can cause Feature Randomness and Feature Drift. In this paper, we
formally characterize these two new concepts. On one hand, Feature Randomness takes
place when a considerable portion of the pseudo-labels is deemed to be random. In this
regard, the trained model can learn non-representative features. On the other hand, Feature …
their own labels based on self-supervision and pseudo-supervision. However, applying
these techniques can cause Feature Randomness and Feature Drift. In this paper, we
formally characterize these two new concepts. On one hand, Feature Randomness takes
place when a considerable portion of the pseudo-labels is deemed to be random. In this
regard, the trained model can learn non-representative features. On the other hand, Feature …
To overcome the absence of concrete supervisory signals, deep clustering models construct their own labels based on self-supervision and pseudo-supervision. However, applying these techniques can cause Feature Randomness and Feature Drift. In this paper, we formally characterize these two new concepts. On one hand, Feature Randomness takes place when a considerable portion of the pseudo-labels is deemed to be random. In this regard, the trained model can learn non-representative features. On the other hand, Feature Drift takes place when the pseudo-supervised and the reconstruction losses are jointly minimized. While penalizing the reconstruction loss aims to preserve all the inherent data information, optimizing the embedded-clustering objective drops the latent between-cluster variances. Due to this compromise, the clustering-friendly representations can be easily drifted. In this context, we propose ADEC (Adversarial Deep Embedded Clustering) a novel autoencoder-based clustering model, which relies on a discriminator network to reduce random features while avoiding the drifting effect. Our new metrics and allows to, respectively, assess the level of Feature Randomness and Feature Drift. We empirically demonstrate the suitability of our model on handling these problems using benchmark real datasets. Experimental results validate that our model outperforms state-of-the-art autoencoder-based clustering methods.
ieeexplore.ieee.org