Abstract
A recent work introduced the concept of deep dictionary learning. The first level is a dictionary learning stage where the inputs are the training data and the outputs are the dictionary and learned coefficients. In subsequent levels of deep dictionary learning, the learned coefficients from the previous level acts as inputs. This is an unsupervised representation learning technique. In this work we empirically compare and contrast with similar deep representation learning techniques – deep belief network and stacked autoencoder. We delve into two aspects; the first one is the robustness of the learning tool in the presence of noise and the second one is the robustness with respect to variations in the number of training samples. The experiments have been carried out on several benchmark datasets. We find that the deep dictionary learning method is the most robust.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Sutskever, I., Tieleman, T.: On the convergence properties of contrastive divergence. In: AISTATS (2010)
Lee, D.D., Seung, H.S.: Learning the parts of objects by non-negative matrix factorization. Nature 401(6755), 788–791 (1999)
Rubinstein, R., Bruckstein, A.M., Elad, M.: Dictionaries for sparse representation modeling. Proc. IEEE 98(6), 1045–1057 (2010)
Tariyal, S., Majumdar, A., Singh, R., Vatsa, M.: Greedy Deep Dictionary Learning, arXiv:1602.00203v1
Engan, K., Aase, S., Hakon-Husoy, J.: Method of optimal directions for frame design. In: IEEE ICASSP (1999)
Jain, P., Netrapalli, P., Sanghavi, S.: Low-rank matrix completion using alternating minimization. In: Symposium on Theory of Computing (2013)
Agarwal, A., Anandkumar, A., Jain, P., Netrapalli, P.: Learning sparsely used overcomplete dictionaries via alternating minimization. In: International Conference on Learning Theory (2014)
Spielman, D.A., Wang, H., Wright, J.: Exact recovery of sparsely-used dictionaries. In: International Conference on Learning Theory (2012)
Arora, S., Bhaskara, A., Ge, R., Ma, T.: More Algorithms for Provable Dictionary Learning, arXiv:1401.0579v1
Courville, A., Bergstra, J., Bengio, Y.: An empirical evaluation of deep architectures on problems with many factors of variation. In: ICML (2007)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Singhal, V., Gogna, A., Majumdar, A. (2016). Deep Dictionary Learning vs Deep Belief Network vs Stacked Autoencoder: An Empirical Analysis. In: Hirose, A., Ozawa, S., Doya, K., Ikeda, K., Lee, M., Liu, D. (eds) Neural Information Processing. ICONIP 2016. Lecture Notes in Computer Science(), vol 9950. Springer, Cham. https://doi.org/10.1007/978-3-319-46681-1_41
Download citation
DOI: https://doi.org/10.1007/978-3-319-46681-1_41
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-46680-4
Online ISBN: 978-3-319-46681-1
eBook Packages: Computer ScienceComputer Science (R0)