Authors
Minmin Chen, Kilian Q Weinberger, Zhixiang Xu, Fei Sha
Publication date
2015/1/1
Journal
The Journal of Machine Learning Research
Volume
16
Issue
1
Pages
3849-3875
Publisher
JMLR. org
Description
Stacked denoising autoencoders (SDAs) have been successfully used to learn new representations for domain adaptation. They have attained record accuracy on standard benchmark tasks of sentiment analysis across different text domains. SDAs learn robust data representations by reconstruction, recovering original features from data that are artificially corrupted with noise. In this paper, we propose marginalized Stacked Linear Denoising Autoencoder (mSLDA) that addresses two crucial limitations of SDAs: high computational cost and lack of scalability to high-dimensional features. In contrast to SDAs, our approach of mSLDA marginalizes noise and thus does not require stochastic gradient descent or other optimization algorithms to learn parameters—in fact, the linear formulation gives rise to a closed-form solution. Consequently, mSLDA, which can be implemented in only 20 lines of MATLABTM, is about two orders of magnitude faster than a corresponding SDA. Furthermore, the representations learnt by mSLDA are as effective as the traditional SDAs, attaining almost identical accuracies in benchmark tasks.
Total citations
201620172018201920202021202220232024148996554
Scholar articles
M Chen, KQ Weinberger, Z Xu, F Sha - The Journal of Machine Learning Research, 2015