Citation: | LIAN Zifeng, JING Xiaojun, WANG Xiaohan, et al., “DropConnect Regularization Method with Sparsity Constraint for Neural Networks,” Chinese Journal of Electronics, vol. 25, no. 1, pp. 152-158, 2016, doi: 10.1049/cje.2016.01.023 |
G.E. Hinton, N. Srivastava, A. Krizhevsky, et al., “Improving neural networks by preventing co-adaptation of feature detectors”, arXiv preprint arXiv:1207.0580, 2012.
|
L. Wan, et al., “Regularization of neural networks using Drop- Connect”, Proceedings of the 30th International Conference on Machine Learning (ICML-13), pp.1058-1066, 2013.
|
Y. Bengio, “Learning deep architectures for AI”, Foundations and trends in Machine Learning, Vol.2, No.1, pp.1-127, 2009
|
Y. Bengio, et al., “Representation learning: A review and new perspectives”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol.35, No.8, pp.1798-1828, 2013.
|
A.S. Weigend, D.E. Rumelhart and B.A. Huberman, “Generalization by weight-elimination with application to forecasting”, Neural Information Processing Systems (NIPS), 1991.
|
D.J.C. Mackay, “Probable networks and plausible predictions — A review of practical bayesian methods for supervised neural networks”, Network Computation in Neural Systems, Vol.6, No.3, pp.469-505, 1995.
|
P. Vincent, et al., “Extracting and composing robust features with denoising autoencoders”, Proceedings of the 25th International Conference on Machine Learning, ACM, pp.1096-1103, 2008.
|
P. Vincent, et al., “Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion”, Proceedings of the 27th International Conference on Machine Learning, ACM, pp.3371-3408, 2010.
|
Y. LeCun, L. Bottou, Y. Bengio and P. Haner, “Gradient-based learning applied to document recognition”, Proceedings of the IEEE, Vol.86, No.11, pp.2278-2324, 1998.
|
Hinton, Geoffrey, S. Osindero and Yee-Whye Teh, “A fast learning algorithm for deep belief nets”, Neural computation, Vol.18, No.7, pp.1527-1554, 2006.
|
D.J. Field, “What is the goal of sensory coding?”, Neural Computation, Vol.6, pp.559-601, 1994.
|
P. Zhao, G. Rocha and B. Yu, “The composite absolute penalties family for grouped and hierarchical variable selection”, Annals of Statistics, Vol.37, No.6A, pp.3468-3497, 2009.
|
P.O. Hoyer, “Non-negative matrix factorization with sparseness constraints”, The Journal of Machine Learning Research, Vol.5, pp.1457-1469, 2004.
|
A. Krizhevsky, “Learning multiple layers of features from tiny images”, Master's Thesis, University of Toronto, 2009.
|
A. Krizhevsky, “Cuda-convnet”, available at http://code.google.c om/p/cuda-convnet/, 2015-12-18.
|
Torralba, Antonio, R. Fergus and W.T. Freeman, “80 million tiny images: A large data set for nonparametric object and scene recognition”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol.30, No.11, pp.1958-1970, 2008.
|