Abstract
The restricted Boltzmann machine (RBM) has been used as building blocks for many successful deep learning models, e.g., deep belief networks (DBN) and deep Boltzmann machine (DBM) etc. The training of RBM can be extremely slow in pathological regions. The second order optimization methods, such as quasi-Newton methods, were proposed to deal with this problem. However, the non-convexity results in many obstructions for training RBM, including the infeasibility of applying second order optimization methods. In order to overcome this obstruction, we introduce an em-like iterative project quasi-Newton (IPQN) algorithm. Specifically, we iteratively perform the sampling procedure where it is not necessary to update parameters, and the sub-training procedure that is convex. In sub-training procedures, we apply quasi-Newton methods to deal with the pathological problem. We further show that Newton's method turns out to be a good approximation of the natural gradient (NG) method in RBM training. We evaluate IPQN in a series of density estimation experiments on the artificial dataset and the MNIST digit dataset. Experimental results indicate that IPQN achieves an improved convergent performance over the traditional CD method.
Original language | English |
---|---|
Title of host publication | 30th AAAI Conference on Artificial Intelligence, AAAI 2016 |
Publisher | AAAI press |
Pages | 4236-4237 |
Number of pages | 2 |
ISBN (Electronic) | 9781577357605 |
Publication status | Published - 1 Jan 2016 |
Externally published | Yes |
Event | 30th AAAI Conference on Artificial Intelligence, AAAI 2016 - Phoenix Convention Center, Phoenix, United States Duration: 12 Feb 2016 → 17 Feb 2016 |
Conference
Conference | 30th AAAI Conference on Artificial Intelligence, AAAI 2016 |
---|---|
Country/Territory | United States |
City | Phoenix |
Period | 12/02/16 → 17/02/16 |
ASJC Scopus subject areas
- Artificial Intelligence