z-logo
open-access-imgOpen Access
Effects of Sparse Initialization in Deep Belief Networks
Author(s) -
Grzegorczyk Karol,
Marcin Kurdziel,
Piotr Iwo Wójcik
Publication year - 2015
Publication title -
computer science
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.145
H-Index - 5
eISSN - 2300-7036
pISSN - 1508-2806
DOI - 10.7494/csci.2015.16.4.313
Subject(s) - initialization , computer science , discriminative model , deep belief network , artificial intelligence , backpropagation , set (abstract data type) , artificial neural network , deep neural networks , deep learning , pattern recognition (psychology) , machine learning , programming language
Deep neural networks are often trained in two phases: first hidden layers are pretrained in an unsupervised manner and then network is fine-tuned with error backpropagation. Pretraining is often carried out using Deep Belief Networks (DBNs), with initial weights set to small random values. However, recent results established that well-designed initialization schemes, e.g. Sparse Initialization (SI), can greatly improve performance of networks that do not use pretraining. An interesting question arising from these results is whether such initialization techniques wouldn't also improve pretrained networks? To shed light on this question, in this work we evaluate SI in DBNs that are used to pretrain discriminative networks. The motivation behind this research is our observation that SI has an impact on the features learned by a DBN during pretraining. Our results demonstrate that this improves network performance: when pretraining starts from sparsely initialized weight matrices networks achieve lower classification error after fine-tuning.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom