Publication | Closed Access
On the problem of local minima in backpropagation
651
Citations
29
References
1992
Year
Mathematical ProgrammingEngineeringMachine LearningNeural Networks (Machine Learning)Variational AnalysisRecurrent Neural NetworkSocial SciencesPattern RecognitionNull CostRegularization (Mathematics)Approximation TheoryComputational Learning TheoryMachine Learning ModelTheoretical FrameworkInverse ProblemsNeural Networks (Computational Neuroscience)Computer ScienceNondifferentiable OptimizationLocal MinimaDeep Neural NetworksClassical Machine Learning
The authors propose a theoretical framework for backpropagation (BP) in order to identify some of its limitations as a general learning procedure and the reasons for its success in several experiments on pattern recognition. The first important conclusion is that examples can be found in which BP gets stuck in local minima. A simple example in which BP can get stuck during gradient descent without having learned the entire training set is presented. This example guarantees the existence of a solution with null cost. Some conditions on the network architecture and the learning environment that ensure the convergence of the BP algorithm are proposed. It is proven in particular that the convergence holds if the classes are linearly separable. In this case, the experience gained in several experiments shows that multilayered neural networks (MLNs) exceed perceptrons in generalization to new examples.< <ETX xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">></ETX>
| Year | Citations | |
|---|---|---|
Page 1
Page 1