Concepedia

Publication | Closed Access

Gradient calculations for dynamic recurrent neural networks: a survey

590

Citations

113

References

1995

Year

Abstract

Surveys learning algorithms for recurrent neural networks with hidden units and puts the various techniques into a common framework. The authors discuss fixed point learning algorithms, namely recurrent backpropagation and deterministic Boltzmann machines, and nonfixed point algorithms, namely backpropagation through time, Elman's history cutoff, and Jordan's output feedback architecture. Forward propagation, an on-line technique that uses adjoint equations, and variations thereof, are also discussed. In many cases, the unified presentation leads to generalizations of various sorts. The author discusses advantages and disadvantages of temporally continuous neural networks in contrast to clocked ones continues with some "tricks of the trade" for training, using, and simulating continuous time and recurrent neural networks. The author presents some simulations, and at the end, addresses issues of computational complexity and learning speed.

References

YearCitations

1983

44K

1960

30.4K

1996

14.4K

1989

11.6K

1990

10.6K

1990

8K

1985

6K

1990

4.8K

1989

4.4K

1985

3.3K

Page 1