## 抄録

The back-propagation algorithm has been applied to many fields, and has shown large capability of neural networks. Many people use the back-propagation algorithm together with a momentum term to accelerate its convergence. However, in spite of the importance for theoretical studies, theoretical background of a momentum term has been unknown so far. First, this paper explains clearly the theoretical origin of a momentum term in the back-propagation algorithm for both a batch mode learning and a pattern-by-pattern learning. We will prove that the back-propagation algorithm having a momentum term can be derived through the following two assumptions: 1) The cost function is E^{n} = n/Σ/μ α^{n-μ} E_{μ}, where E_{μ} is the summation of squared error at the output layer at the μth learning time and a is the momentum coefficient. 2) The latest weights are assumed in calculating the cost function E^{n}. Next, we derive a simple relationship between momentum, learning rate, and learning speed and then further discussion is made with computer simulation.

本文言語 | English |
---|---|

ページ（範囲） | 1080-1086 |

ページ数 | 7 |

ジャーナル | IEICE Transactions on Information and Systems |

巻 | E78-D |

号 | 8 |

出版ステータス | Published - 1995 8月 1 |

## ASJC Scopus subject areas

- ソフトウェア
- ハードウェアとアーキテクチャ
- コンピュータ ビジョンおよびパターン認識
- 電子工学および電気工学
- 人工知能