This article proposes new algorithms both for Boltzmann Machine (BM) and Mean Field Theory (MFT) learning. They use momentum terms that are derived theoretically to accelerate their learning speeds. The derivation of the new algorithms is based on the following assumptions: (1) The alternate cost function is Gn = Στnζn-τGτ, where Gτ is the information-theoretical measure at the learning time τ, not G which is the commonly used information-theoretical measure in the derivation of BM learning. (2) The most recent weights are assumed in calculating Gn, which technique is used in the derivation of the recursive least-squares algorithm. As a result, momentum terms that accelerate learning can be derived in the BM and MFT learning algorithms. In addition, note that the proposed methods can be used both in batch-mode and pattern-by-pattern learning. Computer simulation is carried out to conform the effectiveness of the proposed MFT algorithm by comparing it with the conventional MFT algorithm.
|Journal of artificial neural networks
|Published - 1995 12月 1
ASJC Scopus subject areas