## 抄録

The α-logarithm extends the logarithm as the special case of α = -1. Usage of α-related information measures based upon this extended logarithm is expected to be effective to speedup of convergence, i.e., on the improvement of learning aptitude. In this paper, two typical cases are investigated. One is the α-EM algorithm (α-Expectation-Maximization algorithm) which is derived from the α-log-likelihood ratio. The other is the α-ICA (α-Independent Component Analysis) which is formulated as minimizing the α-mutual information. In the derivation of both algorithms, the α-divergence plays the main role. For the α-EM algorithm, the reason for the speedup is explained using Hessian and Jacobian matrices for learning. For the α-ICA learning, methods of exploiting the past and future information are presented. Examples are shown on single-loop α-EM's and sample-based α-ICA's. In all cases, effective speedups are observed. Thus, this paper's examples together with formerly reported ones are evidences that the speed improvement by the α-logarithm is a general property beyond individual problems.

本文言語 | English |
---|---|

ホスト出版物のタイトル | Proceedings of the International Joint Conference on Neural Networks |

Place of Publication | Piscataway, NJ, United States |

出版社 | IEEE |

ページ | 351-356 |

ページ数 | 6 |

巻 | 3 |

出版ステータス | Published - 2000 |

イベント | International Joint Conference on Neural Networks (IJCNN'2000) - Como, Italy 継続期間: 2000 7月 24 → 2000 7月 27 |

### Other

Other | International Joint Conference on Neural Networks (IJCNN'2000) |
---|---|

City | Como, Italy |

Period | 00/7/24 → 00/7/27 |

## ASJC Scopus subject areas

- ソフトウェア