Optimization of overtraining and overgeneration

Goutam Chakraborty, Norio Shiratori, Shoichi Noguchi

    研究成果: Conference contribution

    1 引用 (Scopus)


    The task of any supervised classifier is to assign optimum boundaries in the input space, for the different class membership. This is done using informations from the available set of known samples. This mapping of sample position in the input space to sample class is further used to classify unknown samples. The available set of known sample is generally a finite set. A boundary exactly defined by those finite sample set is usually not the best boundary to classify the new unknown samples. We end up with a overfitted boundary i.e. a overtrained classifier, resulting in poor classification for unknown new samples. We therefore need to smooth the boundary to be able to generalize for the unknown samples. But to what extent? If we smooth the boundary too much, we will not be exploiting all the class informations contained in the known sample set, and the classification result will again be poor. Depending on the number of known samples and the dimension of the actual solution (which, of course, is not known in any of the practical problems), there will be a certain amount of smoothness, which is optimum for generalization. In this paper, we are trying to focus on this problem. We introduce some practical ways to arrive at optimum smoothness, with regards to single hidden layer neural network classifier using radial basis function.

    ホスト出版物のタイトルProceedings of the International Joint Conference on Neural Networks
    出版場所Piscataway, NJ, United States
    出版者Publ by IEEE
    ISBN(印刷物)0780314212, 9780780314214
    出版物ステータスPublished - 1993
    イベントProceedings of 1993 International Joint Conference on Neural Networks. Part 1 (of 3) - Nagoya, Jpn
    継続期間: 1993 10 251993 10 29


    OtherProceedings of 1993 International Joint Conference on Neural Networks. Part 1 (of 3)
    Nagoya, Jpn

    ASJC Scopus subject areas

    • Engineering(all)

    フィンガープリント Optimization of overtraining and overgeneration' の研究トピックを掘り下げます。これらはともに一意のフィンガープリントを構成します。

  • これを引用

    Chakraborty, G., Shiratori, N., & Noguchi, S. (1993). Optimization of overtraining and overgeneration. : Proceedings of the International Joint Conference on Neural Networks (巻 3, pp. 2257-2262). Publ by IEEE.