Condensed vector machines: Learning fast machine for large data

Dung Duc Nguyen, Kazunori Matsumoto, Yasuhiro Takishima, Kazuo Hashimoto

研究成果: Article査読

13 被引用数 (Scopus)

抄録

Scalability is one of the main challenges for kernel-based methods and support vector machines (SVMs). The quadratic demand in memory for storing kernel matrices makes it impossible for training on million-size data. Sophisticated decomposition algorithms have been proposed to efficiently train SVMs using only important examples, which ideally are the final support vectors (SVs). However, the ability of the decomposition method is limited to large-scale applications where the number of SVs is still too large for a computer's capacity. From another perspective, the large number of SVs slows down SVMs in the testing phase, making it impractical for many applications. In this paper, we introduce the integration of a vector combination scheme to simplify the SVM solution into an incremental working set selection for SVM training. The main objective of the integration is to maintain a minimal number of final SVs, bringing a minimum resource demand and faster training time. Consequently, the learning machines are more compact and run faster thanks to the small number of vectors included in their solution. Experimental results on large benchmark datasets shows that the proposed condensed SVMs achieve both training and testing efficiency while maintaining a generalization ability equivalent to that of normal SVMs.

本文言語English
論文番号5605254
ページ(範囲)1903-1914
ページ数12
ジャーナルIEEE Transactions on Neural Networks
21
12
DOI
出版ステータスPublished - 2010 12
外部発表はい

ASJC Scopus subject areas

  • Software
  • Computer Science Applications
  • Computer Networks and Communications
  • Artificial Intelligence

フィンガープリント 「Condensed vector machines: Learning fast machine for large data」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル