您的瀏覽器不支援JavaScript語法,網站的部份功能在JavaScript沒有啟用的狀態下無法正常使用。

中央研究院 資訊科學研究所

活動訊息

友善列印

列印可使用瀏覽器提供的(Ctrl+P)功能

學術演講

:::

Combine the Passive and Aggressive Algorithm with a Proximal Model

  • 講者李育杰 教授 (國立台灣科技大學資訊工程系)
    邀請人:張復老師
  • 時間2011-08-15 (Mon.) 16:00 ~ 18:00
  • 地點本所新館一樓106演講廳
摘要

Due to the nature of the online learning setting, the online learning algorithms have been successfully applied to large scale classification tasks. This type of algorithms can be interpreted as a stochastic gradient descent method that try to find the solution of the underlying minimization problem with the objective function consisting the sum of training loss and regularization term. How to decide the learning rate becomes an important issue in this kind of learning algorithms. The passive and aggressive (PA) algorithm proposed an updating scheme to determine the new classifier. It suggests that the new classifier should not only classify the new arriving data correctly but also as close to current classifier as possible. A closed form of updating rule was derived that makes PA algorithm training extremely fast. However, a lack of memory for previous instances might hurt the learning efficiency. We propose a new updating rule that takes the accumulated variances and means information for different classes into account. Using these information, a good proximal classification model can be generated. We augment the proximal model into PA algorithm updating rule in a closed form. Thus, our proposed method has the same computational advantage as PA algorithm. The preliminary numerical results show that our proposed method is less sensitive to the input order of training data than PA algorithm. It will return a near optimal classifier in a single pass. Two consecutive classifiers generated by two training epochs are very close. These evidences show that our method is suitable for the learning task with extremely large dataset.