Doorkeeper

数理科学チームセミナー

2018-01-23(火)13:00 - 15:00 JST

日本橋オフィス

申し込む

申し込み受付は終了しました

今後イベント情報を受け取る

参加費無料

詳細

Train faster, generalize better: Stability of stochastic gradient descent

We would talk about “train faster generalize better” by Moritz Hardt et al.
https://arxiv.org/pdf/1509.01240.pdf
This paper gives us the notion of the “stability” of stochastic algorithms.
Intuitively, the output of the stable algorithm changes just a little (bounded by a epsilon), when we change a sample in the training data set.
With this notion, we show two theorems about stable algorithms.
One is that if the stochastic algorithm is stable, then the generalization gap is bounded.
The second is that if the number of the iteration is bounded, then SGD is stable.
As a consequence, these two theorems indicate that if we can train DNNs faster, then they generalize better. (Remenber the case of "random labelling" by Zhang, et al.)

コミュニティについて

RIKEN AIP Public

RIKEN AIP Public

Public events of RIKEN Center for Advanced Intelligence Project (AIP)

メンバーになる