LOSSGRAD: Automatic Learning Rate in Gradient Descent
cytuj
pobierz pliki
RIS BIB ENDNOTEChoose format
RIS BIB ENDNOTELOSSGRAD: Automatic Learning Rate in Gradient Descent
Publication date: 2018
Schedae Informaticae, 2018, Volume 27, pp. 47 - 57
https://doi.org/10.4467/20838476SI.18.004.10409Authors
LOSSGRAD: Automatic Learning Rate in Gradient Descent
In this paper, we propose a simple, fast and easy to implement algorithm LOSSGRAD (locally optimal step-size in gradient descent), which automatically modifies the step-size in gradient descent during neural networks training. Given a function f, a point x, and the gradient ▽xf of f, we aim to find the step-size h which is (locally) optimal, i.e. satisfies:
h = arg min f(x - t▽xf).
t≥0
Making use of quadratic approximation, we show that the algorithm satisfies the above assumption. We experimentally show that our method is insensitive to the choice of initial learning rate while achieving results comparable to other methods.
Information: Schedae Informaticae, 2018, Volume 27, pp. 47 - 57
Article type: Original article
Titles:
LOSSGRAD: Automatic Learning Rate in Gradient Descent
LOSSGRAD: Automatic Learning Rate in Gradient Descent
Institute of Mathematics, Faculty of Physics, Mathematics and Computer Science, Cracow University of Technology
Faculty of Mathematics and Computer Science, Jagiellonian University, Krakow, Poland
Faculty of Mathematics and Computer Science, Jagiellonian University ul. Łojasiewicza 6, 30-348 Kraków, Poland
Published at: 2018
Article status: Open
Licence: CC BY-NC-ND
Percentage share of authors:
Article corrections:
-Publication languages:
EnglishView count: 1855
Number of downloads: 1352