Minimizing Average of Loss Functions Using Gradient Descent and Stochastic Gradient Descent
DOI:
https://doi.org/10.3329/dujs.v64i2.54490Keywords:
Gradient Descent, Stochastic Gradient Descent, Convex Function, Unconstrained Optimization Problems.Abstract
This paper deals with minimizing average of loss functions using Gradient Descent (GD) and Stochastic Gradient Descent (SGD). We present these two algorithms for minimizing average of a large number of smooth convex functions. We provide some discussions on their complexity analysis, also illustrate the algorithms geometrically. At the end, we compare their performance through numerical experiments.
Dhaka Univ. J. Sci. 64(2): 141-145, 2016 (July)
Downloads
44
28