Samantika S. Sury

We don’t have enough information about this author to calculate their statistics. If you think this is an error let us know.
Learn More
Stochastic Gradient Descent (SGD) is a popular optimization method used to train a variety of machine learning models. Most of SGD work to-date has concentrated on improving its statistical efficiency, in terms of rate of convergence to the optimal solution. At the same time, as parallelism of modern CPUs continues to increase through progressively higher(More)
  • 1