Image Home
Image People
Image Publication
Image Applications
Image Data & Code
Image Library
Image Seminar
Image Link
Image Album

Search LAMDA

Seminar abstract

Distributed and Stochastic Machine Learning on Big Data

James Tin-Yau Kwok
Hong Kong University of Science and Technology

Abstract: On big data sets, it is often challenging to learn the parameters in a machine learning model. A popular technique is the use of stochastic gradient, which computes the gradient at a single sample instead of over the whole data set. Another alternative is distributed processing, which is particularly natural when a single computer cannot store or process the whole data set. In this talk, some recent extensions will be presented. For stochastic gradient, instead of using the information from only one sample, we incrementally approximate the full gradient by also using old gradient values from the other samples.It enjoys the same computational simplicity as existing stochastic algorithms, but has faster convergence. As for existing distributed machine learning algorithms,they are often synchronized and the system can move forward only at the pace of the slowest worker. I will present an asynchronous algorithm which requires only partial synchronization, and updates from the faster workers can be incorporated more often by the master.

Bio: Prof. Kwok is a Professor in the Department of Computer Science and Engineering, Hong Kong University of Science and Technology. He received his B.Sc. degree in Electrical and Electronic Engineering from the University of Hong Kong and his Ph.D. degree in computer science from the Hong Kong University of Science and Technology. Prof. Kwok has served as an Associate Editor for the IEEE Transactions on Neural Networks and Learning Systems from 2006—2012. Currently he is an Associate Editor for the Neurocomputing journal.
  Name Size

(for FireFox 3+ and IE 7+)
Contact LAMDA: (email) (tel) +86-025-89681608 © LAMDA, 2016