Distributed Machine Learning Algorithms: CommunicationComputation Tradeoffs  Part 2 Sundararajan Sellamanickam
Distributed machine learning is an important area that has been receiving considerable attention from academic and industrial communities, as data is growing in unprecedented rate. In the first part of the talk, we review several popular approaches that are proposed/used to learn classifier models in the big data scenario. With commodity clusters priced on system configurations becoming popular, machine learning algorithms have to be aware of the computation and communication costs involved in order to be cost effective and efficient. In the second part of the talk, we focus on methods that address this problem; in particular, considering different data distribution settings (e.g., example and feature partitions), we present efficient distributed learning algorithms that tradeoff computation and communication costs. 

Distributed Machine Learning Algorithms: CommunicationComputation Tradeoffs  Part 1 Sundararajan Sellamanickam
Distributed machine learning is an important area that has been receiving considerable attention from academic and industrial communities, as data is growing in unprecedented rate. In the first part of the talk, we review several popular approaches that are proposed/used to learn classifier models in the big data scenario. With commodity clusters priced on system configurations becoming popular, machine learning algorithms have to be aware of the computation and communication costs involved in order to be cost effective and efficient. In the second part of the talk, we focus on methods that address this problem; in particular, considering different data distribution settings (e.g., example and feature partitions), we present efficient distributed learning algorithms that tradeoff computation and communication costs. 

Scaling Up Reinforcement Learning B. Ravindran
Distributed machine learning is an important area that has been receiving considerable attention from academic and industrial communities, as data is growing in unprecedented rate. In the first part of the talk, we review several popular approaches that are proposed/used to learn classifier models in the big data scenario. With commodity clusters priced on system configurations becoming popular, machine learning algorithms have to be aware of the computation and communication costs involved in order to be cost effective and efficient. In the second part of the talk, we focus on methods that address this problem; in particular, considering different data distribution settings (e.g., example and feature partitions), we present efficient distributed learning algorithms that tradeoff computation and communication costs. 

Reinforcement Learning: An Introduction B. Ravindran
Distributed machine learning is an important area that has been receiving considerable attention from academic and industrial communities, as data is growing in unprecedented rate. In the first part of the talk, we review several popular approaches that are proposed/used to learn classifier models in the big data scenario. With commodity clusters priced on system configurations becoming popular, machine learning algorithms have to be aware of the computation and communication costs involved in order to be cost effective and efficient. In the second part of the talk, we focus on methods that address this problem; in particular, considering different data distribution settings (e.g., example and feature partitions), we present efficient distributed learning algorithms that tradeoff computation and communication costs. 

Submodular Optimization and Machine Learning  Part 2 Stefanie Jegelka
Many problems in machine learning that involve discrete structures or subset selection may be phrased in the language of submodular set functions. The property of submodularity, also referred to as a 'discrete analog of convexity', expresses the notion of diminishing marginal returns, and captures combinatorial versions of rank and dependence. Submodular functions occur in a variety of areas including graph theory, information theory, combinatorial optimization, stochastic processes and game theory. In machine learning, they emerge in different forms as the potential functions of graphical models, as the utility functions in active learning and sensing, in models of diversity, in structured sparse estimation or network inference. The lectures will give an introduction to the theory of submodular functions, some applications in machine learning and algorithms for minimizing and maximizing submodular functions that exploit ties to both convexity and concavity. 

Submodular Optimization and Machine Learning  Part 1 Stefanie Jegelka
Many problems in machine learning that involve discrete structures or subset selection may be phrased in the language of submodular set functions. The property of submodularity, also referred to as a 'discrete analog of convexity', expresses the notion of diminishing marginal returns, and captures combinatorial versions of rank and dependence. Submodular functions occur in a variety of areas including graph theory, information theory, combinatorial optimization, stochastic processes and game theory. In machine learning, they emerge in different forms as the potential functions of graphical models, as the utility functions in active learning and sensing, in models of diversity, in structured sparse estimation or network inference. The lectures will give an introduction to the theory of submodular functions, some applications in machine learning and algorithms for minimizing and maximizing submodular functions that exploit ties to both convexity and concavity. 

Panel Q and A Prateek Jain, ChinJen Lin, Aditya Gopalan, Suvrit Sra, and Stefanie Jegelka


Introduction to largescale optimization  Part 2 Suvrit Sra
These lectures will cover both basics as well as cuttingedge topics in largescale convex and nonconvex optimization (continuous case only). Examples include stochastic convex optimization, variance reduced stochastic gradient, coordinate descent methods, proximalmethods, operator splitting techniques, and more. The lectures will also cover relevant mathematical background, as well as some pointers to interesting directions of future research. 

Introduction to largescale optimization  Part1 Suvrit Sra
These lectures will cover both basics as well as cuttingedge topics in largescale convex and nonconvex optimization (continuous case only). Examples include stochastic convex optimization, variance reduced stochastic gradient, coordinate descent methods, proximalmethods, operator splitting techniques, and more. The lectures will also cover relevant mathematical background, as well as some pointers to interesting directions of future research (time permitting). 

Provable Nonconvex Projections for Highdimensional Learning Problems  Part1 Prateek Jain
Typical highdimensional learning problems such as sparse regression, lowrank matrix completion, robust PCA etc can be solved using projections onto nonconvex sets. However, providing theoretical guarantees for such methods is difficult due to the nonconvexity in projections. In this talk, we will discuss some of our recent results that show that nonconvex projections based methods can be used to solve several important problems in this area such as: a) sparse regression, b) lowrank matrix completion, c) robust PCA. In this talk, we will give an overview of the stateoftheart for these problems and also discuss how simple nonconvex techniques can significantly outperform stateoftheart convex relaxation based techniques and provide solid theoretical results as well. For example, for robust PCA, we provide first provable algorithm with time complexity O(n^{2} r) which matches the time complexity of normal SVD and is faster than the usual nuclear+L_{1}regularization methods that incur O(n^{3}) time complexity. This talk is based on joint works with Ambuj Tewari, Purushottam Kar, Praneeth Netrapalli, Animashree Anandkumar, U N Niranjan, and Sujay Sanghavi. 
 Microsoft Research India
“Scientia”
196/36 2nd Main
Sadashivnagar, Bangalore 560 080
India