Learning and stochastic optimization with non-i.i.d. data

2 391
29.5
Следующее
17.08.16 – 1 1581:26:59
PCPs and Expander Graphs
Популярные
Опубликовано 17 августа 2016, 3:51
ABSTRACT: We study learning and optimization scenarios where the samples we receive do not obey the frequently made i.i.d. assumption, but are coupled over time. We show that as long as the samples come from a suitably mixing process, i.e. the dependence weakens over time, a large class of learning algorithms continue to enjoy good generalization guarantees. The result also has implications for stochastic optimization with non-i.i.d. samples. Specifically, we show that a large class of suitably stable online learning algorithms produce a predictor with a small optimization error, as long as the samples are from a suitably ergodic process. Our mixing assumptions are satisfied by finite-state Markov chains, autoregressive processes, certain infinite and continuous state Markov chains, and various queuing processes. The talk will discuss applications to machine learning with non-i.i.d. data samples, optimization over high-dimensional and combinatorial spaces, and distributed optimization as a few examples. Based on joint work with John Duchi, Michael Jordan and Mikael Johansson.
автотехномузыкадетское