NIPS: Oral Session 2 - Xiangyu Wang

129
Опубликовано 18 августа 2016, 18:12
Median Selection Subset Aggregation for Parallel Inference ΓÇïFor massive data sets, efficient computation commonly relies on distributed algorithms that store and process subsets of the data on different machines, minimizing communication costs. Our focus is on regression and classification problems involving many features. A variety of distributed algorithms have been proposed in this context, but challenges arise in defining an algorithm with low communication, theoretical guarantees and excellent practical performance in general settings. We propose a MEdian Selection Subset AGgregation Estimator (message) algorithm, which attempts to solve these problems. The algorithm applies feature selection in parallel for each subset using Lasso or another method, calculates the `median' feature inclusion index, estimates coefficients for the selected features in parallel for each subset, and then averages these estimates. The algorithm is simple, involves very minimal communication, scales efficiently in both sample and feature size, and has theoretical guarantees. In particular, we show model selection consistency and coefficient estimation efficiency. Extensive experiments show excellent performance in variable selection, estimation, prediction, and computation time relative to usual competitors.
Свежие видео
5 дней – 40 78514:57
Ryzen Turbo Mode Tested On All Cpus
6 дней – 282 9790:16
Android x Wicked: Wickedly open style
12 дней – 87 5619:45
9800X3D vs. tiny cooler.....
164 дня – 5 16711:48
Let’s Talk about Apple Intelligence
Случайные видео
356 дней – 4 4690:50
Growth: Custom store listings #shorts
26.12.21 – 3 687 75717:03
Building the ULTIMATE Racing Setup
22.04.08 – 89 67610:22
Google Ajax Apis Go Restful
30 дней – 997 4890:55
Blast the BASS💦
автотехномузыкадетское