Authors:
Nery Riquelme-Granada
;
Khuong An Nguyen
and
Zhiyuan Luo
Affiliation:
Department of Computer Science, Royal Holloway University of London, Egham, Surrey, TW20 0EX, U.K.
Keyword(s):
Coresets, Data Summaries, Logistic Regression, Large-data, Computing Time.
Abstract:
In the era of datasets of unprecedented sizes, data compression techniques are an attractive approach for speeding up machine learning algorithms. One of the most successful paradigms for achieving good-quality compression is that of coresets: small summaries of data that act as proxies to the original input data. Even though coresets proved to be extremely useful to accelerate unsupervised learning problems, applying them to supervised learning problems may bring unexpected computational bottlenecks. We show that this is the case for Logistic Regression classification, and hence propose two methods for accelerating the computation of coresets for this problem. When coresets are computed using our methods on three public datasets, computing the coreset and learning from it is, in the worst case, 11 times faster than learning directly from the full input data, and 34 times faster in the best case. Furthermore, our results indicate that our accelerating approaches do not degrade the em
pirical performance of coresets.
(More)