TY  - CONF
TI  - Optimal Rates for the Random Fourier Feature Technique
UR  - http://www.gatsby.ucl.ac.uk/~szabo/talks/invited_talk/Zoltan_Szabo_invited_talk_Ecole_Polytechnique_14_03_2016_abstract.pdf
T2  - invited talk at École Polytechnique
AV  - public
ID  - discovery1476385
A1  - Szabo, Z
M2  - Palaiseau, France
Y1  - 2016/03/14/
N2  - Kernel methods represent one of the most powerful tools in machine learning to tackle problems expressed in terms of function values and derivatives. While these methods show good versatility, they are computationally intensive and have poor scalability to large data as they require operations on Gram matrices. In order to mitigate this serious computational limitation, recently randomized methods have been proposed in the literature, which allow the application of fast linear algorithms. Random Fourier features (RFF) are among the most popular and widely applied constructions: they provide an easily computable, low-dimensional feature representation for shift-invariant kernels. Despite the popularity of RFFs, very little is understood theoretically about their approximation quality. In this talk, I am going to present the main ideas and results of a detailed finite-sample theoretical analysis about the approximation quality of RFFs by (i) establishing optimal (in terms of the RFF dimension, and growing set size) performance guarantees in uniform norm, and (ii) providing guarantees in Lp (1 ? p < ?) norms. [Joint work with Bharath K. Sriperumbudur]
ER  -