High-Dimensional Non-Linear Variable Selection through Hierarchical Kernel Learning

author: Francis R. Bach, INRIA - SIERRA project-team
published: May 6, 2009,   recorded: April 2009,   views: 4358
Categories

See Also:

Download slides icon Download slides: smls09_bach_hdnlv_01.pdf (680.2 KB)


Help icon Streaming Video Help

Related content

Report a problem or upload files

If you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Lecture popularity: You need to login to cast your vote.
  Delicious Bibliography

Description

We consider the problem of high-dimensional non-linear variable selection for supervised learning. Our approach is based on performing linear selection among exponentially many well-defined groups of features or positive definite kernels, that characterize non-linear interactions between the original variables. To select efficiently from these many kernels, we use the natural hierarchical structure of the kernels to extend the multiple kernel learning framework to kernels that can be embedded in a directed acyclic graph; we show that it is then possible to perform kernel selection through a graph-adapted sparsity-inducing norm, in polynomial time in the number of selected kernels. Moreover, we study the consistency of variable selection in high-dimensional settings, showing that under certain assumptions, our regularization framework allows a number of irrelevant variables which is sub-exponential in the number of observations.

Link this page

Would you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !

Write your own review or comment:

make sure you have javascript enabled or clear this field: