Pattern Classification and Large Margin Classifiers

author: Peter L. Bartlett, UC Berkeley
published: Feb. 25, 2007,   recorded: August 2006,   views: 1394
Categories

Slides

Related content

Report a problem or upload files

If you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Lecture popularity: You need to login to cast your vote.
  Delicious Bibliography

 Watch videos:   (click on thumbnail to launch)

Watch Part 1
Part 1 1:17:06
!NOW PLAYING
Watch Part 2
Part 2 1:17:53
!NOW PLAYING
Watch Part 3
Part 3 1:19:53
!NOW PLAYING
Watch Part 4
Part 4 1:14:48
!NOW PLAYING

Description

These lectures will provide an introduction to the theory of pattern classification methods. They will focus on relationships between the minimax performance of a learning system and its complexity. There will be four lectures. The first will review the formulation of the pattern classification problem, and several popular pattern classification methods, and present general risk bounds in terms of Rademacher averages, a measure of the complexity of a class of functions. The second lecture will consider pattern classification in a minimax setting, and show that, in this setting, the Vapnik-Chervonenkis dimension is the key measure of complexity. The third lecture will focus on a theme of computational complexity. It will present the elegant relationship between the complexity of a class, as measured by its VC-dimension, and the computational complexity of functions from the class. This lecture will also review general results on the computational complexity of the pattern classification problem, and its tight relationship with that of an associated empirical risk optimization problems. The fourth lecture will consider large margin classification methods, such as AdaBoost, support vector machines, and neural networks, viewing them as convex relaxations of intractable empirical minimization problems. It will review several statistical properties of these large margin methods, in particular, a characterization of the convex optimization problems that lead to accurate classifiers, and relationships between these methods and probability models.

See Also:

Download slides icon Download slides: mlss06tw_bartlett_pclmc.pdf (327.3┬áKB)


Help icon Streaming Video Help

Link this page

Would you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !

Write your own review or comment:

make sure you have javascript enabled or clear this field: