Machine Learning: Overview - PowerPoint PPT Presentation

About This Presentation
Title:

Machine Learning: Overview

Description:

Title: Machine Learning: Overview Author: James Hays Last modified by: James Hays Created Date: 8/16/2006 12:00:00 AM Document presentation format – PowerPoint PPT presentation

Number of Views:143
Avg rating:3.0/5.0
Slides: 21
Provided by: JamesH170
Learn more at: https://cs.brown.edu
Category:

less

Transcript and Presenter's Notes

Title: Machine Learning: Overview


1
The blue and green colors are actually the same
2
http//blogs.discovermagazine.com/badastronomy/200
9/06/24/the-blue-and-the-green/
3
Machine Learning Overview
09/19/11
  • Computer Vision
  • James Hays, Brown

Slides Isabelle Guyon, Erik Sudderth, Mark
Johnson,Derek Hoiem
Photo CMU Machine Learning Department protests
G20
4
Machine learning Overview
  • Core of ML Making predictions or decisions from
    Data.
  • This overview will not go in to depth about the
    statistical underpinnings of learning methods.
    Were looking at ML as a tool. Take CSCI 1950-F
    Introduction to Machine Learning to learn more
    about ML.

5
Impact of Machine Learning
  • Machine Learning is arguably the greatest export
    from computing to other scientific fields.

6
Machine Learning Applications
Slide Isabelle Guyon
7
Image Categorization
Training Labels
Training
Classifier Training
Image Features
Trained Classifier
Slide Derek Hoiem
8
Image Categorization
Training Labels
Training
Classifier Training
Image Features
Trained Classifier
Testing
Prediction
Image Features
Trained Classifier
Outdoor
Test Image
Slide Derek Hoiem
9
  • Claim
  • The decision to use machine learning is more
    important than the choice of a particular
    learning method.
  • If you hear somebody talking of a specific
    learning mechanism, be wary (e.g. YouTube comment
    "Oooh, we could plug this in to a Neural
    networkand blah blah blah)

10
Example Boundary Detection
  • Is this a boundary?

11
Image features
Training Labels
Training
Classifier Training
Image Features
Trained Classifier
Slide Derek Hoiem
12
General Principles of Representation
  • Coverage
  • Ensure that all relevant info is captured
  • Concision
  • Minimize number of features without sacrificing
    coverage
  • Directness
  • Ideal features are independently useful for
    prediction

Image Intensity
Slide Derek Hoiem
13
Image representations
  • Templates
  • Intensity, gradients, etc.
  • Histograms
  • Color, texture, SIFT descriptors, etc.

Slide Derek Hoiem
14
Classifiers
Training Labels
Training
Classifier Training
Image Features
Trained Classifier
Slide Derek Hoiem
15
Learning a classifier
  • Given some set of features with corresponding
    labels, learn a function to predict the labels
    from the features

Slide Derek Hoiem
16
One way to think about it
  • Training labels dictate that two examples are the
    same or different, in some sense
  • Features and distance measures define visual
    similarity
  • Classifiers try to learn weights or parameters
    for features and distance measures so that visual
    similarity predicts label similarity

Slide Derek Hoiem
17
Slide Erik Sudderth
18
Dimensionality Reduction
  • PCA, ICA, LLE, Isomap
  • PCA is the most important technique to know.  It
    takes advantage of correlations in data
    dimensions to produce the best possible lower
    dimensional representation, according to
    reconstruction error.
  • PCA should be used for dimensionality reduction,
    not for discovering patterns or making
    predictions. Don't try to assign semantic meaning
    to the bases.

19
Many classifiers to choose from
  • SVM
  • Neural networks
  • Naïve Bayes
  • Bayesian network
  • Logistic regression
  • Randomized Forests
  • Boosted Decision Trees
  • K-nearest neighbor
  • RBMs
  • Etc.

Which is the best one?
Slide Derek Hoiem
20
Next Two Lectures
  • Friday we'll talk about clustering methods
    (k-means, mean shift) and their common usage in
    computer vision -- building "bag of words
    representations inspired by the NLP community.
    We'll be using these models for projects 2 and 3.
  • Monday we'll focus specifically on classification
    methods, e.g. nearest neighbor, naïve-Bayes,
    decision trees, linear SVM, Kernel methods. Well
    be using these for projects 3 and 4 (and
    optionally 2).
Write a Comment
User Comments (0)
About PowerShow.com