SlideShare une entreprise Scribd logo
1  sur  52
Télécharger pour lire hors ligne
Robert Collins
CSE486, Penn State



                CSE/EE 486: Computer Vision I
                  Fall 2007 Course Overview
Robert Collins
CSE486, Penn State
                                            Textbook
                required                                         optional




     Introductory Techniques for 3-D Computer Vision   Computer Vision: A Modern Approach, by
     by E. Trucco and A. Verri, Prentice Hall, 1998.   D.Forsyth and J.Ponce, Prentice Hall, 2002.
Robert Collins
CSE486, Penn State           Fall 2007 Syllabus (excerpts)




      See full syllabus posted on Angel course web site
               https://cms.psu.edu/
Robert Collins
CSE486, Penn State
                          Programming Groups
 excerpts from syllabus
Robert Collins
CSE486, Penn State
                     Tentative Schedule   tentative
Robert Collins
CSE486, Penn State




                     Lecture 1: Introduction
                      to Computer Vision

              Readings T&V: Section 1, Section 2.1-2.2
Robert Collins
CSE486, Penn State

                         What is Vision?


            “Vision is the act of knowing what is
                where by looking.” --Aristotle



       Special emphasis: relationship between 3D world
       and a 2D image. Location and identity of objects.
Robert Collins
CSE486, Penn State

                        What is Computer Vision?


            It is related to, but not equivalent to:

                     • Photogrammetry
                     • Image Processing
                     • Artificial Intelligence
Robert Collins
CSE486, Penn State

                     Why study Computer Vision?

         • Images and movies are everywhere
         • Fast-growing collection of useful applications
              –   building representations of the 3D world from pictures
              –   automated surveillance (who’s doing what)
              –   movie post-processing
              –   face finding
         • Various deep and attractive scientific mysteries
              – how does object recognition work?
         • Greater understanding of human vision
Robert Collins
CSE486, Penn State

                     Course Goals and Objectives
     • Introduce the fundamental problems of computer
       vision.

     • Introduce the main concepts and techniques used to
       solve those problems.

     • Enable students to implement vision algorithms

     • Enable students to make sense of the vision literature
Robert Collins
CSE486, Penn State
                     Input: Digital Images
    2D arrays (matrices) of numbers:




      If color image, we have 3 arrays - red, green, blue
Robert Collins
CSE486, Penn State
                     Why is Computer Vision Hard?

     We are trying to infer things about the world
     from an array of numbers




                                                     Shoulder
                                                     of a cow…




                                 problems: too local; lack of context;
                                 mismatch between levels of abstraction.
                                 But wait, it’s even worse than that...
Robert Collins
CSE486, Penn State
                     Why is Computer Vision Hard?

     If we already know the geometry, surface material
     and lighting conditions, it is well-understood how to
     generate the value at each pixel. [this is Computer
     Graphics]

     But this confluence of factors contributing to each
     pixel can not be easily decomposed. The process
     can not be inverted.
Robert Collins
CSE486, Penn State




   tell me
   what you
   see...
Robert Collins
CSE486, Penn State




            Congratulations! You just did
         something mathematically impossible.

                                   How?
                       You used assumptions based on
                     prior knowledge / experience about
                          the way the world works.
Robert Collins
CSE486, Penn State   Recovering 3D from a single image is
                     a mathematically ill-posed problem.




        So we can’t solve the problem using only math.
Robert Collins
CSE486, Penn State


                 Good news: with more than one
                  camera, we can recover 3D!




                     Example: Stereo Vision
Robert Collins
CSE486, Penn State

                      Structure from Motion
                 We can also infer 3D from only one camera,
                  provided we move it around “enough”.

                         t-3              t-2
              t-4

                                                       Tx
                                  pitch
                                                                      t+2
                          t-1                         Ty
                                roll            yaw              Tz



                                          t                t+1
Robert Collins
CSE486, Penn State

                     Structure from Motion
             So how much is “enough” motion? It depends.




                                         www.grand-illusions.com

(play movie in external player)
Robert Collins
CSE486, Penn State
                     More Difficulties
       Object appearance changes with respect to viewpoint
Robert Collins
CSE486, Penn State
                     Effects of Lighting




                          Object appearance also
                          varies with respect to lighting
                          magnitude and direction
Robert Collins
CSE486, Penn State

                       Photometry Overview
    Source emits photons
                                                          And then some reach
                                                          an eye/camera and
                                                          are measured.


                     Photons travel in a
                     straight line




                                           They hit an object. Some are
                                           absorbed, some bounce off
                                           in a new direction.
Robert Collins
CSE486, Penn State

                            Light Transport
    Source emits photons
                                                          And then some reach
                     Illumination                         an eye/camera and
                                                          are measured.


                     Photons travel in a
                     straight line




                                           They hit an object. Some are
                                           absorbed, some bounce off
                                           in a new direction.
Robert Collins
CSE486, Penn State

                            Light Transport
    Source emits photons
                                                          And then some reach
                                                          an eye/camera and
                                                          are measured.


                     Photons travel in a
                     straight line




      Surface
    Reflection                             They hit an object. Some are
                                           absorbed, some bounce off
                                           in a new direction.
Robert Collins
CSE486, Penn State

                           Light Transport
    Source emits photons
                                                And then some reach
                                                an eye/camera and
                                                are measured.


                              Sensor Response




                                 They hit an object. Some are
                                 absorbed, some bounce off
                                 in a new direction.
Robert Collins
CSE486, Penn State

                     What do we mean by “red”
     Color percepts are a composition of three factors
       (illumination, surface reflectance, sensor response)

     We can’t easily factor the color we see in the image to
      infer illumination and material (even if sensor
      properties are fixed and known).


       Again, this is counterintuitive to human experience.
Robert Collins
CSE486, Penn State

                     Perception: Color Constancy
           Humans are very good at recognizing the same material
           colors under different illumination. Not clear how this
           is achieved in the general case.




                      Again, math won’t save us.
Robert Collins
CSE486, Penn State

                     What do we mean by “red”
     Color percepts are a composition of three factors
       (illumination, surface reflectance, sensor response)

     Some things to think about:
     •   “red” typically means “appears red to a human observer under white light”.
     •   white objects appear red under a red light.
     •   nothing looks red if you are red/green color blind.




                        “normal” color perception   red/green color blind
Robert Collins
CSE486, Penn State



 perceived as
 a 3D scene




                     but really just a planar
                     surface (screen) under
                     nonuniform lighting
                     (projector).
Robert Collins
CSE486, Penn State




                     What we will be studying
                        in thus course...

                     with a few examples of why.
Robert Collins
CSE486, Penn State

                       Filtering and Smoothing




                     Linear operators
                     Convolution
                     Smoothing
Robert Collins
CSE486, Penn State

                     Feature Extraction




           Corners
                                            Edges


      Image derivatives
                               Why?
      Gradient operators
                               Seek more unique descriptors
      DoG/LoG operators
                               (than pixels) for matching
      Harris corner detector
Robert Collins
CSE486, Penn State

                           Color and Light




               Radiance / Reflection
               Illumination / Shading
               Chromaticity
               Color Constancy
Robert Collins
CSE486, Penn State

                     Application : Skin Detection




                        Skin detection has been used for web
                     filtering based on identifying adult content
Robert Collins
CSE486, Penn State

                     Camera Projection Models
                                                W
                 Y           y
         X           Z   x

                                                    V
                                            U

              Projection Models
              Intrinsic (lens) Paramters
              Extrinsic (pose) Parameters
              Camera Calibration
Robert Collins
CSE486, Penn State
                              Application: Eyevision System
                               CAMERA CALIBRATION!
       Master pan/tilt                               Slave pan/tilt
       zoom/focus                          Control   zoom/focus




  Human
  operator
                         Virtual surface



                                                                       30 Cameras

                         Playing field




                                                             Stadium           Control Room
Robert Collins
CSE486, Penn State

                     Eyevision : SuperBowl XXXV
                            January 28, 2001
Robert Collins
CSE486, Penn State

                     EyeVision Examples




                     Super Bowl XXXV
                     January 28, 2001
                     Courtesy of CBS Sports
Robert Collins
CSE486, Penn State

                     Plane to Plane Mappings




            Rigid, Similarity, Affine, & Projective Mappings
            Homography Estimation
            Image Warping
Robert Collins
CSE486, Penn State

                      Mosaicing and Stabilization




                     Camera Rotation Homography
                     Mosaicing
                     Video Stabilization
Robert Collins
CSE486, Penn State

                     Example : Quicktime VR




                           http://www.panoguide.com/gallery/
Robert Collins
CSE486, Penn State

                                 Stereo Vision




                     Stereo Camera Setups
                     Stereo Disparity / Parallax
                     Epipolar Geometry
                     Correspondence Matching
                     Triangulation / Depth Recovery
Robert Collins
CSE486, Penn State

                              Camera Motion




                     Motion Field vs Optic Flow
                     Flow Estimation
                     Egomotion Estimation
                     Structure from Motion
Robert Collins
CSE486, Penn State

                     Application: Match Move
   Track a set of feature points through a movie sequence

   Deduce where the cameras are
   and the 3D locations of the
   points that were tracked

   Render synthetic objects
   with respect to the
   deduced 3D geometry
   of the scene / cameras


                              “Graham Kimpton” example from www.realviz.com
                              MatchMover Professional gallery. Copyrighted.
Robert Collins
CSE486, Penn State

               Application: Autonomous Driving
                     Have your vehicle chauffeur you around.




                                  Stanley, winner 2005 Darpa Grand Challenge
Robert Collins
CSE486, Penn State

                       Video Change Detection




                     Video Sequences
                     Background Modeling
                     Change Detection
Robert Collins
CSE486, Penn State

                     Application: Video Surveillance



         Automatic detection, classification and tracking of moving people and vehicles




           Object locations are determined with
             respect to a 3D campus model




         A single operator can monitor results from
           many sensors spread over a large area
Robert Collins
CSE486, Penn State
                         Automated Surveillance
                 Trigger regions for detecting motion and detecting motion going in
                 the wrong direction. These are pretty well “solved” problems.




Guardian Systems
Robert Collins
CSE486, Penn State

                           Video Tracking




                     Appearance-Based Tracking
                     Sample Tracking Algorithms
                       (e.g. Mean-Shift, Lucas-Kanade)
Robert Collins
CSE486, Penn State

                         Object Recognition
                                         Library of models




                     Model matches (using Lowe’s SIFT keys)

  Approaches: PCA; Sift Keys; boosted cascade of detectors
Robert Collins
CSE486, Penn State

                     Application: Face/Eye Detection
                                           Henry Schneiderman




       Finding users to interact with.
       Red-eye removal from photos.
       Drowsy-driver detection.
Robert Collins
CSE486, Penn State

                              Reminder...


                 “Vision is the act of knowing what is
                     where by looking.” --Aristotle

Contenu connexe

Plus de zukun

My lyn tutorial 2009
My lyn tutorial 2009My lyn tutorial 2009
My lyn tutorial 2009zukun
 
ETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCVETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCVzukun
 
ETHZ CV2012: Information
ETHZ CV2012: InformationETHZ CV2012: Information
ETHZ CV2012: Informationzukun
 
Siwei lyu: natural image statistics
Siwei lyu: natural image statisticsSiwei lyu: natural image statistics
Siwei lyu: natural image statisticszukun
 
Lecture9 camera calibration
Lecture9 camera calibrationLecture9 camera calibration
Lecture9 camera calibrationzukun
 
Brunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer visionBrunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer visionzukun
 
Modern features-part-4-evaluation
Modern features-part-4-evaluationModern features-part-4-evaluation
Modern features-part-4-evaluationzukun
 
Modern features-part-3-software
Modern features-part-3-softwareModern features-part-3-software
Modern features-part-3-softwarezukun
 
Modern features-part-2-descriptors
Modern features-part-2-descriptorsModern features-part-2-descriptors
Modern features-part-2-descriptorszukun
 
Modern features-part-1-detectors
Modern features-part-1-detectorsModern features-part-1-detectors
Modern features-part-1-detectorszukun
 
Modern features-part-0-intro
Modern features-part-0-introModern features-part-0-intro
Modern features-part-0-introzukun
 
Lecture 02 internet video search
Lecture 02 internet video searchLecture 02 internet video search
Lecture 02 internet video searchzukun
 
Lecture 01 internet video search
Lecture 01 internet video searchLecture 01 internet video search
Lecture 01 internet video searchzukun
 
Lecture 03 internet video search
Lecture 03 internet video searchLecture 03 internet video search
Lecture 03 internet video searchzukun
 
Icml2012 tutorial representation_learning
Icml2012 tutorial representation_learningIcml2012 tutorial representation_learning
Icml2012 tutorial representation_learningzukun
 
Advances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer visionAdvances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer visionzukun
 
Gephi tutorial: quick start
Gephi tutorial: quick startGephi tutorial: quick start
Gephi tutorial: quick startzukun
 
EM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysisEM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysiszukun
 
Object recognition with pictorial structures
Object recognition with pictorial structuresObject recognition with pictorial structures
Object recognition with pictorial structureszukun
 
Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities zukun
 

Plus de zukun (20)

My lyn tutorial 2009
My lyn tutorial 2009My lyn tutorial 2009
My lyn tutorial 2009
 
ETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCVETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCV
 
ETHZ CV2012: Information
ETHZ CV2012: InformationETHZ CV2012: Information
ETHZ CV2012: Information
 
Siwei lyu: natural image statistics
Siwei lyu: natural image statisticsSiwei lyu: natural image statistics
Siwei lyu: natural image statistics
 
Lecture9 camera calibration
Lecture9 camera calibrationLecture9 camera calibration
Lecture9 camera calibration
 
Brunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer visionBrunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer vision
 
Modern features-part-4-evaluation
Modern features-part-4-evaluationModern features-part-4-evaluation
Modern features-part-4-evaluation
 
Modern features-part-3-software
Modern features-part-3-softwareModern features-part-3-software
Modern features-part-3-software
 
Modern features-part-2-descriptors
Modern features-part-2-descriptorsModern features-part-2-descriptors
Modern features-part-2-descriptors
 
Modern features-part-1-detectors
Modern features-part-1-detectorsModern features-part-1-detectors
Modern features-part-1-detectors
 
Modern features-part-0-intro
Modern features-part-0-introModern features-part-0-intro
Modern features-part-0-intro
 
Lecture 02 internet video search
Lecture 02 internet video searchLecture 02 internet video search
Lecture 02 internet video search
 
Lecture 01 internet video search
Lecture 01 internet video searchLecture 01 internet video search
Lecture 01 internet video search
 
Lecture 03 internet video search
Lecture 03 internet video searchLecture 03 internet video search
Lecture 03 internet video search
 
Icml2012 tutorial representation_learning
Icml2012 tutorial representation_learningIcml2012 tutorial representation_learning
Icml2012 tutorial representation_learning
 
Advances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer visionAdvances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer vision
 
Gephi tutorial: quick start
Gephi tutorial: quick startGephi tutorial: quick start
Gephi tutorial: quick start
 
EM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysisEM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysis
 
Object recognition with pictorial structures
Object recognition with pictorial structuresObject recognition with pictorial structures
Object recognition with pictorial structures
 
Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities
 

Lecture01

  • 1. Robert Collins CSE486, Penn State CSE/EE 486: Computer Vision I Fall 2007 Course Overview
  • 2. Robert Collins CSE486, Penn State Textbook required optional Introductory Techniques for 3-D Computer Vision Computer Vision: A Modern Approach, by by E. Trucco and A. Verri, Prentice Hall, 1998. D.Forsyth and J.Ponce, Prentice Hall, 2002.
  • 3. Robert Collins CSE486, Penn State Fall 2007 Syllabus (excerpts) See full syllabus posted on Angel course web site https://cms.psu.edu/
  • 4. Robert Collins CSE486, Penn State Programming Groups excerpts from syllabus
  • 5. Robert Collins CSE486, Penn State Tentative Schedule tentative
  • 6. Robert Collins CSE486, Penn State Lecture 1: Introduction to Computer Vision Readings T&V: Section 1, Section 2.1-2.2
  • 7. Robert Collins CSE486, Penn State What is Vision? “Vision is the act of knowing what is where by looking.” --Aristotle Special emphasis: relationship between 3D world and a 2D image. Location and identity of objects.
  • 8. Robert Collins CSE486, Penn State What is Computer Vision? It is related to, but not equivalent to: • Photogrammetry • Image Processing • Artificial Intelligence
  • 9. Robert Collins CSE486, Penn State Why study Computer Vision? • Images and movies are everywhere • Fast-growing collection of useful applications – building representations of the 3D world from pictures – automated surveillance (who’s doing what) – movie post-processing – face finding • Various deep and attractive scientific mysteries – how does object recognition work? • Greater understanding of human vision
  • 10. Robert Collins CSE486, Penn State Course Goals and Objectives • Introduce the fundamental problems of computer vision. • Introduce the main concepts and techniques used to solve those problems. • Enable students to implement vision algorithms • Enable students to make sense of the vision literature
  • 11. Robert Collins CSE486, Penn State Input: Digital Images 2D arrays (matrices) of numbers: If color image, we have 3 arrays - red, green, blue
  • 12. Robert Collins CSE486, Penn State Why is Computer Vision Hard? We are trying to infer things about the world from an array of numbers Shoulder of a cow… problems: too local; lack of context; mismatch between levels of abstraction. But wait, it’s even worse than that...
  • 13. Robert Collins CSE486, Penn State Why is Computer Vision Hard? If we already know the geometry, surface material and lighting conditions, it is well-understood how to generate the value at each pixel. [this is Computer Graphics] But this confluence of factors contributing to each pixel can not be easily decomposed. The process can not be inverted.
  • 14. Robert Collins CSE486, Penn State tell me what you see...
  • 15. Robert Collins CSE486, Penn State Congratulations! You just did something mathematically impossible. How? You used assumptions based on prior knowledge / experience about the way the world works.
  • 16. Robert Collins CSE486, Penn State Recovering 3D from a single image is a mathematically ill-posed problem. So we can’t solve the problem using only math.
  • 17. Robert Collins CSE486, Penn State Good news: with more than one camera, we can recover 3D! Example: Stereo Vision
  • 18. Robert Collins CSE486, Penn State Structure from Motion We can also infer 3D from only one camera, provided we move it around “enough”. t-3 t-2 t-4 Tx pitch t+2 t-1 Ty roll yaw Tz t t+1
  • 19. Robert Collins CSE486, Penn State Structure from Motion So how much is “enough” motion? It depends. www.grand-illusions.com (play movie in external player)
  • 20. Robert Collins CSE486, Penn State More Difficulties Object appearance changes with respect to viewpoint
  • 21. Robert Collins CSE486, Penn State Effects of Lighting Object appearance also varies with respect to lighting magnitude and direction
  • 22. Robert Collins CSE486, Penn State Photometry Overview Source emits photons And then some reach an eye/camera and are measured. Photons travel in a straight line They hit an object. Some are absorbed, some bounce off in a new direction.
  • 23. Robert Collins CSE486, Penn State Light Transport Source emits photons And then some reach Illumination an eye/camera and are measured. Photons travel in a straight line They hit an object. Some are absorbed, some bounce off in a new direction.
  • 24. Robert Collins CSE486, Penn State Light Transport Source emits photons And then some reach an eye/camera and are measured. Photons travel in a straight line Surface Reflection They hit an object. Some are absorbed, some bounce off in a new direction.
  • 25. Robert Collins CSE486, Penn State Light Transport Source emits photons And then some reach an eye/camera and are measured. Sensor Response They hit an object. Some are absorbed, some bounce off in a new direction.
  • 26. Robert Collins CSE486, Penn State What do we mean by “red” Color percepts are a composition of three factors (illumination, surface reflectance, sensor response) We can’t easily factor the color we see in the image to infer illumination and material (even if sensor properties are fixed and known). Again, this is counterintuitive to human experience.
  • 27. Robert Collins CSE486, Penn State Perception: Color Constancy Humans are very good at recognizing the same material colors under different illumination. Not clear how this is achieved in the general case. Again, math won’t save us.
  • 28. Robert Collins CSE486, Penn State What do we mean by “red” Color percepts are a composition of three factors (illumination, surface reflectance, sensor response) Some things to think about: • “red” typically means “appears red to a human observer under white light”. • white objects appear red under a red light. • nothing looks red if you are red/green color blind. “normal” color perception red/green color blind
  • 29. Robert Collins CSE486, Penn State perceived as a 3D scene but really just a planar surface (screen) under nonuniform lighting (projector).
  • 30. Robert Collins CSE486, Penn State What we will be studying in thus course... with a few examples of why.
  • 31. Robert Collins CSE486, Penn State Filtering and Smoothing Linear operators Convolution Smoothing
  • 32. Robert Collins CSE486, Penn State Feature Extraction Corners Edges Image derivatives Why? Gradient operators Seek more unique descriptors DoG/LoG operators (than pixels) for matching Harris corner detector
  • 33. Robert Collins CSE486, Penn State Color and Light Radiance / Reflection Illumination / Shading Chromaticity Color Constancy
  • 34. Robert Collins CSE486, Penn State Application : Skin Detection Skin detection has been used for web filtering based on identifying adult content
  • 35. Robert Collins CSE486, Penn State Camera Projection Models W Y y X Z x V U Projection Models Intrinsic (lens) Paramters Extrinsic (pose) Parameters Camera Calibration
  • 36. Robert Collins CSE486, Penn State Application: Eyevision System CAMERA CALIBRATION! Master pan/tilt Slave pan/tilt zoom/focus Control zoom/focus Human operator Virtual surface 30 Cameras Playing field Stadium Control Room
  • 37. Robert Collins CSE486, Penn State Eyevision : SuperBowl XXXV January 28, 2001
  • 38. Robert Collins CSE486, Penn State EyeVision Examples Super Bowl XXXV January 28, 2001 Courtesy of CBS Sports
  • 39. Robert Collins CSE486, Penn State Plane to Plane Mappings Rigid, Similarity, Affine, & Projective Mappings Homography Estimation Image Warping
  • 40. Robert Collins CSE486, Penn State Mosaicing and Stabilization Camera Rotation Homography Mosaicing Video Stabilization
  • 41. Robert Collins CSE486, Penn State Example : Quicktime VR http://www.panoguide.com/gallery/
  • 42. Robert Collins CSE486, Penn State Stereo Vision Stereo Camera Setups Stereo Disparity / Parallax Epipolar Geometry Correspondence Matching Triangulation / Depth Recovery
  • 43. Robert Collins CSE486, Penn State Camera Motion Motion Field vs Optic Flow Flow Estimation Egomotion Estimation Structure from Motion
  • 44. Robert Collins CSE486, Penn State Application: Match Move Track a set of feature points through a movie sequence Deduce where the cameras are and the 3D locations of the points that were tracked Render synthetic objects with respect to the deduced 3D geometry of the scene / cameras “Graham Kimpton” example from www.realviz.com MatchMover Professional gallery. Copyrighted.
  • 45. Robert Collins CSE486, Penn State Application: Autonomous Driving Have your vehicle chauffeur you around. Stanley, winner 2005 Darpa Grand Challenge
  • 46. Robert Collins CSE486, Penn State Video Change Detection Video Sequences Background Modeling Change Detection
  • 47. Robert Collins CSE486, Penn State Application: Video Surveillance Automatic detection, classification and tracking of moving people and vehicles Object locations are determined with respect to a 3D campus model A single operator can monitor results from many sensors spread over a large area
  • 48. Robert Collins CSE486, Penn State Automated Surveillance Trigger regions for detecting motion and detecting motion going in the wrong direction. These are pretty well “solved” problems. Guardian Systems
  • 49. Robert Collins CSE486, Penn State Video Tracking Appearance-Based Tracking Sample Tracking Algorithms (e.g. Mean-Shift, Lucas-Kanade)
  • 50. Robert Collins CSE486, Penn State Object Recognition Library of models Model matches (using Lowe’s SIFT keys) Approaches: PCA; Sift Keys; boosted cascade of detectors
  • 51. Robert Collins CSE486, Penn State Application: Face/Eye Detection Henry Schneiderman Finding users to interact with. Red-eye removal from photos. Drowsy-driver detection.
  • 52. Robert Collins CSE486, Penn State Reminder... “Vision is the act of knowing what is where by looking.” --Aristotle