SlideShare une entreprise Scribd logo
1  sur  36
Depth & Space
Becca Kennedy
Perception in Real & Virtual Environments
9/19/12
Overview
• Edges, lines, and texture elements must be interpreted in
terms of 3D structure to understand the world
• Observer must determine:
▫ Depth – distance of the surface from the observer
▫ Surface orientation – slant and tilt
• Depth and surface orientation are recovered together
▫ 3D orientation determines distances of object parts
from the observer, and distance of parts determines
3D orientation
Overview
• Slant – size of the angle between the observer’s line of sight
and the surface normal
▫ Surface normal – virtual line sticking out perpendicularly
out of the surface at that point
• Tilt – the direction of the depth gradient relative to the
frontal plane
The Problem of Depth Perception
• Depth perception from a 2-D retinal image is ambiguous
The Problem of Depth Perception –
Heuristic Assumptions
• Visual system implicitly makes heuristic assumptions
about the nature of the world
• Our visual system is fooled by 3-D movies
▫ Visual system implicitly assumes that both eyes are
looking at the same scene
▫ The different image presented to each eye is
interpreted as depth
▫ But usually this heuristic is correct
The Problem of Depth Perception –
Marr’s 2.5-D Sketch
• There are many independent processing modules
computing depth information from separate sources
▫ Each module processes different kinds of information
• The final common depth interpretation is expressed as a
2.5-D sketch
Sources of Depth Information
• Ocular information vs. optical information
▫ Ocular information arises from factors that depend on the state of
the eyes themselves
▫ Optical information arises from the structure of the light entering
the eyes
• Binocular information vs. monocular information
• Static information vs. dynamic information
Ocular Information
• Accommodation – the process through which the ciliary
muscles in the eye control the optical focus of the lens by
temporarily changing its shape
▫ Monocular cue
▫ Thick lens for close objects, thin lens for far objects
▫ Weak source of depth information, but used at close distances
Ocular Information
• Convergence – the extent to which the two eyes are
turned inward to fixate an object
▫ Binocular cue
▫ Fixating on a close object results in a large convergence angle
▫ Fixating on a far object results in a small convergence angle
▫ Visual system uses the angle of eye convergence to determine
distance to the fixated point
Stereoscopic Information
• Stereopsis is the process of perceiving the relative
distance to objects based on their lateral displacement in
the two retinal images
▫ This relative lateral displacement is binocular disparity
 Direction of binocular disparity provides info about which points are closer
and which are farther than the fixated point
 Magnitude of binocular disparity provides information about how much
closer or farther they are
• Specifies ratios of distances to objects rather than simply
which is farther and which is closer
Corresponding Retinal Positions
• Corresponding positions on the two retinae are positions
that would coincide if the two foveae were superimposed
by simple lateral displacement
▫ Binocular disparity occurs when a given point in the external world
doesn’t project to corresponding positions
 Crossed disparity indicate that a point is closer than the fixated point
 Uncrossed disparity indicates that a point is farther away than the fixated point
Corresponding Retinal Positions
• The horopter is the set of environmental points that
stimulate corresponding points on the two retinae
▫ Theoretical horopter –defined as the locus of points which make
the same angle at the eyes
▫ Empirical horopter –defined by singleness of vision; larger than
theoretical horopter
• Panum’s fusional area is the area around the horopter
within which disparate images are perceptually fused, so
we don’t see double images
▫ Points that lie outside Panum’s area create disparity that we
experience as depth
The Correspondence Problem
• How does the visual system determine which features in
one retinal image correspond to which features in the
other?
• For many years, theorists assumed that this problem was
solved by a shape analysis for each left and right image
that occurred before stereopsis
The Correspondence Problem
• The alternative possibility is that stereopsis occurs first
▫ Random dot stereograms
 When each image is viewed alone, the dots look random
 Shape-first theory would predict that depth perception of random-dot
images stereoscopically would be impossible
 Random dot stereograms show that stereoscopic depth can be
perceived without monocular shape information
Computational Theories
• Most dots in the left image have a corresponding dot in the
right image
▫ The visual system needs to figure out which pairs of dots go together
• The first Marr-Poggio algorithm (1976, 1977)
▫ Individual pixels in the left and right images are matched according to
location and color
 Among these matches, there are the correct ones that correspond to the visible
portions of the actual surfaces in the real world
▫ Two heuristic constraints help provide the correct solution
 Surface opacity – only the nearest surface can be seen
 Surface continuity – correct solution will tend to be one in which matches are close
together in depth
Edge-Based Algorithms
• Marr and Poggio suggested a second algorithm in 1979
• Differed from the first in the following ways:
▫ Edge-based matching – matches edges in the left and right
images rather than pixels
▫ Multiple scales – visual system first looks for corresponding
edges at a large spatial scale, followed by more detailed matching
at finer-grained levels
▫ Single-pass operation – noniterative; finds best edge-based
correspondence in a single pass through a multistage operation
Multi-Orientation, Multi-Scale (MOMS) filters
• Jones and Malik (1990)
▫ A process of matching the vector representing a given position in
one eye to each of the vectors representing laterally displaced
positions in the other eye
 Specifies the most likely correspondence
▫ Better and more robust matches because MOMS vectors carry a
lot of spatial information
 Compared to outputs of single receptors or edge detectors
Physiology of Stereoscopic Vision
• Binocular depth cells
▫ Hubel and Wiesel (1962)
 Discovered cells in V1 of the visual cortex that were sensitive to binocular stimulation
▫ Barlow, Blakemore, and Pettigrew (1967)
 Reported that some binocular cells in area V1 responded optimally to stimulation in
disparate locations of the two retinae
• To show that these cells are involved in depth perception, need to
also demonstrate a connection between disparity and behavior
▫ Blake and Hirsch (1975)
 Reared cats so that their vision was alternated between left and right eyes for 6 months
 These cats had few binocular neurons and they were not able to use binocular disparity to
perceive depth
▫ Recent brain imaging experiments have shown that many different areas
are activated by stimuli that create binocular disparity
 Depth perception involves many stages of processing that extend from primary visual cortex
Environmental depth cues
Dynamic Information
• Motion parallax
▫ The differential motion of pairs of points due to their different depths
relative to the fixation point
 Nearby objects move quickly, far off objects appear stationary
• Optic flow caused by a moving observer
▫ Relative to the fixation point…
 Points closer to the observer flow in the direction opposite the observer’s motion
 Points farther than fixation point flow in the same direction as the observer’s motion
Dynamic Information
• Another pattern of optic flow is optic expansion or looming
▫ Fixated point is stationary on the retina
▫ Other points flow outward, faster with more distance from fixation
point
Dynamic Information
• Optic flow caused by moving objects
▫ Kinetic depth effect (KDE; Wallach & O’Connell, 1953) – ability to
perceive depth from object motion
▫ Visual system uses a rigidity heuristic
 Biased toward perceiving rigid motions rather than plastic motions
• Accretion/Deletion of Texture
▫ Appearance and disappearance of texture behind a moving edge
Pictorial Information
• Convergence of parallel lines
• Position relative to the horizon
Pictorial Information
• Relative size
• Familiar size
▫ In a VE, if not enough depth cues are present, the observer begins to
depend on retinal size (Kenyon, Sandin, Smith, Pawlicki, & Defanti, 2007)
• Texture gradients
Pictorial Information – Edge Interpretation
• Edge and contour interpretations
▫ E.g., occlusion or interposition – blocking of light from an object by an
opaque object causing occlusion or interposition
▫ Edges provide relative rather than absolute depth information
▫ Available from virtually unlimited distances within visible range
• Vertex (edge intersection) classification
▫ Guzman’s (1968, 1969) program SEE attempted to interpret line
drawings of simple configurations of blocks
 He developed a classification scheme for edge intersections (vertices): Ts, Ys, Ks, Xs,
Ls, etc.
▫ Huffman and Clowes (1971) developed a complete catalog of the vertex
types that arise in viewing simple trihedral angles from all possible
viewpoints
Pictorial Information – Edge Interpretation
• Four types of edges:
1. Orientation edges – places where there are discontinuities in
surface orientation; when two different orientations meet along
an edge
2. Depth edges – places where there is a spatial discontinuity in
depth between surfaces; places where one surface occludes
another that extends behind it, with space between
3. Illumination edges – places where there is a difference in the
amount of light falling on a homogenous surface; edge of a
shadow, highlight, or spotlight
4. Reflectance edges – places where there is a change in the
light-reflecting properties of the surface material; e.g., designs
painted on a surface
Pictorial Information – Edge Interpretation
• Edge labels
▫ Two kinds of orientation edges
 Convex orientation edges are labeled with a +
 Concave orientation edges are labeled with a -
▫ Arrows indicate that the closer surface is on the right
Pictorial Information – Edge Interpretation
Pictorial Information – Edge Interpretation
• Physical constraints
▫ Not all logically possible labelings are physically possible
Pictorial Information – Edge Interpretation
• Extensions and Generalizations
▫ Waltz (1975) extended the Huffman-Clowes analysis to include 11
types of edges, including shadows and “cracks” (orientation edges
at 180 degree angles)
 Adding shadows making interpretation more accurate because it
provides further constraints
▫ Malik (1987) extended analysis of edge labeling to curved
objects
 New depth edge type extremal edge or limb (double arrow) occurs
when a surface curves smoothly around to partly exclude itself
Pictorial Information – Edge Interpretation
• Extensions and Generalizations
▫ Barrow and Tennenbaum (1978)’s analysis contained additional
constraints:
 The smoothness assumption – if an occluding edge in the image is
smooth, then so is the contour of the surface that produced it
 The general viewpoint assumption – small changes in viewpoint will
not cause qualitative differences in the image
Pictorial Information
• Shading information
▫ Shading – variations in the amount of light reflected from the surface as
a result of variations in the orientation of the surface relative to a light
source
▫ Horn’s (1975, 1977) Computational Analysis
 Showed that percentage changes in image luminance are directly proportional to
percentage changes in the orientation of the surface
▫ Humans are able to interpret surfaces with significantly specular
characteristics, like glossy surfaces that reflect light more coherently
than matte surfaces do
 How?
Pictorial Information
• Shading information
▫ Cast shadows
 Shadows of one objet that fall on the surface of another object provide more
depth information
 Distance between object and its shadow cast on the surface gives height of its
bottom above the surface
Pictorial Information
• Aerial perspective
▫ Refers to certain systematic differences in the contrast and color of
objects that occur when they are viewed from great distances
 Contrast is reduced by additional atmosphere through which they are viewed, which
contains particles of dust, water, or pollutants that scatter light
 Mountains that are far away appear bluer because the atmosphere scatters longer
wavelengths of light more than shorter wavelengths
Integrating Information Sources
• Depth cues are often highly correlated, making them easy to integrate
• What happens when cues are in conflict with one another? 3 possibilities:
1. One source dominates a conflicting source
 E.g., In Ames room, perspective information dominates familiar size
2. A compromise is achieved between two conflicting sources
 Visual system makes independent estimates of depth from each source alone, then
integrates them according to a mathematical rule
 Bruno and Cutting (1988) found that information integration was
additive; sum independent effects of sources
3. The two sources interact to arrive at an optimal solution
 E.g., convergence specifies absolute depth, binocular disparity specifies ratios of
distances; together they can provide a complete depth map
Depth Perception and VEs
• Our visual system is really good at depth perception in real
environments, but this is hard to replicate in virtual scenes
▫ Ocular depth information (accommodation, convergence) is less useful
▫ Stereoscopic depth information may not be available
▫ Motion cues may not be faithfully represented
▫ Depth cues may be conflicting
▫ Etc.!
• But augmented reality can also improve real-world depth
perception

Contenu connexe

Tendances

Psychology of Perception & Attention
Psychology of Perception & Attention   Psychology of Perception & Attention
Psychology of Perception & Attention Dr.Mohammad Hussein
 
Chapter 14 The Human Eye Lesson 3 - The 2 Mechanisms of the Eye 1) Accommodat...
Chapter 14 The Human Eye Lesson 3 - The 2 Mechanisms of the Eye 1) Accommodat...Chapter 14 The Human Eye Lesson 3 - The 2 Mechanisms of the Eye 1) Accommodat...
Chapter 14 The Human Eye Lesson 3 - The 2 Mechanisms of the Eye 1) Accommodat...j3di79
 
The eye in the control of attention
The eye in the control of attentionThe eye in the control of attention
The eye in the control of attentioneyetech
 
Binocular Single Vision
Binocular Single VisionBinocular Single Vision
Binocular Single VisionMinal Zahid
 
Chapter 14 The Human Eye Lesson 1 - Anatomy of the Human Eye
Chapter 14 The Human Eye Lesson 1 - Anatomy of the Human EyeChapter 14 The Human Eye Lesson 1 - Anatomy of the Human Eye
Chapter 14 The Human Eye Lesson 1 - Anatomy of the Human Eyej3di79
 
Psychology: Visual perception 3. By Janice Fung.
Psychology: Visual perception 3. By Janice Fung.Psychology: Visual perception 3. By Janice Fung.
Psychology: Visual perception 3. By Janice Fung.Janice Fung
 
Physics ( human eye and the colourful world).
Physics ( human eye and the colourful world).Physics ( human eye and the colourful world).
Physics ( human eye and the colourful world).Nikhil Dahiya
 
Physics human eye and colorful world/ © krazzy4™
Physics human  eye and colorful world/ © krazzy4™Physics human  eye and colorful world/ © krazzy4™
Physics human eye and colorful world/ © krazzy4™Vishnu J Kumar
 
MIT6.870 Grounding Object Recognition and Scene Understanding: lecture 1
MIT6.870 Grounding Object Recognition and Scene Understanding: lecture 1MIT6.870 Grounding Object Recognition and Scene Understanding: lecture 1
MIT6.870 Grounding Object Recognition and Scene Understanding: lecture 1zukun
 
Light human eye and the colourful world
Light human eye and the colourful worldLight human eye and the colourful world
Light human eye and the colourful worldUtkarshIshwar
 
Human eye and the colourful world
Human eye and the colourful worldHuman eye and the colourful world
Human eye and the colourful worldaditya315
 
Chapter 14 The Human Eye Lesson 2 - Light Detection
Chapter 14 The Human Eye Lesson 2 - Light DetectionChapter 14 The Human Eye Lesson 2 - Light Detection
Chapter 14 The Human Eye Lesson 2 - Light Detectionj3di79
 
Chapter 6 pwrpt 1 1
Chapter 6  pwrpt 1 1Chapter 6  pwrpt 1 1
Chapter 6 pwrpt 1 1jamiesamel
 
The human eye 10 13-2011
The human eye 10 13-2011The human eye 10 13-2011
The human eye 10 13-2011Varunitya
 
Vision, sensation & perception
Vision, sensation & perceptionVision, sensation & perception
Vision, sensation & perceptionirenek
 
The human eye and the colourful world
The human eye and the colourful worldThe human eye and the colourful world
The human eye and the colourful worldmadhuparna bhowmik
 

Tendances (18)

Psychology of Perception & Attention
Psychology of Perception & Attention   Psychology of Perception & Attention
Psychology of Perception & Attention
 
Chapter 14 The Human Eye Lesson 3 - The 2 Mechanisms of the Eye 1) Accommodat...
Chapter 14 The Human Eye Lesson 3 - The 2 Mechanisms of the Eye 1) Accommodat...Chapter 14 The Human Eye Lesson 3 - The 2 Mechanisms of the Eye 1) Accommodat...
Chapter 14 The Human Eye Lesson 3 - The 2 Mechanisms of the Eye 1) Accommodat...
 
The eye in the control of attention
The eye in the control of attentionThe eye in the control of attention
The eye in the control of attention
 
Binocular Single Vision
Binocular Single VisionBinocular Single Vision
Binocular Single Vision
 
Chapter 14 The Human Eye Lesson 1 - Anatomy of the Human Eye
Chapter 14 The Human Eye Lesson 1 - Anatomy of the Human EyeChapter 14 The Human Eye Lesson 1 - Anatomy of the Human Eye
Chapter 14 The Human Eye Lesson 1 - Anatomy of the Human Eye
 
Psychology: Visual perception 3. By Janice Fung.
Psychology: Visual perception 3. By Janice Fung.Psychology: Visual perception 3. By Janice Fung.
Psychology: Visual perception 3. By Janice Fung.
 
Physics ( human eye and the colourful world).
Physics ( human eye and the colourful world).Physics ( human eye and the colourful world).
Physics ( human eye and the colourful world).
 
Physics human eye and colorful world/ © krazzy4™
Physics human  eye and colorful world/ © krazzy4™Physics human  eye and colorful world/ © krazzy4™
Physics human eye and colorful world/ © krazzy4™
 
MIT6.870 Grounding Object Recognition and Scene Understanding: lecture 1
MIT6.870 Grounding Object Recognition and Scene Understanding: lecture 1MIT6.870 Grounding Object Recognition and Scene Understanding: lecture 1
MIT6.870 Grounding Object Recognition and Scene Understanding: lecture 1
 
Light human eye and the colourful world
Light human eye and the colourful worldLight human eye and the colourful world
Light human eye and the colourful world
 
Human eye and the colourful world
Human eye and the colourful worldHuman eye and the colourful world
Human eye and the colourful world
 
Chapter 14 The Human Eye Lesson 2 - Light Detection
Chapter 14 The Human Eye Lesson 2 - Light DetectionChapter 14 The Human Eye Lesson 2 - Light Detection
Chapter 14 The Human Eye Lesson 2 - Light Detection
 
Physics project
Physics projectPhysics project
Physics project
 
Chapter 6 pwrpt 1 1
Chapter 6  pwrpt 1 1Chapter 6  pwrpt 1 1
Chapter 6 pwrpt 1 1
 
The human eye 10 13-2011
The human eye 10 13-2011The human eye 10 13-2011
The human eye 10 13-2011
 
The human eye
The human eye  The human eye
The human eye
 
Vision, sensation & perception
Vision, sensation & perceptionVision, sensation & perception
Vision, sensation & perception
 
The human eye and the colourful world
The human eye and the colourful worldThe human eye and the colourful world
The human eye and the colourful world
 

En vedette

En vedette (6)

Binocular vision
Binocular visionBinocular vision
Binocular vision
 
Binocular vision basics
Binocular vision basicsBinocular vision basics
Binocular vision basics
 
Basics of binocular vision
Basics of binocular visionBasics of binocular vision
Basics of binocular vision
 
Binocular vision
Binocular visionBinocular vision
Binocular vision
 
binocular single vision
binocular single visionbinocular single vision
binocular single vision
 
Stereoscopic Vision
Stereoscopic VisionStereoscopic Vision
Stereoscopic Vision
 

Similaire à Depth & space

Similaire à Depth & space (20)

Binocular vision and stereopsis
Binocular vision and stereopsisBinocular vision and stereopsis
Binocular vision and stereopsis
 
Binocular vision
Binocular visionBinocular vision
Binocular vision
 
EYE 1.pptx
EYE 1.pptxEYE 1.pptx
EYE 1.pptx
 
Dynamic retinoscopy
Dynamic retinoscopyDynamic retinoscopy
Dynamic retinoscopy
 
Binocular vision final
Binocular vision finalBinocular vision final
Binocular vision final
 
Binocular vision
Binocular visionBinocular vision
Binocular vision
 
Introduction to binocular vision and ocular motility
Introduction to binocular vision and ocular motilityIntroduction to binocular vision and ocular motility
Introduction to binocular vision and ocular motility
 
Introduction vision
Introduction visionIntroduction vision
Introduction vision
 
5. perception
5. perception5. perception
5. perception
 
Elements of visual perception
Elements of visual perceptionElements of visual perception
Elements of visual perception
 
Binocular Single Vision Ophthalmology .pptx
Binocular Single Vision Ophthalmology .pptxBinocular Single Vision Ophthalmology .pptx
Binocular Single Vision Ophthalmology .pptx
 
009457343.pdf
009457343.pdf009457343.pdf
009457343.pdf
 
Stereography 1234.pptx
Stereography 1234.pptxStereography 1234.pptx
Stereography 1234.pptx
 
Binocular vision
Binocular visionBinocular vision
Binocular vision
 
Accomodative convergance and accommodation
Accomodative convergance and accommodationAccomodative convergance and accommodation
Accomodative convergance and accommodation
 
Perception
PerceptionPerception
Perception
 
Remote Sensing
Remote SensingRemote Sensing
Remote Sensing
 
Introduction to binocular single vision (BSV)
Introduction to binocular single vision (BSV)Introduction to binocular single vision (BSV)
Introduction to binocular single vision (BSV)
 
UVC100_Fall16_Class3.1
UVC100_Fall16_Class3.1UVC100_Fall16_Class3.1
UVC100_Fall16_Class3.1
 
chapter 3 - Sensation and perception 2013
chapter 3 - Sensation and perception 2013chapter 3 - Sensation and perception 2013
chapter 3 - Sensation and perception 2013
 

Plus de becca_kennedy

Inclusive Design Thinking: IxDD 2018
Inclusive Design Thinking: IxDD 2018Inclusive Design Thinking: IxDD 2018
Inclusive Design Thinking: IxDD 2018becca_kennedy
 
Do You Trust Your Robots?
Do You Trust Your Robots?Do You Trust Your Robots?
Do You Trust Your Robots?becca_kennedy
 
Usability testing for older adults
Usability testing for older adultsUsability testing for older adults
Usability testing for older adultsbecca_kennedy
 
A novel ergonomic laptop keyboard layout
A novel ergonomic laptop keyboard layoutA novel ergonomic laptop keyboard layout
A novel ergonomic laptop keyboard layoutbecca_kennedy
 
Designing Technology for Children
Designing Technology for ChildrenDesigning Technology for Children
Designing Technology for Childrenbecca_kennedy
 
Kennedy imsh2012 lap_depth_final
Kennedy imsh2012 lap_depth_finalKennedy imsh2012 lap_depth_final
Kennedy imsh2012 lap_depth_finalbecca_kennedy
 
R kennedy hfes 2012 visual aids
R kennedy hfes 2012 visual aidsR kennedy hfes 2012 visual aids
R kennedy hfes 2012 visual aidsbecca_kennedy
 

Plus de becca_kennedy (7)

Inclusive Design Thinking: IxDD 2018
Inclusive Design Thinking: IxDD 2018Inclusive Design Thinking: IxDD 2018
Inclusive Design Thinking: IxDD 2018
 
Do You Trust Your Robots?
Do You Trust Your Robots?Do You Trust Your Robots?
Do You Trust Your Robots?
 
Usability testing for older adults
Usability testing for older adultsUsability testing for older adults
Usability testing for older adults
 
A novel ergonomic laptop keyboard layout
A novel ergonomic laptop keyboard layoutA novel ergonomic laptop keyboard layout
A novel ergonomic laptop keyboard layout
 
Designing Technology for Children
Designing Technology for ChildrenDesigning Technology for Children
Designing Technology for Children
 
Kennedy imsh2012 lap_depth_final
Kennedy imsh2012 lap_depth_finalKennedy imsh2012 lap_depth_final
Kennedy imsh2012 lap_depth_final
 
R kennedy hfes 2012 visual aids
R kennedy hfes 2012 visual aidsR kennedy hfes 2012 visual aids
R kennedy hfes 2012 visual aids
 

Dernier

Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Victor Rentea
 
AI in Action: Real World Use Cases by Anitaraj
AI in Action: Real World Use Cases by AnitarajAI in Action: Real World Use Cases by Anitaraj
AI in Action: Real World Use Cases by AnitarajAnitaRaj43
 
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)Samir Dash
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businesspanagenda
 
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWEREMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWERMadyBayot
 
WSO2's API Vision: Unifying Control, Empowering Developers
WSO2's API Vision: Unifying Control, Empowering DevelopersWSO2's API Vision: Unifying Control, Empowering Developers
WSO2's API Vision: Unifying Control, Empowering DevelopersWSO2
 
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ..."I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...Zilliz
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodJuan lago vázquez
 
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...apidays
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FMESafe Software
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native ApplicationsWSO2
 
Vector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptxVector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptxRemote DBA Services
 
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...apidays
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...DianaGray10
 
Introduction to use of FHIR Documents in ABDM
Introduction to use of FHIR Documents in ABDMIntroduction to use of FHIR Documents in ABDM
Introduction to use of FHIR Documents in ABDMKumar Satyam
 
Corporate and higher education May webinar.pptx
Corporate and higher education May webinar.pptxCorporate and higher education May webinar.pptx
Corporate and higher education May webinar.pptxRustici Software
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingEdi Saputra
 

Dernier (20)

Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
 
AI in Action: Real World Use Cases by Anitaraj
AI in Action: Real World Use Cases by AnitarajAI in Action: Real World Use Cases by Anitaraj
AI in Action: Real World Use Cases by Anitaraj
 
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWEREMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
 
WSO2's API Vision: Unifying Control, Empowering Developers
WSO2's API Vision: Unifying Control, Empowering DevelopersWSO2's API Vision: Unifying Control, Empowering Developers
WSO2's API Vision: Unifying Control, Empowering Developers
 
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ..."I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
 
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
 
Vector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptxVector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptx
 
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
Introduction to use of FHIR Documents in ABDM
Introduction to use of FHIR Documents in ABDMIntroduction to use of FHIR Documents in ABDM
Introduction to use of FHIR Documents in ABDM
 
Corporate and higher education May webinar.pptx
Corporate and higher education May webinar.pptxCorporate and higher education May webinar.pptx
Corporate and higher education May webinar.pptx
 
Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 

Depth & space

  • 1. Depth & Space Becca Kennedy Perception in Real & Virtual Environments 9/19/12
  • 2. Overview • Edges, lines, and texture elements must be interpreted in terms of 3D structure to understand the world • Observer must determine: ▫ Depth – distance of the surface from the observer ▫ Surface orientation – slant and tilt • Depth and surface orientation are recovered together ▫ 3D orientation determines distances of object parts from the observer, and distance of parts determines 3D orientation
  • 3. Overview • Slant – size of the angle between the observer’s line of sight and the surface normal ▫ Surface normal – virtual line sticking out perpendicularly out of the surface at that point • Tilt – the direction of the depth gradient relative to the frontal plane
  • 4. The Problem of Depth Perception • Depth perception from a 2-D retinal image is ambiguous
  • 5. The Problem of Depth Perception – Heuristic Assumptions • Visual system implicitly makes heuristic assumptions about the nature of the world • Our visual system is fooled by 3-D movies ▫ Visual system implicitly assumes that both eyes are looking at the same scene ▫ The different image presented to each eye is interpreted as depth ▫ But usually this heuristic is correct
  • 6. The Problem of Depth Perception – Marr’s 2.5-D Sketch • There are many independent processing modules computing depth information from separate sources ▫ Each module processes different kinds of information • The final common depth interpretation is expressed as a 2.5-D sketch
  • 7. Sources of Depth Information • Ocular information vs. optical information ▫ Ocular information arises from factors that depend on the state of the eyes themselves ▫ Optical information arises from the structure of the light entering the eyes • Binocular information vs. monocular information • Static information vs. dynamic information
  • 8. Ocular Information • Accommodation – the process through which the ciliary muscles in the eye control the optical focus of the lens by temporarily changing its shape ▫ Monocular cue ▫ Thick lens for close objects, thin lens for far objects ▫ Weak source of depth information, but used at close distances
  • 9. Ocular Information • Convergence – the extent to which the two eyes are turned inward to fixate an object ▫ Binocular cue ▫ Fixating on a close object results in a large convergence angle ▫ Fixating on a far object results in a small convergence angle ▫ Visual system uses the angle of eye convergence to determine distance to the fixated point
  • 10. Stereoscopic Information • Stereopsis is the process of perceiving the relative distance to objects based on their lateral displacement in the two retinal images ▫ This relative lateral displacement is binocular disparity  Direction of binocular disparity provides info about which points are closer and which are farther than the fixated point  Magnitude of binocular disparity provides information about how much closer or farther they are • Specifies ratios of distances to objects rather than simply which is farther and which is closer
  • 11. Corresponding Retinal Positions • Corresponding positions on the two retinae are positions that would coincide if the two foveae were superimposed by simple lateral displacement ▫ Binocular disparity occurs when a given point in the external world doesn’t project to corresponding positions  Crossed disparity indicate that a point is closer than the fixated point  Uncrossed disparity indicates that a point is farther away than the fixated point
  • 12. Corresponding Retinal Positions • The horopter is the set of environmental points that stimulate corresponding points on the two retinae ▫ Theoretical horopter –defined as the locus of points which make the same angle at the eyes ▫ Empirical horopter –defined by singleness of vision; larger than theoretical horopter • Panum’s fusional area is the area around the horopter within which disparate images are perceptually fused, so we don’t see double images ▫ Points that lie outside Panum’s area create disparity that we experience as depth
  • 13. The Correspondence Problem • How does the visual system determine which features in one retinal image correspond to which features in the other? • For many years, theorists assumed that this problem was solved by a shape analysis for each left and right image that occurred before stereopsis
  • 14. The Correspondence Problem • The alternative possibility is that stereopsis occurs first ▫ Random dot stereograms  When each image is viewed alone, the dots look random  Shape-first theory would predict that depth perception of random-dot images stereoscopically would be impossible  Random dot stereograms show that stereoscopic depth can be perceived without monocular shape information
  • 15. Computational Theories • Most dots in the left image have a corresponding dot in the right image ▫ The visual system needs to figure out which pairs of dots go together • The first Marr-Poggio algorithm (1976, 1977) ▫ Individual pixels in the left and right images are matched according to location and color  Among these matches, there are the correct ones that correspond to the visible portions of the actual surfaces in the real world ▫ Two heuristic constraints help provide the correct solution  Surface opacity – only the nearest surface can be seen  Surface continuity – correct solution will tend to be one in which matches are close together in depth
  • 16. Edge-Based Algorithms • Marr and Poggio suggested a second algorithm in 1979 • Differed from the first in the following ways: ▫ Edge-based matching – matches edges in the left and right images rather than pixels ▫ Multiple scales – visual system first looks for corresponding edges at a large spatial scale, followed by more detailed matching at finer-grained levels ▫ Single-pass operation – noniterative; finds best edge-based correspondence in a single pass through a multistage operation
  • 17. Multi-Orientation, Multi-Scale (MOMS) filters • Jones and Malik (1990) ▫ A process of matching the vector representing a given position in one eye to each of the vectors representing laterally displaced positions in the other eye  Specifies the most likely correspondence ▫ Better and more robust matches because MOMS vectors carry a lot of spatial information  Compared to outputs of single receptors or edge detectors
  • 18. Physiology of Stereoscopic Vision • Binocular depth cells ▫ Hubel and Wiesel (1962)  Discovered cells in V1 of the visual cortex that were sensitive to binocular stimulation ▫ Barlow, Blakemore, and Pettigrew (1967)  Reported that some binocular cells in area V1 responded optimally to stimulation in disparate locations of the two retinae • To show that these cells are involved in depth perception, need to also demonstrate a connection between disparity and behavior ▫ Blake and Hirsch (1975)  Reared cats so that their vision was alternated between left and right eyes for 6 months  These cats had few binocular neurons and they were not able to use binocular disparity to perceive depth ▫ Recent brain imaging experiments have shown that many different areas are activated by stimuli that create binocular disparity  Depth perception involves many stages of processing that extend from primary visual cortex
  • 20. Dynamic Information • Motion parallax ▫ The differential motion of pairs of points due to their different depths relative to the fixation point  Nearby objects move quickly, far off objects appear stationary • Optic flow caused by a moving observer ▫ Relative to the fixation point…  Points closer to the observer flow in the direction opposite the observer’s motion  Points farther than fixation point flow in the same direction as the observer’s motion
  • 21. Dynamic Information • Another pattern of optic flow is optic expansion or looming ▫ Fixated point is stationary on the retina ▫ Other points flow outward, faster with more distance from fixation point
  • 22. Dynamic Information • Optic flow caused by moving objects ▫ Kinetic depth effect (KDE; Wallach & O’Connell, 1953) – ability to perceive depth from object motion ▫ Visual system uses a rigidity heuristic  Biased toward perceiving rigid motions rather than plastic motions • Accretion/Deletion of Texture ▫ Appearance and disappearance of texture behind a moving edge
  • 23. Pictorial Information • Convergence of parallel lines • Position relative to the horizon
  • 24. Pictorial Information • Relative size • Familiar size ▫ In a VE, if not enough depth cues are present, the observer begins to depend on retinal size (Kenyon, Sandin, Smith, Pawlicki, & Defanti, 2007) • Texture gradients
  • 25. Pictorial Information – Edge Interpretation • Edge and contour interpretations ▫ E.g., occlusion or interposition – blocking of light from an object by an opaque object causing occlusion or interposition ▫ Edges provide relative rather than absolute depth information ▫ Available from virtually unlimited distances within visible range • Vertex (edge intersection) classification ▫ Guzman’s (1968, 1969) program SEE attempted to interpret line drawings of simple configurations of blocks  He developed a classification scheme for edge intersections (vertices): Ts, Ys, Ks, Xs, Ls, etc. ▫ Huffman and Clowes (1971) developed a complete catalog of the vertex types that arise in viewing simple trihedral angles from all possible viewpoints
  • 26. Pictorial Information – Edge Interpretation • Four types of edges: 1. Orientation edges – places where there are discontinuities in surface orientation; when two different orientations meet along an edge 2. Depth edges – places where there is a spatial discontinuity in depth between surfaces; places where one surface occludes another that extends behind it, with space between 3. Illumination edges – places where there is a difference in the amount of light falling on a homogenous surface; edge of a shadow, highlight, or spotlight 4. Reflectance edges – places where there is a change in the light-reflecting properties of the surface material; e.g., designs painted on a surface
  • 27. Pictorial Information – Edge Interpretation • Edge labels ▫ Two kinds of orientation edges  Convex orientation edges are labeled with a +  Concave orientation edges are labeled with a - ▫ Arrows indicate that the closer surface is on the right
  • 28. Pictorial Information – Edge Interpretation
  • 29. Pictorial Information – Edge Interpretation • Physical constraints ▫ Not all logically possible labelings are physically possible
  • 30. Pictorial Information – Edge Interpretation • Extensions and Generalizations ▫ Waltz (1975) extended the Huffman-Clowes analysis to include 11 types of edges, including shadows and “cracks” (orientation edges at 180 degree angles)  Adding shadows making interpretation more accurate because it provides further constraints ▫ Malik (1987) extended analysis of edge labeling to curved objects  New depth edge type extremal edge or limb (double arrow) occurs when a surface curves smoothly around to partly exclude itself
  • 31. Pictorial Information – Edge Interpretation • Extensions and Generalizations ▫ Barrow and Tennenbaum (1978)’s analysis contained additional constraints:  The smoothness assumption – if an occluding edge in the image is smooth, then so is the contour of the surface that produced it  The general viewpoint assumption – small changes in viewpoint will not cause qualitative differences in the image
  • 32. Pictorial Information • Shading information ▫ Shading – variations in the amount of light reflected from the surface as a result of variations in the orientation of the surface relative to a light source ▫ Horn’s (1975, 1977) Computational Analysis  Showed that percentage changes in image luminance are directly proportional to percentage changes in the orientation of the surface ▫ Humans are able to interpret surfaces with significantly specular characteristics, like glossy surfaces that reflect light more coherently than matte surfaces do  How?
  • 33. Pictorial Information • Shading information ▫ Cast shadows  Shadows of one objet that fall on the surface of another object provide more depth information  Distance between object and its shadow cast on the surface gives height of its bottom above the surface
  • 34. Pictorial Information • Aerial perspective ▫ Refers to certain systematic differences in the contrast and color of objects that occur when they are viewed from great distances  Contrast is reduced by additional atmosphere through which they are viewed, which contains particles of dust, water, or pollutants that scatter light  Mountains that are far away appear bluer because the atmosphere scatters longer wavelengths of light more than shorter wavelengths
  • 35. Integrating Information Sources • Depth cues are often highly correlated, making them easy to integrate • What happens when cues are in conflict with one another? 3 possibilities: 1. One source dominates a conflicting source  E.g., In Ames room, perspective information dominates familiar size 2. A compromise is achieved between two conflicting sources  Visual system makes independent estimates of depth from each source alone, then integrates them according to a mathematical rule  Bruno and Cutting (1988) found that information integration was additive; sum independent effects of sources 3. The two sources interact to arrive at an optimal solution  E.g., convergence specifies absolute depth, binocular disparity specifies ratios of distances; together they can provide a complete depth map
  • 36. Depth Perception and VEs • Our visual system is really good at depth perception in real environments, but this is hard to replicate in virtual scenes ▫ Ocular depth information (accommodation, convergence) is less useful ▫ Stereoscopic depth information may not be available ▫ Motion cues may not be faithfully represented ▫ Depth cues may be conflicting ▫ Etc.! • But augmented reality can also improve real-world depth perception