SlideShare une entreprise Scribd logo
1  sur  22
Seminar By Melisha
Monteiro
 The volume of uncompressed video data could be extremly
large
 The reason is that video contains much spatial and temporal
redundancy.
 In a single frame, nearby pixels are often correlated with each
other. This is called spatial redundancy, or the intraframe
correlation.
 Another one is temporal redundancy, which means adjacent
frames are highly correlated, or called the interframe
correlation. Therefore, our goal is to efficiently reduce spatial
and temporal redundancy to achieve video compression.
 Video consist of time ordered sequence of
frames.
 An obvious solution to video compression
would be predictive coding based on previous
frames.
 Video compression deals with the compression of
visual video data.
 Video contains spatial & temporal redundancy.
 We reduce the redundency using lossy compression
 Image compression techniques to reduce spatial
redundency
 Motion compensation techniques to reduce temporal
redundency
 This was invented in 1960s
 To reduce temporal redundancy in video sequence, we use the
motion-compensated prediction method.
 Temporal redundancy is exploited so that not every frame of the
video needs to be coded independently as a new image. The
difference between the current frame and other frame(s) in the
sequence will be coded — small values and low entropy, good for
compression.
• A practical and widely-used method of motion compensation
is to compensate for movement of blocks of the current frame,
as show in Fig.
• The following procedure is carried out for each MxN block in
the current frame:
 Search an area in the reference frame (past or future
frame, previously coded and transmitted) to find the
best matching M x N block. This is carried out by
comparing the MxN block in the current frame with all
of the possible MxN regions in the search area. This
process of finding the best match is known as Motion
Estimation (ME).
 When the best matching block is found in the reference
frame by motion estimation, we subtract the best
matching block from the current macroblock to
produce a residual macroblock.
 Within the encoder, the residual is encoded and
decoded and added to the matching region to form a
reconstructed macroblock which is stored as a
reference for further motion-compensated prediction.
 The residual block is encoded and transmitted
and the offset between the current block and
the position of the candidate region motion
vectors is also transmitted.
 The displacement between current block and
reference block .
 – The current image frame is referred to as Target
Frame.
 – A match is sought between the macroblock in the
Target Frame
 and the most similar macroblock in previous and/or
future frame(s) (referred to as Reference frame(s)).
 – The displacement of the reference macroblock to the
target macroblock is called a motion vector MV.
 – Figure shows the case of forward prediction in
which the Reference frame is taken to be a previous
frame.
 MV search is usually limited to a small immediate
neighborhood — both horizontal and vertical
displacements in the range [−p, p].
 This makes a search window of size (2p+1) ×
(2p+1).
 The difference between two macroblocks can then
be measured by their Mean Absolute Difference
(MAD):
 N – size of the macroblock,
 k and l – indices for pixels in the macroblock,
 i and j – horizontal and vertical displacements,
 C(x+k, y +l) – pixels in macroblock in Target
frame,
 R(x+i+k, y +j +l) – pixels in macroblock in
Reference frame.
 Black and white TV picture is generated by exciting
the phosphor on the television screen using an electron
beam whose intensity is modulated to generate the
image .
 The line created by horizontal traversal of electron
beam is called a line of the image.
 In order to trace second line electron beam as to be
deflected back to the left screen .
 The gun is turned off in order to prevent retrace from
becoming visible.
 In order to keep the cost of bandwidth low it was
decided to send 525 lines 30 times in a seconds .
 525 lines said to constitue -frame
 To avoid the flicker , devide the into two interlaced
fields .
IMAGE
FIELD FIELD
Sixtieth
second
262.5 lines traced by
electron beam
262.5 lines traced
btn lines of first
field
 In colour tv instead of single electron gun, we have
three electron guns ,these excite red green blue
phosphor dots embeded in the screen .
 To control this we require three signals red green
blue signals ,if they transmit individually they
require three times the bandwidth –problem
backward compatibility.
 The human visual system (HVS) is less sensitive
to color than to luminance (brightness).
 It is a popular way of efficiently representing color
images. Y is luminance (luma) component and the
color information can be represented as color
 difference (chrominance or chroma) components,
where each chrominance component is the
difference between R, G or B and the luminance Y
The lumiance component is expressed as
Y = 0.299R + 0.587G +0.114B

Contenu connexe

Tendances

Advanced Lighting Techniques Dan Baker (Meltdown 2005)
Advanced Lighting Techniques   Dan Baker (Meltdown 2005)Advanced Lighting Techniques   Dan Baker (Meltdown 2005)
Advanced Lighting Techniques Dan Baker (Meltdown 2005)mobius.cn
 
Two dimensional 3 d venn - dual
Two dimensional 3 d venn - dualTwo dimensional 3 d venn - dual
Two dimensional 3 d venn - dualRalph Sherman
 
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View SynthesisPR-302: NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View SynthesisHyeongmin Lee
 
Multimedia lossy compression algorithms
Multimedia lossy compression algorithmsMultimedia lossy compression algorithms
Multimedia lossy compression algorithmsMazin Alwaaly
 
2020 12-04-shake shake
2020 12-04-shake shake2020 12-04-shake shake
2020 12-04-shake shakeJAEMINJEONG5
 
PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...
PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...
PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...Jinwon Lee
 
Semi fragile watermarking
Semi fragile watermarkingSemi fragile watermarking
Semi fragile watermarkingYash Diwakar
 
2021 03-01-on the relationship between self-attention and convolutional layers
2021 03-01-on the relationship between self-attention and convolutional layers2021 03-01-on the relationship between self-attention and convolutional layers
2021 03-01-on the relationship between self-attention and convolutional layersJAEMINJEONG5
 
A primer for colour computer vision
A primer for colour computer visionA primer for colour computer vision
A primer for colour computer visionpotaters
 
Build Your Own 3D Scanner: Surface Reconstruction
Build Your Own 3D Scanner: Surface ReconstructionBuild Your Own 3D Scanner: Surface Reconstruction
Build Your Own 3D Scanner: Surface ReconstructionDouglas Lanman
 
B140715
B140715B140715
B140715irjes
 
Performance of MMSE Denoise Signal Using LS-MMSE Technique
Performance of MMSE Denoise Signal Using LS-MMSE  TechniquePerformance of MMSE Denoise Signal Using LS-MMSE  Technique
Performance of MMSE Denoise Signal Using LS-MMSE TechniqueIJMER
 

Tendances (18)

Advanced Lighting Techniques Dan Baker (Meltdown 2005)
Advanced Lighting Techniques   Dan Baker (Meltdown 2005)Advanced Lighting Techniques   Dan Baker (Meltdown 2005)
Advanced Lighting Techniques Dan Baker (Meltdown 2005)
 
Two dimensional 3 d venn - dual
Two dimensional 3 d venn - dualTwo dimensional 3 d venn - dual
Two dimensional 3 d venn - dual
 
Chapter 9 newer
Chapter 9   newerChapter 9   newer
Chapter 9 newer
 
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View SynthesisPR-302: NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
PR-302: NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
 
Multimedia lossy compression algorithms
Multimedia lossy compression algorithmsMultimedia lossy compression algorithms
Multimedia lossy compression algorithms
 
2020 12-04-shake shake
2020 12-04-shake shake2020 12-04-shake shake
2020 12-04-shake shake
 
Antennas
AntennasAntennas
Antennas
 
Sid39 2
Sid39 2Sid39 2
Sid39 2
 
PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...
PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...
PR-344: A Battle of Network Structures: An Empirical Study of CNN, Transforme...
 
Semi fragile watermarking
Semi fragile watermarkingSemi fragile watermarking
Semi fragile watermarking
 
2021 03-01-on the relationship between self-attention and convolutional layers
2021 03-01-on the relationship between self-attention and convolutional layers2021 03-01-on the relationship between self-attention and convolutional layers
2021 03-01-on the relationship between self-attention and convolutional layers
 
A primer for colour computer vision
A primer for colour computer visionA primer for colour computer vision
A primer for colour computer vision
 
Build Your Own 3D Scanner: Surface Reconstruction
Build Your Own 3D Scanner: Surface ReconstructionBuild Your Own 3D Scanner: Surface Reconstruction
Build Your Own 3D Scanner: Surface Reconstruction
 
Slides Hair Rendering
Slides Hair RenderingSlides Hair Rendering
Slides Hair Rendering
 
ResearchPaper_Final
ResearchPaper_FinalResearchPaper_Final
ResearchPaper_Final
 
B140715
B140715B140715
B140715
 
Performance of MMSE Denoise Signal Using LS-MMSE Technique
Performance of MMSE Denoise Signal Using LS-MMSE  TechniquePerformance of MMSE Denoise Signal Using LS-MMSE  Technique
Performance of MMSE Denoise Signal Using LS-MMSE Technique
 
floor planning
floor planningfloor planning
floor planning
 

En vedette

En vedette (6)

Mmclass4
Mmclass4Mmclass4
Mmclass4
 
PPT
PPTPPT
PPT
 
ARPS Architecture
ARPS ArchitectureARPS Architecture
ARPS Architecture
 
Dsc
DscDsc
Dsc
 
M.TECH Project, Sem 3 presentation
M.TECH Project, Sem 3 presentation M.TECH Project, Sem 3 presentation
M.TECH Project, Sem 3 presentation
 
COM2304: Digital Image Fundamentals - I
COM2304: Digital Image Fundamentals - I COM2304: Digital Image Fundamentals - I
COM2304: Digital Image Fundamentals - I
 

Similaire à Melisha Monteiro Seminar on Video Compression Techniques

Multimedia basic video compression techniques
Multimedia basic video compression techniquesMultimedia basic video compression techniques
Multimedia basic video compression techniquesMazin Alwaaly
 
Kernel Estimation of Videodeblurringalgorithm and Motion Compensation of Resi...
Kernel Estimation of Videodeblurringalgorithm and Motion Compensation of Resi...Kernel Estimation of Videodeblurringalgorithm and Motion Compensation of Resi...
Kernel Estimation of Videodeblurringalgorithm and Motion Compensation of Resi...IJERA Editor
 
IRJET- Comparison and Simulation based Analysis of an Optimized Block Mat...
IRJET-  	  Comparison and Simulation based Analysis of an Optimized Block Mat...IRJET-  	  Comparison and Simulation based Analysis of an Optimized Block Mat...
IRJET- Comparison and Simulation based Analysis of an Optimized Block Mat...IRJET Journal
 
Motion detection in compressed video using macroblock classification
Motion detection in compressed video using macroblock classificationMotion detection in compressed video using macroblock classification
Motion detection in compressed video using macroblock classificationacijjournal
 
EFFICIENT IMAGE COMPRESSION USING LAPLACIAN PYRAMIDAL FILTERS FOR EDGE IMAGES
EFFICIENT IMAGE COMPRESSION USING LAPLACIAN PYRAMIDAL FILTERS FOR EDGE IMAGESEFFICIENT IMAGE COMPRESSION USING LAPLACIAN PYRAMIDAL FILTERS FOR EDGE IMAGES
EFFICIENT IMAGE COMPRESSION USING LAPLACIAN PYRAMIDAL FILTERS FOR EDGE IMAGESijcnac
 
Video Compression Advanced.pdf
Video Compression Advanced.pdfVideo Compression Advanced.pdf
Video Compression Advanced.pdfSMohiuddin1
 
THE TELEVISION SYSTEM IN INDIA
THE TELEVISION SYSTEM IN INDIATHE TELEVISION SYSTEM IN INDIA
THE TELEVISION SYSTEM IN INDIAIshank Ranjan
 
Motion Estimation - umit 5 (II).pdf
Motion Estimation  - umit 5 (II).pdfMotion Estimation  - umit 5 (II).pdf
Motion Estimation - umit 5 (II).pdfHeenaSyed6
 
Back projection geometry in cbct
Back projection geometry in cbctBack projection geometry in cbct
Back projection geometry in cbctDrGayatriMehrotra
 
Frequency Domain Blockiness and Blurriness Meter for Image Quality Assessment
Frequency Domain Blockiness and Blurriness Meter for Image Quality AssessmentFrequency Domain Blockiness and Blurriness Meter for Image Quality Assessment
Frequency Domain Blockiness and Blurriness Meter for Image Quality AssessmentCSCJournals
 
Fractal Image Compression of Satellite Color Imageries Using Variable Size of...
Fractal Image Compression of Satellite Color Imageries Using Variable Size of...Fractal Image Compression of Satellite Color Imageries Using Variable Size of...
Fractal Image Compression of Satellite Color Imageries Using Variable Size of...CSCJournals
 
Lossless Huffman coding image compression implementation in spatial domain by...
Lossless Huffman coding image compression implementation in spatial domain by...Lossless Huffman coding image compression implementation in spatial domain by...
Lossless Huffman coding image compression implementation in spatial domain by...IRJET Journal
 

Similaire à Melisha Monteiro Seminar on Video Compression Techniques (20)

Multimedia basic video compression techniques
Multimedia basic video compression techniquesMultimedia basic video compression techniques
Multimedia basic video compression techniques
 
Kernel Estimation of Videodeblurringalgorithm and Motion Compensation of Resi...
Kernel Estimation of Videodeblurringalgorithm and Motion Compensation of Resi...Kernel Estimation of Videodeblurringalgorithm and Motion Compensation of Resi...
Kernel Estimation of Videodeblurringalgorithm and Motion Compensation of Resi...
 
IRJET- Comparison and Simulation based Analysis of an Optimized Block Mat...
IRJET-  	  Comparison and Simulation based Analysis of an Optimized Block Mat...IRJET-  	  Comparison and Simulation based Analysis of an Optimized Block Mat...
IRJET- Comparison and Simulation based Analysis of an Optimized Block Mat...
 
Motion detection in compressed video using macroblock classification
Motion detection in compressed video using macroblock classificationMotion detection in compressed video using macroblock classification
Motion detection in compressed video using macroblock classification
 
EFFICIENT IMAGE COMPRESSION USING LAPLACIAN PYRAMIDAL FILTERS FOR EDGE IMAGES
EFFICIENT IMAGE COMPRESSION USING LAPLACIAN PYRAMIDAL FILTERS FOR EDGE IMAGESEFFICIENT IMAGE COMPRESSION USING LAPLACIAN PYRAMIDAL FILTERS FOR EDGE IMAGES
EFFICIENT IMAGE COMPRESSION USING LAPLACIAN PYRAMIDAL FILTERS FOR EDGE IMAGES
 
Video Compression Advanced.pdf
Video Compression Advanced.pdfVideo Compression Advanced.pdf
Video Compression Advanced.pdf
 
40120140503006
4012014050300640120140503006
40120140503006
 
THE TELEVISION SYSTEM IN INDIA
THE TELEVISION SYSTEM IN INDIATHE TELEVISION SYSTEM IN INDIA
THE TELEVISION SYSTEM IN INDIA
 
Motion Estimation - umit 5 (II).pdf
Motion Estimation  - umit 5 (II).pdfMotion Estimation  - umit 5 (II).pdf
Motion Estimation - umit 5 (II).pdf
 
L0936775
L0936775L0936775
L0936775
 
Back projection geometry in cbct
Back projection geometry in cbctBack projection geometry in cbct
Back projection geometry in cbct
 
Broadcaster Notes
Broadcaster NotesBroadcaster Notes
Broadcaster Notes
 
Frequency Domain Blockiness and Blurriness Meter for Image Quality Assessment
Frequency Domain Blockiness and Blurriness Meter for Image Quality AssessmentFrequency Domain Blockiness and Blurriness Meter for Image Quality Assessment
Frequency Domain Blockiness and Blurriness Meter for Image Quality Assessment
 
www.ijerd.com
www.ijerd.comwww.ijerd.com
www.ijerd.com
 
Fractal Image Compression of Satellite Color Imageries Using Variable Size of...
Fractal Image Compression of Satellite Color Imageries Using Variable Size of...Fractal Image Compression of Satellite Color Imageries Using Variable Size of...
Fractal Image Compression of Satellite Color Imageries Using Variable Size of...
 
538 207-219
538 207-219538 207-219
538 207-219
 
557 480-486
557 480-486557 480-486
557 480-486
 
056-2004 SID Manuscript
056-2004 SID Manuscript056-2004 SID Manuscript
056-2004 SID Manuscript
 
Lossless Huffman coding image compression implementation in spatial domain by...
Lossless Huffman coding image compression implementation in spatial domain by...Lossless Huffman coding image compression implementation in spatial domain by...
Lossless Huffman coding image compression implementation in spatial domain by...
 
L0351007379
L0351007379L0351007379
L0351007379
 

Melisha Monteiro Seminar on Video Compression Techniques

  • 2.
  • 3.  The volume of uncompressed video data could be extremly large  The reason is that video contains much spatial and temporal redundancy.  In a single frame, nearby pixels are often correlated with each other. This is called spatial redundancy, or the intraframe correlation.  Another one is temporal redundancy, which means adjacent frames are highly correlated, or called the interframe correlation. Therefore, our goal is to efficiently reduce spatial and temporal redundancy to achieve video compression.
  • 4.  Video consist of time ordered sequence of frames.  An obvious solution to video compression would be predictive coding based on previous frames.
  • 5.  Video compression deals with the compression of visual video data.  Video contains spatial & temporal redundancy.  We reduce the redundency using lossy compression  Image compression techniques to reduce spatial redundency  Motion compensation techniques to reduce temporal redundency
  • 6.  This was invented in 1960s  To reduce temporal redundancy in video sequence, we use the motion-compensated prediction method.  Temporal redundancy is exploited so that not every frame of the video needs to be coded independently as a new image. The difference between the current frame and other frame(s) in the sequence will be coded — small values and low entropy, good for compression.
  • 7. • A practical and widely-used method of motion compensation is to compensate for movement of blocks of the current frame, as show in Fig. • The following procedure is carried out for each MxN block in the current frame:
  • 8.  Search an area in the reference frame (past or future frame, previously coded and transmitted) to find the best matching M x N block. This is carried out by comparing the MxN block in the current frame with all of the possible MxN regions in the search area. This process of finding the best match is known as Motion Estimation (ME).
  • 9.  When the best matching block is found in the reference frame by motion estimation, we subtract the best matching block from the current macroblock to produce a residual macroblock.  Within the encoder, the residual is encoded and decoded and added to the matching region to form a reconstructed macroblock which is stored as a reference for further motion-compensated prediction.
  • 10.  The residual block is encoded and transmitted and the offset between the current block and the position of the candidate region motion vectors is also transmitted.  The displacement between current block and reference block .
  • 11.  – The current image frame is referred to as Target Frame.  – A match is sought between the macroblock in the Target Frame  and the most similar macroblock in previous and/or future frame(s) (referred to as Reference frame(s)).  – The displacement of the reference macroblock to the target macroblock is called a motion vector MV.  – Figure shows the case of forward prediction in which the Reference frame is taken to be a previous frame.
  • 12.
  • 13.  MV search is usually limited to a small immediate neighborhood — both horizontal and vertical displacements in the range [−p, p].  This makes a search window of size (2p+1) × (2p+1).
  • 14.  The difference between two macroblocks can then be measured by their Mean Absolute Difference (MAD):
  • 15.  N – size of the macroblock,  k and l – indices for pixels in the macroblock,  i and j – horizontal and vertical displacements,  C(x+k, y +l) – pixels in macroblock in Target frame,  R(x+i+k, y +j +l) – pixels in macroblock in Reference frame.
  • 16.
  • 17.  Black and white TV picture is generated by exciting the phosphor on the television screen using an electron beam whose intensity is modulated to generate the image .  The line created by horizontal traversal of electron beam is called a line of the image.  In order to trace second line electron beam as to be deflected back to the left screen .  The gun is turned off in order to prevent retrace from becoming visible.
  • 18.  In order to keep the cost of bandwidth low it was decided to send 525 lines 30 times in a seconds .  525 lines said to constitue -frame  To avoid the flicker , devide the into two interlaced fields .
  • 19. IMAGE FIELD FIELD Sixtieth second 262.5 lines traced by electron beam 262.5 lines traced btn lines of first field
  • 20.  In colour tv instead of single electron gun, we have three electron guns ,these excite red green blue phosphor dots embeded in the screen .  To control this we require three signals red green blue signals ,if they transmit individually they require three times the bandwidth –problem backward compatibility.
  • 21.  The human visual system (HVS) is less sensitive to color than to luminance (brightness).  It is a popular way of efficiently representing color images. Y is luminance (luma) component and the color information can be represented as color  difference (chrominance or chroma) components, where each chrominance component is the difference between R, G or B and the luminance Y
  • 22. The lumiance component is expressed as Y = 0.299R + 0.587G +0.114B