Digital image processing involves the manipulation of digital images through various algorithms and techniques. The key steps involve image acquisition through sensors, preprocessing such as sampling and quantization, processing such as enhancement and analysis, and output. Digital image processing has applications in fields such as medicine, astronomy, security, and more. It allows analysis and manipulation of images to improve quality or extract useful information.
Vector graphics use mathematical formulas to define images as objects made of points and paths, allowing resolution-independent scaling. Raster graphics are composed of pixels arranged in a grid to form images. Key factors that determine raster image quality include resolution, color depth, and file format. Common file formats like JPEG, PNG, and GIF vary in their compression algorithms and support for animation and transparency.
Digital image processing refers to manipulating, enhancing, and analyzing digital images using computer algorithms and techniques. It involves applying mathematical operations to digital images, which are treated as two-dimensional arrays of pixels where each pixel represents a point of color and brightness. The basic steps in digital image processing are image acquisition, enhancement, restoration, segmentation, representation/description, analysis, synthesis/compression. Digital image processing is widely used in applications like medical imaging, computer vision, and multimedia.
The document outlines the course objectives, outcomes, examination scheme, and units of a Computer Graphics course. The course aims to acquaint students with basic concepts, algorithms, and techniques of computer graphics through understanding, applying, and creating graphics using OpenGL. Students will learn about primitives, transformations, projections, lighting, shading, animation and gaming. The course assessment includes a mid-semester test, end-semester test, and covers topics ranging from graphics primitives to fractals and animation.
There are two types of images - analog and digital. Digital images are represented as a matrix of pixels, with each pixel represented by a numerical value. The structure and characteristics of digital images, such as pixel bit depth, matrix size, and field of view, determine image quality and detail. Digital images allow for processing and analysis that analog images do not.
The document discusses image processing and provides information on several key topics:
1. Image processing can be grouped into compression, preprocessing, and analysis. Preprocessing improves image quality by reducing noise and enhancing edges. Analysis extracts numeric or graphical information for tasks like classification.
2. Images are 2D matrices of intensity values represented by pixels. Common digital formats include grayscale, RGB, and RGBA. Higher bit depths allow more intensity levels to be represented.
3. Basic measurements of images include spatial resolution in pixels per unit, bit depth determining representable intensity levels, and factors like saturation and noise.
Computer vision is a field that deals with how computers can understand digital images and videos. It seeks to automate tasks that the human visual system can perform, such as object recognition. One example is computer vision systems in self-driving cars that can identify objects on the road to help drivers or prevent collisions. Computer vision involves various image processing levels from low-level tasks like noise removal to high-level tasks like scene understanding. Digital images come in various types like binary, grayscale, and color images represented by different numbers of bits per pixel.
Digital image processing involves the manipulation of digital images through various algorithms and techniques. The key steps involve image acquisition through sensors, preprocessing such as sampling and quantization, processing such as enhancement and analysis, and output. Digital image processing has applications in fields such as medicine, astronomy, security, and more. It allows analysis and manipulation of images to improve quality or extract useful information.
Vector graphics use mathematical formulas to define images as objects made of points and paths, allowing resolution-independent scaling. Raster graphics are composed of pixels arranged in a grid to form images. Key factors that determine raster image quality include resolution, color depth, and file format. Common file formats like JPEG, PNG, and GIF vary in their compression algorithms and support for animation and transparency.
Digital image processing refers to manipulating, enhancing, and analyzing digital images using computer algorithms and techniques. It involves applying mathematical operations to digital images, which are treated as two-dimensional arrays of pixels where each pixel represents a point of color and brightness. The basic steps in digital image processing are image acquisition, enhancement, restoration, segmentation, representation/description, analysis, synthesis/compression. Digital image processing is widely used in applications like medical imaging, computer vision, and multimedia.
The document outlines the course objectives, outcomes, examination scheme, and units of a Computer Graphics course. The course aims to acquaint students with basic concepts, algorithms, and techniques of computer graphics through understanding, applying, and creating graphics using OpenGL. Students will learn about primitives, transformations, projections, lighting, shading, animation and gaming. The course assessment includes a mid-semester test, end-semester test, and covers topics ranging from graphics primitives to fractals and animation.
There are two types of images - analog and digital. Digital images are represented as a matrix of pixels, with each pixel represented by a numerical value. The structure and characteristics of digital images, such as pixel bit depth, matrix size, and field of view, determine image quality and detail. Digital images allow for processing and analysis that analog images do not.
The document discusses image processing and provides information on several key topics:
1. Image processing can be grouped into compression, preprocessing, and analysis. Preprocessing improves image quality by reducing noise and enhancing edges. Analysis extracts numeric or graphical information for tasks like classification.
2. Images are 2D matrices of intensity values represented by pixels. Common digital formats include grayscale, RGB, and RGBA. Higher bit depths allow more intensity levels to be represented.
3. Basic measurements of images include spatial resolution in pixels per unit, bit depth determining representable intensity levels, and factors like saturation and noise.
Computer vision is a field that deals with how computers can understand digital images and videos. It seeks to automate tasks that the human visual system can perform, such as object recognition. One example is computer vision systems in self-driving cars that can identify objects on the road to help drivers or prevent collisions. Computer vision involves various image processing levels from low-level tasks like noise removal to high-level tasks like scene understanding. Digital images come in various types like binary, grayscale, and color images represented by different numbers of bits per pixel.
This document outlines the syllabus for the course IT6005 - Digital Image Processing. The syllabus is divided into 5 units that cover digital image fundamentals, image enhancement, image restoration and segmentation, wavelets and image compression, and image representation and recognition. Unit 1 introduces key concepts in digital image processing such as pixels, gray levels, sampling and quantization. It also provides a brief history of the origin and development of digital image processing.
This document summarizes various topics related to image processing including image data types, file formats, acquisition, storage, processing, communication, display, and enhancement techniques. It discusses key concepts such as image fundamentals, color models, resolution, bit depth, file formats like JPEG, GIF, TIFF, compression techniques including lossless, lossy, intraframe, interframe, and algorithms like run length encoding and Shannon-Fano coding. Image enhancement topics covered are point processing, spatial filtering, and color image processing.
The Indian Dental Academy is the Leader in continuing dental education , training dentists in all aspects of dentistry and
offering a wide range of dental certified courses in different formats.for more details please visit
www.indiandentalacademy.com
This document provides an introduction to digital image processing. It defines what an image and digital image are, and discusses the first ever digital photograph. It describes digital image processing as processing digital images using computers, with sources including the electromagnetic spectrum from gamma rays to radio waves. Key concepts covered include digital images, image enhancement through spatial and frequency domain methods, image restoration to remove noise and blurring, and image compression to reduce file size through removing different types of data redundancy.
Discover the fundamentals, Characteristics & types of digital image analysis. Learn about pixels, bit depth, challenges, and AI impacts on image processing.
computervision1.pdf it is about computer visionshesnasuneer
This document provides an introduction to digital image processing and computer vision. It discusses how images are represented digitally through sampling and quantization. Low-level image processing techniques like preprocessing, segmentation, and object description are used to simplify computer vision tasks. Fundamental concepts in digital image processing are also introduced, such as how images can be represented as functions and processed using mathematical tools like the Fourier transform and convolution.
The document discusses fundamentals of digital images including representation as pixels, color models like RGB and CMYK, color depth, resolution, and file formats. It also covers topics like dithering, 2D graphics as vector or raster, and image compression standards. Key aspects covered include how pixels and bit depth determine color representation, uses of RGB vs CMYK color schemes, and how dithering creates illusions of additional colors through pixel arrangement.
Optical Watermarking Literature survey....Arif Ahmed
1. The document discusses optical watermarking, a technique for protecting copyright of real-world objects like artwork by embedding watermarks in the illumination of the objects. When photos are taken of the illuminated objects, the watermarks are captured in the digital images and can be extracted.
2. Optical watermarking works by transforming binary data into patterns of light projected onto objects. The patterns differ based on 1s and 0s in the data. When photos of the illuminated objects are taken, the patterns can be read from the captured images. Visible light is used with fine or low-contrast patterns to make the watermarks imperceptible.
3. Optical watermarking provides copyright protection for valuable
This document provides an overview of a digital image processing course. It outlines 4 course outcomes: 1) describing basic concepts and applications of image processing, 2) describing techniques in color, segmentation, and recognition, 3) illustrating pixel relationships and image arithmetic, and 4) analyzing digital image enhancement principles. The document then discusses topics that will be covered in the course, including image types, operations, and applications in various fields.
Digital image processing involves converting analog images to digital format using techniques like sampling and quantization. Sampling converts continuous images to discrete pixel values that define the image's spatial resolution. Quantization divides signal amplitudes into quanta to derive values that determine image quality. Together, sampling and quantization are key to converting images to digital format and improving qualities like sharpness, brightness, and clarity.
The document discusses image processing and provides details about:
1) The stages of image processing - input, editing, and output. The input stage deals with converting analog images to digital form. The editing stage manipulates the image. The output stage saves the transformed image.
2) Image processing operations like geometric transformations, color corrections, digital compositing, and extending dynamic range.
3) Applications of image processing like face detection, medical imaging, and remote sensing.
This document provides an overview of image processing and related concepts. It discusses:
1) What an image is, the basic steps of image processing like acquisition, analysis and output, and the two main types: analogue and digital.
2) Key aspects of digital image processing including what a digital image and pixel are, and common processing steps of pre-processing, enhancement and extraction.
3) Common image processing techniques like enhancement which adjusts images, and segmentation which partitions an image into meaningful segments.
4) Image classification which predicts categories from inputs, and the main types of supervised and unsupervised classification. It provides an example using Landsat satellite images.
This document provides an overview of image analysis, including:
1) It defines image analysis and discusses its use in recognizing, differentiating, and quantifying images across various fields including food quality assessment.
2) It describes the process of creating a digital image through digitization and discusses key aspects of digital images like resolution, pixel bit depth, and color.
3) It outlines common image processing actions like compression, preprocessing, and analysis and provides examples of applying image analysis to evaluate food products.
YCIS_Forensic PArt 1 Digital Image Processing.pptxSharmilaMore5
Basics of Digital Image Processing
Use of DIP in Society
Digital Image Processing Process
Why do we process images?
Image Enhancement and Edge detection
Python
How are we using Python in DIP
Biomedical Image Processing
Topics covered: Biomedical imaging, Need of image processing in medicine, Principles of image processing, Components of image processing, Application of image processing in different medical imaging systems
This document is a mini project report on digital image processing using MATLAB. It discusses various image processing techniques and applications implemented in MATLAB, including image formats, operations, and tools. Applications demonstrated include text recognition, color tracking, solving an engineering problem using image processing, creating a virtual slate using laser tracking, face detection, and distance estimation. The report provides examples of MATLAB functions used for tasks like importing, displaying, converting and cropping images, as well as analyzing and manipulating them.
Digital image processing & computer graphicsAnkit Garg
Digital Image Processing & Computer Graphics document discusses several topics related to digital image processing including:
1. Digital image processing involves manipulating digital images using computer programs. It includes operations like geometric transformations, image refinement to remove noise, color adjustments, and combining multiple images.
2. Computer graphics is focused on constructing images, while digital image processing is focused on manipulating existing images.
3. Common digital image processing techniques discussed include image enhancement to improve image quality, image restoration to remove degradation, image segmentation to separate objects, image resizing, compression, and feature extraction.
4. Image filtering is used to reduce noise in images using techniques like convolution with filters that target different image frequency ranges like low-pass and
This document discusses the basics of computer graphics. It outlines the advantages of computer graphics such as producing high quality images and animation. It also classifies computer graphics systems as either interactive or passive. Interactive systems allow two-way communication between the user and computer while passive systems do not. The document then discusses pixels, color depth, frame buffers, and monitors. It concludes by outlining major areas of computer graphics like display of information, design/modeling, simulation, and user interfaces.
Digital image processing involves performing operations on digital images using computer algorithms. It has several functional categories including image restoration to remove noise and distortions, enhancement to modify the visual impact, and information extraction to analyze images. The main steps are acquisition, enhancement, restoration, color processing, compression, segmentation, and filtering using techniques like pixelization, principal components analysis, and neural networks. It has applications in medical imaging, film, transmission, sensing, and robotics. The advantages are noise removal, flexibility in format and manipulation, and easy storage and retrieval. The disadvantages can include high initial costs and potential data loss if storage devices fail.
This document outlines the syllabus for the course IT6005 - Digital Image Processing. The syllabus is divided into 5 units that cover digital image fundamentals, image enhancement, image restoration and segmentation, wavelets and image compression, and image representation and recognition. Unit 1 introduces key concepts in digital image processing such as pixels, gray levels, sampling and quantization. It also provides a brief history of the origin and development of digital image processing.
This document summarizes various topics related to image processing including image data types, file formats, acquisition, storage, processing, communication, display, and enhancement techniques. It discusses key concepts such as image fundamentals, color models, resolution, bit depth, file formats like JPEG, GIF, TIFF, compression techniques including lossless, lossy, intraframe, interframe, and algorithms like run length encoding and Shannon-Fano coding. Image enhancement topics covered are point processing, spatial filtering, and color image processing.
The Indian Dental Academy is the Leader in continuing dental education , training dentists in all aspects of dentistry and
offering a wide range of dental certified courses in different formats.for more details please visit
www.indiandentalacademy.com
This document provides an introduction to digital image processing. It defines what an image and digital image are, and discusses the first ever digital photograph. It describes digital image processing as processing digital images using computers, with sources including the electromagnetic spectrum from gamma rays to radio waves. Key concepts covered include digital images, image enhancement through spatial and frequency domain methods, image restoration to remove noise and blurring, and image compression to reduce file size through removing different types of data redundancy.
Discover the fundamentals, Characteristics & types of digital image analysis. Learn about pixels, bit depth, challenges, and AI impacts on image processing.
computervision1.pdf it is about computer visionshesnasuneer
This document provides an introduction to digital image processing and computer vision. It discusses how images are represented digitally through sampling and quantization. Low-level image processing techniques like preprocessing, segmentation, and object description are used to simplify computer vision tasks. Fundamental concepts in digital image processing are also introduced, such as how images can be represented as functions and processed using mathematical tools like the Fourier transform and convolution.
The document discusses fundamentals of digital images including representation as pixels, color models like RGB and CMYK, color depth, resolution, and file formats. It also covers topics like dithering, 2D graphics as vector or raster, and image compression standards. Key aspects covered include how pixels and bit depth determine color representation, uses of RGB vs CMYK color schemes, and how dithering creates illusions of additional colors through pixel arrangement.
Optical Watermarking Literature survey....Arif Ahmed
1. The document discusses optical watermarking, a technique for protecting copyright of real-world objects like artwork by embedding watermarks in the illumination of the objects. When photos are taken of the illuminated objects, the watermarks are captured in the digital images and can be extracted.
2. Optical watermarking works by transforming binary data into patterns of light projected onto objects. The patterns differ based on 1s and 0s in the data. When photos of the illuminated objects are taken, the patterns can be read from the captured images. Visible light is used with fine or low-contrast patterns to make the watermarks imperceptible.
3. Optical watermarking provides copyright protection for valuable
This document provides an overview of a digital image processing course. It outlines 4 course outcomes: 1) describing basic concepts and applications of image processing, 2) describing techniques in color, segmentation, and recognition, 3) illustrating pixel relationships and image arithmetic, and 4) analyzing digital image enhancement principles. The document then discusses topics that will be covered in the course, including image types, operations, and applications in various fields.
Digital image processing involves converting analog images to digital format using techniques like sampling and quantization. Sampling converts continuous images to discrete pixel values that define the image's spatial resolution. Quantization divides signal amplitudes into quanta to derive values that determine image quality. Together, sampling and quantization are key to converting images to digital format and improving qualities like sharpness, brightness, and clarity.
The document discusses image processing and provides details about:
1) The stages of image processing - input, editing, and output. The input stage deals with converting analog images to digital form. The editing stage manipulates the image. The output stage saves the transformed image.
2) Image processing operations like geometric transformations, color corrections, digital compositing, and extending dynamic range.
3) Applications of image processing like face detection, medical imaging, and remote sensing.
This document provides an overview of image processing and related concepts. It discusses:
1) What an image is, the basic steps of image processing like acquisition, analysis and output, and the two main types: analogue and digital.
2) Key aspects of digital image processing including what a digital image and pixel are, and common processing steps of pre-processing, enhancement and extraction.
3) Common image processing techniques like enhancement which adjusts images, and segmentation which partitions an image into meaningful segments.
4) Image classification which predicts categories from inputs, and the main types of supervised and unsupervised classification. It provides an example using Landsat satellite images.
This document provides an overview of image analysis, including:
1) It defines image analysis and discusses its use in recognizing, differentiating, and quantifying images across various fields including food quality assessment.
2) It describes the process of creating a digital image through digitization and discusses key aspects of digital images like resolution, pixel bit depth, and color.
3) It outlines common image processing actions like compression, preprocessing, and analysis and provides examples of applying image analysis to evaluate food products.
YCIS_Forensic PArt 1 Digital Image Processing.pptxSharmilaMore5
Basics of Digital Image Processing
Use of DIP in Society
Digital Image Processing Process
Why do we process images?
Image Enhancement and Edge detection
Python
How are we using Python in DIP
Biomedical Image Processing
Topics covered: Biomedical imaging, Need of image processing in medicine, Principles of image processing, Components of image processing, Application of image processing in different medical imaging systems
This document is a mini project report on digital image processing using MATLAB. It discusses various image processing techniques and applications implemented in MATLAB, including image formats, operations, and tools. Applications demonstrated include text recognition, color tracking, solving an engineering problem using image processing, creating a virtual slate using laser tracking, face detection, and distance estimation. The report provides examples of MATLAB functions used for tasks like importing, displaying, converting and cropping images, as well as analyzing and manipulating them.
Digital image processing & computer graphicsAnkit Garg
Digital Image Processing & Computer Graphics document discusses several topics related to digital image processing including:
1. Digital image processing involves manipulating digital images using computer programs. It includes operations like geometric transformations, image refinement to remove noise, color adjustments, and combining multiple images.
2. Computer graphics is focused on constructing images, while digital image processing is focused on manipulating existing images.
3. Common digital image processing techniques discussed include image enhancement to improve image quality, image restoration to remove degradation, image segmentation to separate objects, image resizing, compression, and feature extraction.
4. Image filtering is used to reduce noise in images using techniques like convolution with filters that target different image frequency ranges like low-pass and
This document discusses the basics of computer graphics. It outlines the advantages of computer graphics such as producing high quality images and animation. It also classifies computer graphics systems as either interactive or passive. Interactive systems allow two-way communication between the user and computer while passive systems do not. The document then discusses pixels, color depth, frame buffers, and monitors. It concludes by outlining major areas of computer graphics like display of information, design/modeling, simulation, and user interfaces.
Digital image processing involves performing operations on digital images using computer algorithms. It has several functional categories including image restoration to remove noise and distortions, enhancement to modify the visual impact, and information extraction to analyze images. The main steps are acquisition, enhancement, restoration, color processing, compression, segmentation, and filtering using techniques like pixelization, principal components analysis, and neural networks. It has applications in medical imaging, film, transmission, sensing, and robotics. The advantages are noise removal, flexibility in format and manipulation, and easy storage and retrieval. The disadvantages can include high initial costs and potential data loss if storage devices fail.
Similaire à Introduction to Image Processing_Lecture01 (20)
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Sinan KOZAK
Sinan from the Delivery Hero mobile infrastructure engineering team shares a deep dive into performance acceleration with Gradle build cache optimizations. Sinan shares their journey into solving complex build-cache problems that affect Gradle builds. By understanding the challenges and solutions found in our journey, we aim to demonstrate the possibilities for faster builds. The case study reveals how overlapping outputs and cache misconfigurations led to significant increases in build times, especially as the project scaled up with numerous modules using Paparazzi tests. The journey from diagnosing to defeating cache issues offers invaluable lessons on maintaining cache integrity without sacrificing functionality.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELijaia
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
Generative AI Use cases applications solutions and implementation.pdfmahaffeycheryld
Generative AI solutions encompass a range of capabilities from content creation to complex problem-solving across industries. Implementing generative AI involves identifying specific business needs, developing tailored AI models using techniques like GANs and VAEs, and integrating these models into existing workflows. Data quality and continuous model refinement are crucial for effective implementation. Businesses must also consider ethical implications and ensure transparency in AI decision-making. Generative AI's implementation aims to enhance efficiency, creativity, and innovation by leveraging autonomous generation and sophisticated learning algorithms to meet diverse business challenges.
https://www.leewayhertz.com/generative-ai-use-cases-and-applications/
Home security is of paramount importance in today's world, where we rely more on technology, home
security is crucial. Using technology to make homes safer and easier to control from anywhere is
important. Home security is important for the occupant’s safety. In this paper, we came up with a low cost,
AI based model home security system. The system has a user-friendly interface, allowing users to start
model training and face detection with simple keyboard commands. Our goal is to introduce an innovative
home security system using facial recognition technology. Unlike traditional systems, this system trains
and saves images of friends and family members. The system scans this folder to recognize familiar faces
and provides real-time monitoring. If an unfamiliar face is detected, it promptly sends an email alert,
ensuring a proactive response to potential security threats.
1. Digital Image processing
What is an image?
An image is defined as a two-dimensional function, F(x,y), where x and y are
spatial coordinates, and the amplitude of F at any pair of coordinates (x,y) is
called the intensity of that image at that point. When x,y, and amplitude values
of F are finite, we call it a digital image.
In other words, an image can be defined by a two-dimensional array specifically arranged in rows and columns.
Digital Image is composed of a finite number of elements, each of which elements have a particular value at a
particular location.These elements are referred to as picture elements ,image elements,and pixels.
A Pixel is most widely used to denote the elements of a Digital Image
2. Image as a Matrix
As we know, images are represented in rows and columns we have the
following syntax in which images are represented:
The right side of this equation is digital image by definition. Every element of this matrix
is called image element , picture element , or pixel.
3. Introduction to the Image processing
□ Image processing is a method to perform some operations on an image, in
order to get an enhanced image or to extract some useful information from
it. It is a type of signal processing in which input is an image and output
may be image or characteristics/features associated with that image.
□ Image processing basically includes the following three steps:
▪ Importing the image via image acquisition tools
▪ Analyzing and manipulating the image;
▪ Output in which result can be altered image or report that is based on
image analysis.
4. Introduction to the Image processing
There are two types of methods used for image
processing namely, analogue and digital image processing.
□ Analogue image processing can be used for the hard copies like printouts
and photographs. Image analysts use various fundamentals of
interpretation while using these visual techniques.
□ The analog image processing is applied on analog signals and it processes
only two-dimensional signals. The images are manipulated by electrical
signals. In analog image processing, analog signals can be periodic or
non-periodic
Examples of analog images are television images, photographs, paintings,
5. Introduction to the Image processing
□ Digital image processing techniques help in manipulation of the digital
images by using computers.
□ A digital image processing is applied to digital images (a matrix of small
pixels and elements). For manipulating the images, there is a number of
software and algorithms that are applied to perform changes. Digital image
processing is one of the fastest growing industry which affects everyone's
life.
Examples of digital images are color processing, image recognition, video
processing, etc
7. Analog Image Processing vs. Digital Image Processing
quality of images.
Analog Image Processing Digital Image Processing
The analog image processing is applied on The digital image processing is applied to digital
analog signals and it processes only signals that work on analyzing and manipulating the two-
dimensional signals. images.
Analog signal is time-varying signals so the It improves the digital quality of the image and
images formed under analog image intensity distribution is perfect in it.
processing get varied.
Analog image processing is a slower and Digital image processing is a cheaper and fast image
costlier process. storage and retrieval process.
Analog signal is a real-world but not good It uses good image compression techniques that
reduce the amount of data required and produce good
quality of images
into tiny components.
It is generally continuous and not broken It uses an image segmentation technique which is used
to detect discontinuity which occurs due to a broken
connection path.
8. Introduction to the DIP
□ Digital image processing deals with manipulation of digital images
through a digital computer. It is a subfield of signals and systems but focus
particularly on images. DIP focuses on developing a computer system that
is able to perform processing on an image.
□ Digital Image Processing is a software which is used in image processing.
For example: computer graphics, signals, photography, camera
mechanism, pixels, etc.
□ Digital Image Processing provides a platform to perform various
operations like image enhancing, processing of analog and digital signals,
image signals, voice signals etc.
11. Different Dimention
1-D Pictures: One-dimensional pictures are those containing only one dimension.
This is only possible when you're dealing with a line, as the only dimension you
have is length.
2-D Pictures: One type of picture you can come across in real life is the two-
dimensional one. The two dimensions depicted are length and width and the
objects on the picture are flat.
3-D Pictures: Three-dimensional pictures contain yet another dimension: depth.
This type is the most realistic one, as the depiction of objects or environments
resembles the way we see them through our own eyes
13. Concept of Pixel
The full form of the pixel is "Picture Element." It is also known as "PEL." Pixel
is the smallest element of an image on a computer display, whether they are
LCD or CRT monitors. A screen is made up of a matrix of thousands or
millions of pixels. A pixel is represented with a dot or a square on a computer
screen.
14. Concept of Pixel
Each pixel has a value, or we can say a unique logical address. It can have only one color at
a time. Colour of a pixel is determined by the number of bits which is used to represent it. A
resolution of a computer screen depends upon graphics card and display monitor, the
quantity, size and color combination of pixels.
17. Calculation of the total number of pixels
Below is the formula to calculate the total number of pixel in an
image.
For example: let rows=300 & columns=200
Total number of pixels= 300 X 200
= 500
18. Concept of Pixel
Pixel value(0)
As we know that each pixel has a unique value. 0 is a unique value that means the absence
of light. It means that 0 is used to denote dark.
We have a matrix of 3X3 of an image, and each pixel is of value as shown below:
It means the image formed is of 9 pixels which are black.
The image would be something as below:
19. Types of an image
BINARY IMAGE– The binary image as its name suggests, contain only two pixel
elements i.e 0 & 1,where 0 refers to black and 1 refers to white. This image is also known
as Monochrome.
It is the simplest type of image. It takes only two values i.e, Black and White or 0 and 1.
The binary image consists of a 1-bit image and it takes only 1 binary digit to represent a
pixel. Binary images are mostly used for general shape or outline.
20. Types of an image
Gray-scale images
Grayscale images are monochrome images, Means they have only one color. Grayscale
images do not contain any information about color. Each pixel determines available
different grey levels.
A normal grayscale image contains 8 bits/pixel data, which has 256 different grey levels.
21. Types of an image
Colour images
Colour images are three band monochrome images in which, each band contains a different
color and the actual information is stored in the digital image. The color images contain
gray level information in each spectral band.
The images are represented as red, green and blue (RGB images). And each color image
has 24 bits/pixel means 8 bits for each of the three color band(RGB).
23. Pixels, Dots, and Lines per Inch
Megapixels
A digital camera uses pixel elements (also known as a pixel) to
capture images. More number of pixels makes the better
resolution of that image.
Formula to calculate megapixel of a camera:
24. Pixels, Dots, and Lines per Inch
Example:
Let's take dimension of an image: 2500 X 3192
So, according to formula
(2500 * 3192) / 1 Million = 7.9 = 8 megapixel
(approx).
25. Pixels, Dots, and Lines per Inch
Aspect ratio:
Aspect ratio is the ratio between height and width of an image. A colon is used to
separate two numbers. Different images on different screens have a different
ratio.
Essentially, it describes an image's shape. Aspect ratios are written as a formula
of width to height, like this: 3:2. For example, a square image has an aspect ratio
of 1:1, since the height and width are the same.
Common aspect ratio is as follows:
1.33:1, 1.37:1, 1.43:1, 1.50:1, 1.56:1, 1.66:1, 1.75:1, 1.78:1, 1.85:1, 2.00:1, e.t.c
Advantage
1. It maintains balance in the appearance of an image (ratio between horizontal and vertical
pixel).
2. When the aspect ratio is increased, images do not distort.
26. Pixels, Dots, and Lines per Inch
Math-01: Here is a grayscale image given with aspect ratio 2:5, and the resolution of the image is 480000
pixels.
Calculate the following things:
□ Dimensions of the image.
□ R =?
□ C =?
□ Size of the image.
27. Pixels, Dots, and Lines per Inch
Comparing equation 1 and equation 2