What is Image Processing?
Image processing is a method to perform operations on an image to enhance it or extract useful information. It is a type of signal processing where the input is an image, and the output may be either an image or characteristics/features associated with that image.
Goals of Image Processing
-
Image Enhancement: Improving visual appearance (e.g., contrast, sharpness)
-
Image Restoration: Removing noise or distortion
-
Image Compression: Reducing the amount of data required to represent an image
-
Feature Extraction: Identifying objects, edges, or patterns
-
Image Analysis: Understanding and interpreting image content
-
Object Recognition: Detecting and identifying objects in an image
What is an Image?
An image is a two-dimensional function f(x, y), where x and y are spatial coordinates, and f is the intensity (brightness or color) at that point. For digital images, both x, y, and f are finite and discrete.
Types of Image Representation
-
Spatial Domain Representation: Direct representation using pixel intensity values in a grid.
-
Frequency Domain Representation: Using transforms like Fourier to represent the image in terms of its frequency components.
Types of Images
-
Binary Image: Only black and white (pixel values: 0 or 1)
-
Grayscale Image: Shades of gray (pixel values: 0 to 255)
-
Color Image: Consists of multiple channels, commonly RGB (Red, Green, Blue)
-
Indexed Image: Uses a colormap or palette to store color information
Image Models
-
Geometric Model: Describes the shape and position of image elements.
-
Photometric Model: Describes the brightness/intensity or color of each point.
-
Color Models:
-
RGB: Red, Green, Blue components
-
HSV: Hue, Saturation, Value
-
YCbCr: Used in video compression
-
CMYK: Used in printing
Resolution
-
Spatial Resolution: Amount of detail in an image (measured in pixels)
-
Gray-level Resolution: Number of distinct gray levels available (e.g., 8-bit = 256 levels)
Image Size
-
Described in terms of width × height × number of channels (e.g., 512 × 512 × 3 for RGB)
2D Linear System
-
A 2D linear system in image processing refers to a system where the output image is a linear transformation of the input image, usually involving operations like convolution.
-
Linearity implies two properties:
-
Additivity:
T[f1 + f2] = T[f1] + T[f2]
-
Homogeneity (Scaling):
T[a·f] = a·T[f]
-
-
Spatial Invariance: The system's response doesn’t change when the input is shifted.
-
Example: Applying a kernel (filter) over an image using convolution is a classic example of a 2D linear system:
Luminance
-
The measured intensity of light emitted or reflected from a surface in a given direction.
-
Closely related to the perceived brightness, but it's a physical quantity.
-
Important in grayscale and color image processing.
Contrast
-
The difference in luminance or color that makes an object distinguishable from others or the background.
-
High contrast makes features pop; low contrast makes the image appear flat.
-
Often enhanced using techniques like contrast stretching or histogram equalization.
Brightness
-
A subjective visual perception of how much light an image appears to emit or reflect.
-
Can be increased by adding a constant to all pixel intensities.
Color Representation
Images can be represented using various color models, each suitable for different applications:
RGB (Red, Green, Blue)
-
Additive color model (used in screens).
-
Each color is a mix of Red, Green, and Blue components.
CMY/CMYK (Cyan, Magenta, Yellow, Key/Black)
-
Subtractive color model (used in printing).
HSV (Hue, Saturation, Value)
-
Hue: Color type (0° to 360°)
-
Saturation: Color purity
-
Value: Brightness of the color
YUV / YCbCr
-
Used in video processing.
-
Separates brightness (Y) from color information (U and V or Cb and Cr).
Visibility Functions
-
Visibility functions describe how sensitive the human eye is to different spatial frequencies.
-
The Contrast Sensitivity Function (CSF) is a common example. It shows that humans are:
-
Most sensitive to mid-range spatial frequencies
-
Less sensitive to very low or very high frequencies
-
-
Important in compression algorithms and display optimization.
Monochrome and Color Vision Models
Monochrome Vision Model
-
Uses only intensity (luminance) values.
-
No color, only grayscale from black to white.
-
Basis of early vision systems and useful in medical/scientific imaging.
Color Vision Model
-
Based on how the human eye perceives color using three types of cones:
-
L (long wavelengths) → Red
-
M (medium) → Green
-
S (short) → Blue
-
-
Color models (like RGB, HSV) are built around this biological model.
-
Opponent Process Theory: Human vision processes color differences (Red-Green, Blue-Yellow) rather than absolute colors.
Comments
Post a Comment