Image Formation Basics - iffatAGheyas/computer-vision-handbook GitHub Wiki
🧱 Image Formation Basics
Understanding how images are captured and represented is fundamental to Computer Vision. This guide covers the essentials of image formation, from sensor capture to colour spaces.
1. 📷 How an Image is Captured
At its core, a digital image is a grid of light measurements captured by a camera sensor:
- A camera lens focuses incoming light rays onto an image sensor (e.g. CMOS or CCD).
- Each small element on the sensor is called a pixel (picture element).
- During exposure, each pixel measures the light intensity that hits it.
This process yields a 2D matrix of brightness values.
2. 🌗 Grayscale Image Formation
A grayscale image records only light intensity—no colour. Each pixel value ranges from 0 (black) to 255 (white), with intermediate values representing shades of grey.
Example (3×3 grayscale image matrix):
[ 34 120 200 ]
[ 45 180 210 ]
[ 90 100 255 ]
Example (single pixel):
Red: 120
Green: 80
Blue: 20
4. 🧰 Other Image Formats (You’ll Encounter Later)
- HSV (Hue, Saturation, Value) — better for colour-based processing
- YCrCb, Lab — used in broadcasting and professional vision pipelines
- Binary — simple two-level (e.g. thresholded) images
🔄 Summary
Term | Description |
---|---|
Pixel | Smallest unit of an image |
Grayscale | 2D matrix of brightness values (0–255) |
RGB Image | Stack of three grayscale matrices (R, G & B channels) |
Sensor | Hardware that captures light and converts it to data |
Resolution | Image dimensions in pixels (e.g. 1920 × 1080) |