Photometric Image Formation

Choose a study mode

Play Quiz
Study Flashcards
Spaced Repetition
Chat to Lesson

Podcast

Play an AI-generated podcast conversation about this lesson
Download our mobile app to listen on the go
Get App

Questions and Answers

Which factors contribute to the values of discrete color or intensity in photometric image formation?

  • Lighting and surface properties only
  • Geometry and projection, camera optics and sensor properties, and lighting and surface properties (correct)
  • Camera optics and sensor properties only
  • Geometry and projection only

In the simplified model of photometric image formation, what key aspect is often ignored?

  • The presence of light sources
  • The reflection of light from an object's surface
  • Multiple reflections of light in real-world scenes (correct)
  • The camera's ability to capture light

What property defines point light sources in photometric image formation?

  • Radiating light uniformly in all directions (correct)
  • Producing polarized light
  • Having a variable intensity based on distance
  • Emitting light in a concentrated beam

Which property of light is described as Watts / Area * Solid Angle?

<p>Intensity (D)</p> Signup and view all the answers

Which of the following best describes the "Specular" component in a simplified BRDF model?

<p>Radiated light is reflected across the normal from the incoming light (A)</p> Signup and view all the answers

In the context of the pinhole camera model, if the focal length (f) is 50mm, and the image plane distance (u) is 25mm, what does the ratio Z/X represent?

<p>The ratio of object depth to object size (B)</p> Signup and view all the answers

What is the effect of increasing the aperture size on the "circle of confusion" in an image?

<p>It increases the size, making the image more blurred if objects are not focused on the image plane (B)</p> Signup and view all the answers

What process involves converting continuous image data into discrete values?

<p>Quantization (A)</p> Signup and view all the answers

In the context of images as functions, what does f(x, y) typically represent?

<p>The intensity of the image at coordinates (x, y) (C)</p> Signup and view all the answers

If a digital image has a width of 1920 pixels and a height of 1080 pixels, what do the variables 'a', 'b', 'c', and 'd' represent in defining the image as a function?

<p>a and b represent the x-coordinate range, while c and d represent the y-coordinate range. (A)</p> Signup and view all the answers

How is a digital image fundamentally represented?

<p>As a matrix of integer values (B)</p> Signup and view all the answers

Which of the following best describes the purpose of the Bayer color filter array?

<p>To capture color information by filtering different wavelengths of light (B)</p> Signup and view all the answers

In the context of color sensing using a Bayer grid, what process is applied to estimate the missing color values at each pixel?

<p>Interpolation (B)</p> Signup and view all the answers

What information does imread(filename) return in OpenCV?

<p>A uint8 image (B)</p> Signup and view all the answers

What is the effect of applying im2double to an image in MATLAB?

<p>It converts the image to a double precision format with values between 0 and 1. (D)</p> Signup and view all the answers

Which color space is described as being 'not perceptual' and having grays that 'live' along the diagonal?

<p>RGB (D)</p> Signup and view all the answers

In the HSV color space, what does the 'Saturation' component represent?

<p>The purity or colorfulness of the color (B)</p> Signup and view all the answers

What is the functionality of cv::imread in the OpenCV library?

<p>Read an image from a file (D)</p> Signup and view all the answers

What parameters are required to draw a line using cv.line in OpenCV?

<p>Image, starting point, ending point, color, thickness (D)</p> Signup and view all the answers

Which function parameter specifies the font type when adding text to an image using OpenCV?

<p>A cv.FONT_* constant (A)</p> Signup and view all the answers

If you access a pixel in a color BGR image using row and column coordinates in OpenCV, what does the result represent?

<p>An array of Blue, Green, and Red values (B)</p> Signup and view all the answers

What information does the img.shape attribute provide in OpenCV?

<p>The number of rows, columns, and channels in the image (B)</p> Signup and view all the answers

What is the purpose of the cv2.split() function in OpenCV?

<p>To split an image into its color channels (C)</p> Signup and view all the answers

What is the purpose of image filtering?

<p>To change range of image (A)</p> Signup and view all the answers

What type of transformation is represented by an affine transformation consisting of translation, rotation, and scaling?

<p>A combination of linear transformation and translation (A)</p> Signup and view all the answers

What is a key difference between a linear transformation and other transformations when represented with matrices?

<p>Linear transformations can be fully described using a matrix multiplication (D)</p> Signup and view all the answers

In image transformation, what does 'uniform scaling' imply?

<p>Scaling each coordinate by the same factor (C)</p> Signup and view all the answers

In 2D transformations, what operation does the matrix $\begin{bmatrix} cos(\theta) & -sin(\theta) \ sin(\theta) & cos(\theta) \end{bmatrix}$ perform?

<p>Rotation (A)</p> Signup and view all the answers

What is the inverse of a rotation transformation matrix R?

<p>The transpose of the rotation matrix (Rᵀ) (B)</p> Signup and view all the answers

Which of the following transformations can be represented by a 2x2 matrix?

<p>Shear (D)</p> Signup and view all the answers

Which transformation can NOT be achieved through a 2x2 matrix alone?

<p>Translation (C)</p> Signup and view all the answers

What is the key purpose of using homogeneous coordinates in image transformations?

<p>To represent translation as a matrix operation (A)</p> Signup and view all the answers

What transformation does the following matrix achieve? $\begin{bmatrix} 1 & 0 & t_x \ 0 & 1 & t_y \ 0 & 0 & 1 \end{bmatrix}$

<p>Translation (A)</p> Signup and view all the answers

Which statement is true regarding affine transformation?

<p>Lines map to lines (D)</p> Signup and view all the answers

What function is used to take a Numpy array of type np.float32 in order to pass it into a cv.warpAffine() function?

<p>cv.warpAffine() (C)</p> Signup and view all the answers

Flashcards

What is Photometric Image Formation?

Photometric Image Formation refers to how images are created based on light, color, and intensity values.

How are images formed?

Light is emitted from a light source, hits a surface, interacts with the surface, reflected light enters camera aperture, and the sensor of camera interprets light.

Point light sources

Point light sources radiate (emits) light uniformly in all directions.

Properties of light

Color spectrum (Wavelength distribution) and Intensity (Watts / Area * Solid Angle)

Signup and view all the flashcards

Surface orientation

The orientation of a surface is defined by its normal vector which sticks straight up out of the surface.

Signup and view all the flashcards

Bi-direction reflectance function

The amount, direction, and color spectrum of reflected light depending on the amount, direction, and color spectrum of incoming light

Signup and view all the flashcards

Simplified BRDF two components

“Lambertian”, “flat” or “matte” component: light radiated equally in all directions; “Specular”, “shiny”, or “highlight” component: radiated light is reflected across the normal from the incoming light

Signup and view all the flashcards

Returned light and foreshortening

The diminution of returned light caused by foreshortening depends on viîn, the cosine of the angle between the incident light direction and the surface normal n.

Signup and view all the flashcards

Visible Light Spectrum

The visible light spectrum is the section of the electromagnetic radiation spectrum that is visible to the human eye.

Signup and view all the flashcards

Lens Equation

It is defined by the equation: 1/zo + 1/zi = 1/f

Signup and view all the flashcards

What creates a “blur circle”

A "blur circle" or "circle of confusion" results when projections of objects are not focused on the image plane. The size of the blur circle depends on the distance to the object and the size of the aperture.

Signup and view all the flashcards

Allowable size of the blur circle.

It determines the allowable range of depths in the scene

Signup and view all the flashcards

Sampling and Quantization

Process of converting a continuous image into a discrete digital image.

Signup and view all the flashcards

How to create digital (discrete) images

Sample the 2D space on a regular grid, Quantize each sample (round to nearest integer).

Signup and view all the flashcards

Images as functions

Point (x, y) coordinates correspond to locations on the image, and pixel values represent intensity at those locations.

Signup and view all the flashcards

Digital image range

The values a and b define the range of x (horizontal) coordinates; The values c and d define the range of y (vertical) coordinates

Signup and view all the flashcards

How can a digital image represented

Together, [𝑎,𝑏]×[𝑐,𝑑] represents a finite rectangle where the image exists. Outside this rectangular region, the function 𝑓(𝑥,𝑦) is undefined because the image doesn't exist beyond its borders.

Signup and view all the flashcards

Practical Color Sensing: Bayer Grid

It estimates RGB at 'G' cells from neighboring values

Signup and view all the flashcards

What is a system in image terms?

a unit that converts an input function f(x) into an output (or response) function g(x), where x is an independent variable, such as time or, as in the case of images, spatial position.

Signup and view all the flashcards

What do lenses do?

A lens focuses parallel rays onto a single focal point

Signup and view all the flashcards

Thin lens assumption

It assumes the lens has no thickness

Signup and view all the flashcards

How to transform an image?

Linear Transformations as Change of Basis

Signup and view all the flashcards

What is translation?

Translation is the shifting of an object's location

Signup and view all the flashcards

What is OpenCV?

OpenCV is a comprehensive open-source library for computer vision, image processing, and machine learning, offering a wide array of functions for tasks such as image reading, manipulation, and analysis.

Signup and view all the flashcards

What is cv::imread?

cv::imread reads an image from a specified file.

Signup and view all the flashcards

What is cv::imshow?

cv::imshow displays an image in an OpenCV window.

Signup and view all the flashcards

What is cv::imwrite?

cv::imwrite writes an image to a file.

Signup and view all the flashcards

How to draw a line?

Drawing a line needs line starting and ending coordinates.

Signup and view all the flashcards

cv::rectangle

To draw a rectangle, you need the top-left and bottom-right corners.

Signup and view all the flashcards

What does cv::circle need?

Drawing a circle needs a center coordinate & radius.

Signup and view all the flashcards

Adding Text to Images

For putting text on images, specify text data, position, font, scale.

Signup and view all the flashcards

Image properties

The shape of an image is accessed by img.shape. It returns a tuple of the number of rows, columns, and channels (if the image is color)

Signup and view all the flashcards

Image filtering and warpping

image filtering changes the range of an image, image warping changes the domain of an image

Signup and view all the flashcards

Parametric (global) warping

Parametric (global) warping is transformations that can be described by just a few numbers parameters and is the same for and point

Signup and view all the flashcards

What are Affine Transformations?

Affine transformations are combinations of translations, rotation, scale, and shearing.

Signup and view all the flashcards

Rotation

Rotation of an image for an angle is achieved by the transformation matrix of the form

Signup and view all the flashcards

Study Notes

  • Photometric Image Formation refers to how images are created based on light, color, and intensity values
  • Discrete color or intensity values, geometry & projection
  • Camera optics & sensor properties, lighting & surface properties must be considered for image formation.
  • In Photometric image formation, light is emitted from a source, strikes a surface, interacts with the surface, and then reflected light enters a camera aperture

How are images formed?

  • Light is emitted from a light source
  • Light hits a surface
  • Light interacts with the surface
  • Reflected light enters a camera aperture
  • Sensor of camera interprets light

Simplified Photometric image formation

  • Simplified model emits light from one or more light sources
  • Light is then reflected from an object’s surface
  • A portion of this light is directed towards the camera
  • This simplified model ignores multiple reflections, which often occur in real-world scenes.

Light Emission

  • Point light sources radiate (emit) light uniformly in all directions.
  • Properties of light include color spectrum (wavelength distribution) and intensity (Watts / Area * Solid Angle)
  • The visible light spectrum is the section of the electromagnetic radiation spectrum that is visible to the human eye, ranging from 435-380 to 740-625 nanometers.
  • A solid angle is like a cone
  • “Area” light sources, like fluorescent lights, are a little different.

Light Interaction with a Surface

  • Some light is absorbed by the surface color, while some is reflected.
  • The orientation of a surface is defined by its "normal vector" which sticks straight up out of the surface.
  • The bidirectional reflectance distribution function (BRDF) expresses the amount, direction, and color spectrum of reflected light.
  • BRDF depends on the amount, direction, and color spectrum of incoming light
  • Simplified BRDF is modeled with "Lambertian", "flat" or "matte" component where light is radiated equally in all directions.
  • Also models, "Specular", "shiny", or "highlight" component where radiated light is reflected across the normal from the incoming light
  • The diminution of returned light depends on the cosine of the angle between the incident light direction and the surface normal
  • Mirror (specular) reflection has the incident light ray direction reflected onto the specular direction around the surface normal

Light Entering a Camera

  • Red triangle (behind camera) and blue triangle (in front of camera) are similar
  • The relationship between focal length (f), image distance (u) and object distance (X), is shown by: f/u = Z/X, where Z is the scene depth
  • With any three terms, the fourth can be determined
  • For a given focal length, “Lens Equation” is represented by 1/z₀ + 1/zᵢ = 1/f
  • From the lens equation, Zi = 1 / (1/f - 1/Zo)
  • When projections of objects aren't focused on the image plane, a "blur circle" or "circle of confusion" results
  • The size of the blur circle depends on the distance to the object and the size of the aperture
  • The allowable size of the blur circle (e.g., a pixel) determines the allowable range of depths in the scene ("depth of field")
  • Images are formed through Film, Digital cameras, and the Eye

Sensor Arrays

  • Images from Sensors are created through Continuos image projected onto a sensor array; Result of image sampling and quantization when generating a digital image
  • The basic process of generating a digital image includes taking a scan line from the continuous image, illustrating sampling and quantization
  • Sampling is taking sample digital representations and quantizing by assigning values.

Images as Functions

  • An image can be thought of as a function, f, from R² to R
  • f(x, y) gives the intensity at position (x, y).
  • With images, expect the image only to be defined over a rectangle, with a finite range: f : [a,b]x[c,d] → [0,1]
  • A color image is just three functions pasted together, expressed as a vector-valued function: f(x,y) = r(x,y) g(x,y) b(x,y)

Digital Images

  • (x, y) are the spatial coordinates in an image
  • The values a and b define the range of x (horizontal) coordinates.
  • The values c and d define the range of y (vertical) coordinates.
  • A finite rectangle [a, b] x [c, d] represents where the image exists.
  • The function f(x, y) is undefined outside this rectangular region because the image doesn't exist beyond its borders.
  • A digital image with width = 1920 pixels and height = 1080 pixels is defined as: x ranging from a = 0 to b = 1919, y ranging from c = 0 to d = 1079
  • Each pixel at (x, y) has an intensity value (for grayscale) or RGB values (for color images).
  • Digital (discrete) images are used to Sample the 2D space on a regular grid, and quantize each sample (round to the nearest integer)
  • If our samples are Δ apart, then f[i,j] = Quantize{f(iΔ, jΔ)}
  • The values can be represented as a matrix of integer values.
  • Bayer color filter arrays capture the full colors in an image

Visible light

  • EM energy from the sun, in specific wavelengths, is what is identified as colors.

Color Sensing

  • RGB values are estimated at 'G' cells from neighboring values in the Bayer grid

Image representation

  • Images are represented as a matrix for NxM RGB images.
  • im(1,1,1) = top-left pixel value in R-channel
  • im(y, x, b) = y pixels down, x pixels to the right in the bth channel
  • im(N, M, 3) = bottom-right pixel in B-channel
  • imread(filename) returns a uint8 image (values 0 to 255) that is convertible to a double format (values 0 to 1) with im2double

Color space

  • Colors are represented in multiple combinations of Red, Green and Blue.

RGB Colour Space

  • RGB color space is easy for devices, but not perceptual
  • RGB does not easily identify properties such as, where do the grays live or where is hue and saturation.

CMYK Color Space

  • CMYK works like paint and includes white is background color (all colors)
  • Opposes the RGB color space

HSV Color Space

  • A perceptual dimensions of the color space includes Hue ( "kind" of color), Saturation, and Value (total amount of light)
  • Use rgb2hsv() and hsv2rgb() in Matlab, in Python w/skimage

Linear Systems

  • Linear system is defined as a unit that converts an input function f(x) into an output (or response) function g(x) where x is an independent variable like time, in the case of images it's spatial position
  • Assume for simplicity that x is a continuous variable, results derived should be equally applicable to discrete variables.

Lenses

  • A lens focuses parallel rays onto a single focal point.
  • Focal point is at a distance f beyond the plane of the lens, where f is a function of the shape and index of refraction of the lens
  • Aperture of diameter D restricts the range of rays, which may be on either side of the lens
  • Lenses are typically spherical (easier to produce).
  • Real cameras use many lenses together to correct for aberrations
  • Thin lens equation defines the relationship between object and images as 1/d₀ + 1/dᵢ = 1/f
  • Any object point satisfying this equation is in focus
  • The thin lens assumption assumes the lens has no thickness, which isn't true.

Open CV Library

  • Read an image from file using cv::imread
  • Display an image in an OpenCV window using cv::imshow
  • Write an image to a file using cv::imwrite
  • Basic drawing functions include drawing lines, rectangles, and circles, and adding text to said objects
  • Images are read as Numpy arrays and split into channels
  • dogface = input_image[60:250, 70:350]
  • plt.imshow(dogface)
  • flipped_code_0=cv2.flip(input_image,0) # vertical flip
  • plt.imshow(flipped_code_0)
  • flipped_code_1=cv2.flip(input_image,1) # horizontal flip
  • plt.imshow(flipped_code_1)
  • transposed=cv2.transpose(input_image)
  • plt.imshow(transposed)

Image transformation

  • Image filtering involves changing the range of an image: expressed as g(x) = T(f(x))
  • Image warping involves changing the domain of an image, expressed as g(x) = f(T(x))

Parametric (global) Warping

  • Examples of parametric warps: translation, rotation, aspect, affine, perspective, cylindrical
  • Transformation T is a coordinate-changing machine: p' = T(p)
  • If T is global then it's the same for any point p and it can be described by just a few numbers (parameters)
  • A linear T can be represented as a matrix: p' = Mp

Scaling

  • Scaling a coordinate means multiplying each of its components by a scalar
  • Uniform scaling means this scalar is the same for all components
  • Non-uniform scaling uses different scalars per component
  • In matrix form X'=aX and Y=bY and is expressed as a scaling matrix

2D Rotation

  • The 2d rotation can be captured in matrix form where the non linear functions of is: x' = x cos(θ) - y sin(θ) + y cos(θ)

Homogenous coordinates

  • In homogeneous coordinates, it's possible to add an extra dimension to 2D points.
  • Point (x, y) in 2D becomes (x, y, 1) in homogeneous coordinates
  • Now, translation can be written as a 3x3 matrix multiplication:

x = x+ t and y = y+ t

  • a general 3x3 affine transformation matrix looks like this and its kept compatible with homogeneous coordinate systems

Studying That Suits You

Use AI to generate personalized quizzes and flashcards to suit your learning preferences.

Quiz Team

Related Documents

More Like This

EET-314 Lighting Design Week 2
9 questions
Spectrometry and Chromophores Quiz
45 questions
Illuminotecnica Corso DAMS 2024/2025
51 questions
Use Quizgecko on...
Browser
Browser