Podcast
Questions and Answers
Which factors contribute to the values of discrete color or intensity in photometric image formation?
Which factors contribute to the values of discrete color or intensity in photometric image formation?
- Lighting and surface properties only
- Geometry and projection, camera optics and sensor properties, and lighting and surface properties (correct)
- Camera optics and sensor properties only
- Geometry and projection only
In the simplified model of photometric image formation, what key aspect is often ignored?
In the simplified model of photometric image formation, what key aspect is often ignored?
- The presence of light sources
- The reflection of light from an object's surface
- Multiple reflections of light in real-world scenes (correct)
- The camera's ability to capture light
What property defines point light sources in photometric image formation?
What property defines point light sources in photometric image formation?
- Radiating light uniformly in all directions (correct)
- Producing polarized light
- Having a variable intensity based on distance
- Emitting light in a concentrated beam
Which property of light is described as Watts / Area * Solid Angle?
Which property of light is described as Watts / Area * Solid Angle?
Which of the following best describes the "Specular" component in a simplified BRDF model?
Which of the following best describes the "Specular" component in a simplified BRDF model?
In the context of the pinhole camera model, if the focal length (f) is 50mm, and the image plane distance (u) is 25mm, what does the ratio Z/X represent?
In the context of the pinhole camera model, if the focal length (f) is 50mm, and the image plane distance (u) is 25mm, what does the ratio Z/X represent?
What is the effect of increasing the aperture size on the "circle of confusion" in an image?
What is the effect of increasing the aperture size on the "circle of confusion" in an image?
What process involves converting continuous image data into discrete values?
What process involves converting continuous image data into discrete values?
In the context of images as functions, what does f(x, y)
typically represent?
In the context of images as functions, what does f(x, y)
typically represent?
If a digital image has a width of 1920 pixels and a height of 1080 pixels, what do the variables 'a', 'b', 'c', and 'd' represent in defining the image as a function?
If a digital image has a width of 1920 pixels and a height of 1080 pixels, what do the variables 'a', 'b', 'c', and 'd' represent in defining the image as a function?
How is a digital image fundamentally represented?
How is a digital image fundamentally represented?
Which of the following best describes the purpose of the Bayer color filter array?
Which of the following best describes the purpose of the Bayer color filter array?
In the context of color sensing using a Bayer grid, what process is applied to estimate the missing color values at each pixel?
In the context of color sensing using a Bayer grid, what process is applied to estimate the missing color values at each pixel?
What information does imread(filename)
return in OpenCV?
What information does imread(filename)
return in OpenCV?
What is the effect of applying im2double
to an image in MATLAB?
What is the effect of applying im2double
to an image in MATLAB?
Which color space is described as being 'not perceptual' and having grays that 'live' along the diagonal?
Which color space is described as being 'not perceptual' and having grays that 'live' along the diagonal?
In the HSV color space, what does the 'Saturation' component represent?
In the HSV color space, what does the 'Saturation' component represent?
What is the functionality of cv::imread
in the OpenCV library?
What is the functionality of cv::imread
in the OpenCV library?
What parameters are required to draw a line using cv.line
in OpenCV?
What parameters are required to draw a line using cv.line
in OpenCV?
Which function parameter specifies the font type when adding text to an image using OpenCV?
Which function parameter specifies the font type when adding text to an image using OpenCV?
If you access a pixel in a color BGR image using row and column coordinates in OpenCV, what does the result represent?
If you access a pixel in a color BGR image using row and column coordinates in OpenCV, what does the result represent?
What information does the img.shape
attribute provide in OpenCV?
What information does the img.shape
attribute provide in OpenCV?
What is the purpose of the cv2.split()
function in OpenCV?
What is the purpose of the cv2.split()
function in OpenCV?
What is the purpose of image filtering?
What is the purpose of image filtering?
What type of transformation is represented by an affine transformation consisting of translation, rotation, and scaling?
What type of transformation is represented by an affine transformation consisting of translation, rotation, and scaling?
What is a key difference between a linear transformation and other transformations when represented with matrices?
What is a key difference between a linear transformation and other transformations when represented with matrices?
In image transformation, what does 'uniform scaling' imply?
In image transformation, what does 'uniform scaling' imply?
In 2D transformations, what operation does the matrix $\begin{bmatrix} cos(\theta) & -sin(\theta) \ sin(\theta) & cos(\theta) \end{bmatrix}$ perform?
In 2D transformations, what operation does the matrix $\begin{bmatrix} cos(\theta) & -sin(\theta) \ sin(\theta) & cos(\theta) \end{bmatrix}$ perform?
What is the inverse of a rotation transformation matrix R?
What is the inverse of a rotation transformation matrix R?
Which of the following transformations can be represented by a 2x2 matrix?
Which of the following transformations can be represented by a 2x2 matrix?
Which transformation can NOT be achieved through a 2x2 matrix alone?
Which transformation can NOT be achieved through a 2x2 matrix alone?
What is the key purpose of using homogeneous coordinates in image transformations?
What is the key purpose of using homogeneous coordinates in image transformations?
What transformation does the following matrix achieve? $\begin{bmatrix} 1 & 0 & t_x \ 0 & 1 & t_y \ 0 & 0 & 1 \end{bmatrix}$
What transformation does the following matrix achieve? $\begin{bmatrix} 1 & 0 & t_x \ 0 & 1 & t_y \ 0 & 0 & 1 \end{bmatrix}$
Which statement is true regarding affine transformation?
Which statement is true regarding affine transformation?
What function is used to take a Numpy array of type np.float32 in order to pass it into a cv.warpAffine()
function?
What function is used to take a Numpy array of type np.float32 in order to pass it into a cv.warpAffine()
function?
Flashcards
What is Photometric Image Formation?
What is Photometric Image Formation?
Photometric Image Formation refers to how images are created based on light, color, and intensity values.
How are images formed?
How are images formed?
Light is emitted from a light source, hits a surface, interacts with the surface, reflected light enters camera aperture, and the sensor of camera interprets light.
Point light sources
Point light sources
Point light sources radiate (emits) light uniformly in all directions.
Properties of light
Properties of light
Signup and view all the flashcards
Surface orientation
Surface orientation
Signup and view all the flashcards
Bi-direction reflectance function
Bi-direction reflectance function
Signup and view all the flashcards
Simplified BRDF two components
Simplified BRDF two components
Signup and view all the flashcards
Returned light and foreshortening
Returned light and foreshortening
Signup and view all the flashcards
Visible Light Spectrum
Visible Light Spectrum
Signup and view all the flashcards
Lens Equation
Lens Equation
Signup and view all the flashcards
What creates a “blur circle”
What creates a “blur circle”
Signup and view all the flashcards
Allowable size of the blur circle.
Allowable size of the blur circle.
Signup and view all the flashcards
Sampling and Quantization
Sampling and Quantization
Signup and view all the flashcards
How to create digital (discrete) images
How to create digital (discrete) images
Signup and view all the flashcards
Images as functions
Images as functions
Signup and view all the flashcards
Digital image range
Digital image range
Signup and view all the flashcards
How can a digital image represented
How can a digital image represented
Signup and view all the flashcards
Practical Color Sensing: Bayer Grid
Practical Color Sensing: Bayer Grid
Signup and view all the flashcards
What is a system in image terms?
What is a system in image terms?
Signup and view all the flashcards
What do lenses do?
What do lenses do?
Signup and view all the flashcards
Thin lens assumption
Thin lens assumption
Signup and view all the flashcards
How to transform an image?
How to transform an image?
Signup and view all the flashcards
What is translation?
What is translation?
Signup and view all the flashcards
What is OpenCV?
What is OpenCV?
Signup and view all the flashcards
What is cv::imread?
What is cv::imread?
Signup and view all the flashcards
What is cv::imshow?
What is cv::imshow?
Signup and view all the flashcards
What is cv::imwrite?
What is cv::imwrite?
Signup and view all the flashcards
How to draw a line?
How to draw a line?
Signup and view all the flashcards
cv::rectangle
cv::rectangle
Signup and view all the flashcards
What does cv::circle need?
What does cv::circle need?
Signup and view all the flashcards
Adding Text to Images
Adding Text to Images
Signup and view all the flashcards
Image properties
Image properties
Signup and view all the flashcards
Image filtering and warpping
Image filtering and warpping
Signup and view all the flashcards
Parametric (global) warping
Parametric (global) warping
Signup and view all the flashcards
What are Affine Transformations?
What are Affine Transformations?
Signup and view all the flashcards
Rotation
Rotation
Signup and view all the flashcards
Study Notes
- Photometric Image Formation refers to how images are created based on light, color, and intensity values
- Discrete color or intensity values, geometry & projection
- Camera optics & sensor properties, lighting & surface properties must be considered for image formation.
- In Photometric image formation, light is emitted from a source, strikes a surface, interacts with the surface, and then reflected light enters a camera aperture
How are images formed?
- Light is emitted from a light source
- Light hits a surface
- Light interacts with the surface
- Reflected light enters a camera aperture
- Sensor of camera interprets light
Simplified Photometric image formation
- Simplified model emits light from one or more light sources
- Light is then reflected from an object’s surface
- A portion of this light is directed towards the camera
- This simplified model ignores multiple reflections, which often occur in real-world scenes.
Light Emission
- Point light sources radiate (emit) light uniformly in all directions.
- Properties of light include color spectrum (wavelength distribution) and intensity (Watts / Area * Solid Angle)
- The visible light spectrum is the section of the electromagnetic radiation spectrum that is visible to the human eye, ranging from 435-380 to 740-625 nanometers.
- A solid angle is like a cone
- “Area” light sources, like fluorescent lights, are a little different.
Light Interaction with a Surface
- Some light is absorbed by the surface color, while some is reflected.
- The orientation of a surface is defined by its "normal vector" which sticks straight up out of the surface.
- The bidirectional reflectance distribution function (BRDF) expresses the amount, direction, and color spectrum of reflected light.
- BRDF depends on the amount, direction, and color spectrum of incoming light
- Simplified BRDF is modeled with "Lambertian", "flat" or "matte" component where light is radiated equally in all directions.
- Also models, "Specular", "shiny", or "highlight" component where radiated light is reflected across the normal from the incoming light
- The diminution of returned light depends on the cosine of the angle between the incident light direction and the surface normal
- Mirror (specular) reflection has the incident light ray direction reflected onto the specular direction around the surface normal
Light Entering a Camera
- Red triangle (behind camera) and blue triangle (in front of camera) are similar
- The relationship between focal length (f), image distance (u) and object distance (X), is shown by: f/u = Z/X, where Z is the scene depth
- With any three terms, the fourth can be determined
- For a given focal length, “Lens Equation” is represented by 1/z₀ + 1/zᵢ = 1/f
- From the lens equation, Zi = 1 / (1/f - 1/Zo)
- When projections of objects aren't focused on the image plane, a "blur circle" or "circle of confusion" results
- The size of the blur circle depends on the distance to the object and the size of the aperture
- The allowable size of the blur circle (e.g., a pixel) determines the allowable range of depths in the scene ("depth of field")
- Images are formed through Film, Digital cameras, and the Eye
Sensor Arrays
- Images from Sensors are created through Continuos image projected onto a sensor array; Result of image sampling and quantization when generating a digital image
- The basic process of generating a digital image includes taking a scan line from the continuous image, illustrating sampling and quantization
- Sampling is taking sample digital representations and quantizing by assigning values.
Images as Functions
- An image can be thought of as a function, f, from R² to R
- f(x, y) gives the intensity at position (x, y).
- With images, expect the image only to be defined over a rectangle, with a finite range: f : [a,b]x[c,d] → [0,1]
- A color image is just three functions pasted together, expressed as a vector-valued function: f(x,y) = r(x,y) g(x,y) b(x,y)
Digital Images
- (x, y) are the spatial coordinates in an image
- The values a and b define the range of x (horizontal) coordinates.
- The values c and d define the range of y (vertical) coordinates.
- A finite rectangle [a, b] x [c, d] represents where the image exists.
- The function f(x, y) is undefined outside this rectangular region because the image doesn't exist beyond its borders.
- A digital image with width = 1920 pixels and height = 1080 pixels is defined as: x ranging from a = 0 to b = 1919, y ranging from c = 0 to d = 1079
- Each pixel at (x, y) has an intensity value (for grayscale) or RGB values (for color images).
- Digital (discrete) images are used to Sample the 2D space on a regular grid, and quantize each sample (round to the nearest integer)
- If our samples are Δ apart, then f[i,j] = Quantize{f(iΔ, jΔ)}
- The values can be represented as a matrix of integer values.
- Bayer color filter arrays capture the full colors in an image
Visible light
- EM energy from the sun, in specific wavelengths, is what is identified as colors.
Color Sensing
- RGB values are estimated at 'G' cells from neighboring values in the Bayer grid
Image representation
- Images are represented as a matrix for NxM RGB images.
- im(1,1,1) = top-left pixel value in R-channel
- im(y, x, b) = y pixels down, x pixels to the right in the bth channel
- im(N, M, 3) = bottom-right pixel in B-channel
- imread(filename) returns a uint8 image (values 0 to 255) that is convertible to a double format (values 0 to 1) with im2double
Color space
- Colors are represented in multiple combinations of Red, Green and Blue.
RGB Colour Space
- RGB color space is easy for devices, but not perceptual
- RGB does not easily identify properties such as, where do the grays live or where is hue and saturation.
CMYK Color Space
- CMYK works like paint and includes white is background color (all colors)
- Opposes the RGB color space
HSV Color Space
- A perceptual dimensions of the color space includes Hue ( "kind" of color), Saturation, and Value (total amount of light)
- Use rgb2hsv() and hsv2rgb() in Matlab, in Python w/skimage
Linear Systems
- Linear system is defined as a unit that converts an input function f(x) into an output (or response) function g(x) where x is an independent variable like time, in the case of images it's spatial position
- Assume for simplicity that x is a continuous variable, results derived should be equally applicable to discrete variables.
Lenses
- A lens focuses parallel rays onto a single focal point.
- Focal point is at a distance f beyond the plane of the lens, where f is a function of the shape and index of refraction of the lens
- Aperture of diameter D restricts the range of rays, which may be on either side of the lens
- Lenses are typically spherical (easier to produce).
- Real cameras use many lenses together to correct for aberrations
- Thin lens equation defines the relationship between object and images as 1/d₀ + 1/dᵢ = 1/f
- Any object point satisfying this equation is in focus
- The thin lens assumption assumes the lens has no thickness, which isn't true.
Open CV Library
- Read an image from file using cv::imread
- Display an image in an OpenCV window using cv::imshow
- Write an image to a file using cv::imwrite
- Basic drawing functions include drawing lines, rectangles, and circles, and adding text to said objects
- Images are read as Numpy arrays and split into channels
- dogface = input_image[60:250, 70:350]
- plt.imshow(dogface)
- flipped_code_0=cv2.flip(input_image,0) # vertical flip
- plt.imshow(flipped_code_0)
- flipped_code_1=cv2.flip(input_image,1) # horizontal flip
- plt.imshow(flipped_code_1)
- transposed=cv2.transpose(input_image)
- plt.imshow(transposed)
Image transformation
- Image filtering involves changing the range of an image: expressed as g(x) = T(f(x))
- Image warping involves changing the domain of an image, expressed as g(x) = f(T(x))
Parametric (global) Warping
- Examples of parametric warps: translation, rotation, aspect, affine, perspective, cylindrical
- Transformation T is a coordinate-changing machine: p' = T(p)
- If T is global then it's the same for any point p and it can be described by just a few numbers (parameters)
- A linear T can be represented as a matrix: p' = Mp
Scaling
- Scaling a coordinate means multiplying each of its components by a scalar
- Uniform scaling means this scalar is the same for all components
- Non-uniform scaling uses different scalars per component
- In matrix form X'=aX and Y=bY and is expressed as a scaling matrix
2D Rotation
- The 2d rotation can be captured in matrix form where the non linear functions of is: x' = x cos(θ) - y sin(θ) + y cos(θ)
Homogenous coordinates
- In homogeneous coordinates, it's possible to add an extra dimension to 2D points.
- Point (x, y) in 2D becomes (x, y, 1) in homogeneous coordinates
- Now, translation can be written as a 3x3 matrix multiplication:
x = x+ t and y = y+ t
- a general 3x3 affine transformation matrix looks like this and its kept compatible with homogeneous coordinate systems
Studying That Suits You
Use AI to generate personalized quizzes and flashcards to suit your learning preferences.