[go: up one dir, main page]

0% found this document useful (0 votes)
411 views14 pages

Radiometry of Image Formation (Computer Vision)

The document discusses the physics of image formation in cameras. It explains that an image captures the brightness values at each pixel, which is proportional to the number of photons detected by the sensor. Radiance is the directional quantity of light traveling in a given direction per unit area. The image irradiance, or brightness, is proportional to the scene radiance in the direction of the camera. Inverting the physics of image formation to understand the 3D scene from a 2D image is challenging.

Uploaded by

Anukriti Bansal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
411 views14 pages

Radiometry of Image Formation (Computer Vision)

The document discusses the physics of image formation in cameras. It explains that an image captures the brightness values at each pixel, which is proportional to the number of photons detected by the sensor. Radiance is the directional quantity of light traveling in a given direction per unit area. The image irradiance, or brightness, is proportional to the scene radiance in the direction of the camera. Inverting the physics of image formation to understand the 3D scene from a 2D image is challenging.

Uploaded by

Anukriti Bansal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 14

Radiometry of Image Formation

Jitendra Malik
What is in an image?

The image is an array of brightness


values (three arrays for RGB images)
A camera creates an image …

The image I(x,y) measures how much light is captured at pixel (x,y)

We want to know
• Where does a point (X,Y,Z) in the world get imaged?
• What is the brightness at the resulting point (x,y)?
The pinhole camera models where a scene
point is projected

y
x
Now let us try to understand brightness at
a pixel (x,y) …

The image I(x,y) measures how much light is captured at pixel (x,y).
Proportional to the number of photons captured at the sensor
element (CCD/CMOS/Rod/cone/..) in a time interval.
Radiance is a directional quantity
Radiant power travelling in a given direction per unit
area (measured perpendicular to the direction of
travel) per unit solid angle
Image irradiance is proportional to scene
radiance in the direction of the camera
What causes the outgoing radiance at a scene patch?
What causes the outgoing radiance at a scene patch?

Two special cases:


• Specular surfaces - Outgoing radiance direction obeys angle of
incidence=angle of reflection, and co-planarity of incident &
reflected rays & the surface normal.
• Lambertian surfaces - Outgoing radiance same in all directions
The Lambertian model

We often model reflectance by a combination of a Lambertian term and a specular term. If we


want to be precise, we use a BRDF (Bidirectional Reflectance Distribution function) which is a 4D
function corresponding to the ratio of outgoing radiance in a particular direction to the
incoming irradiance in some other direction. This can be measured empirically.
Edges are important
• Edges are curves in the image, across which the brightness
changes “a lot”. These arise because of discontinuities in
reflectance, illumination or surface geometry.
Real world scenes have additional complexity…

• Objects are illuminated not just by light sources, but also by


reflected light from other surfaces. In computer graphics, ray
tracing and radiosity are techniques that address this issue.
• Shadows
Inverting the physics of image formation is hard

• Shape-from-shading (SFS) seeks to go from the measured


irradiance values in the image to the scene geometry,
reflectances and illumination that caused it.
• This is the inverse of the computer graphics rendering
problem where the goal is to produce the image, given the
scene.
• The inverse problem is much harder than the forward
problem; traditional SFS only works under gross simplifying
assumptions on the physics.
• Computer vision has been much more successful in exploiting
the geometry of image formation with multiple views.

You might also like