PowerPoint - People

Download Report

Transcript PowerPoint - People

Joint Noise Level Estimation from Personal Photo Collections
YiChang
1*,2
Shih
1
Kwatra
Vivek
Troy
1Google Research
1
Chinen
Hui
2MIT CSAIL
1
Fang
Sergey
*Internship work at Google
Goal
Contributions
ο‚§ Given a set of face images from the same person, taken
under different lighting and cameras, estimate the noise
levels in each image
ο‚§ Key observation: given two noisy images, the noise
levels are correlated if they share the same underlying
2
2
image content, since 𝜎1 βˆ’ 𝜎2 = π‘£π‘Žπ‘Ÿ[𝑰𝒏,𝟏 ] βˆ’ π‘£π‘Žπ‘Ÿ[𝑰𝒏,𝟐 ]
Overview
Starting from a face image collection:
ο‚§ Preprocess: geometrically and photometrically align the
images with affine transform and color match
ο‚§ We formulate the estimation as maximizing the joint
probability distribution between all images’ noise levels
ο‚§ The joint distribution is conditioned on the pair-wise
2
2
relative noise levels {πœŒπ‘–π‘— |πœŒπ‘–π‘— β‰œ πœŽπ‘– βˆ’ πœŽπ‘— }. We use a twostage optimization that first estimates {πœŒπ‘–π‘— }, then {πœŽπ‘– }
ο‚§ 𝑰𝑛 = π‘°π‘œπ‘Ÿπ‘–π‘” + 𝒏, i.i.d, zero mean. 𝜎 = noise level β‰œ 𝑠𝑑𝑑[𝒏]
ο‚§ This is difficult because we cannot decouple 𝒏 from 𝑰𝑛
1
Ioffe
ο‚§ Two-stage optimization:
οƒ˜ Estimating {πœŒπ‘–π‘— }: We take a patch-based method. We
first find the patch correspondence between π‘°π’Š and 𝑰𝒋 ,
βˆ—
then find the best estimated relative noise {πœŒπ‘–π‘— } from
the patch pairs.
βˆ—
2
2
βˆ—
οƒ˜ With {πœŒπ‘–π‘— }, estimate {πœŽπ‘– } by constraining πœŽπ‘– βˆ’ πœŽπ‘— = πœŒπ‘–π‘—
Pair-wise Relative Noise πœŒπ‘–π‘— Estimation
Results
Ground Truth Experiment and Comparison
ο‚§ The two faces are not perfectly aligned
ο‚§ We show one example below with estimated
noise levels and denoised result using BM3D +
our method for noise parameter
ο‚§ Add synthetic Gaussian noise with different parameters
ο‚§ Compare estimated noise levels and denoised result by BM3D
q
β€’ π‘π‘π‘ž = exp(βˆ’πœ…π‘π‘ž 𝒑1𝑝 βˆ’ 𝒑2π‘ž
2
Input
BM3D
Input
y=x Line (Ground Truth)
Liu et. al. (Mean)
Metric-Q (Mean)
Our Method (Mean)
BM3D
Mean PSNR (Metric-Q)
Mean PSNR (Our Method)
Mean PSNR(Best BM3D)
30
I2
βˆ—
𝜌12
=
Οƒ=23.2
Οƒ=14.5
𝑝,π‘ž π‘π‘π‘ž πœπ‘π‘ž
𝑝,π‘ž π‘π‘π‘ž
), confidence that (𝑝, π‘ž) is a true correspondence
26
40.00
22
18
14
10
6
ο‚§ More subjects
38.00
37.00
36.00
35.00
33.00
2
Οƒ=38.2
39.00
34.00
2
Οƒ=23.9
β€’ For computational efficiency, we selected the best 5 π‘ž s for each 𝑝
41.00
Mean PSNR
I1
ο‚§ Compute pair-wise relative noise by aggregating πœπ‘π‘ž :
q
Estimated Noise Sigma
ο‚§ We break down the image into patches, and
estimate the patch-wise relative noise
levels πœπ‘π‘ž by πœπ‘π‘ž β‰œ π‘£π‘Žπ‘Ÿ[𝒑1𝑝 ] βˆ’ π‘£π‘Žπ‘Ÿ[𝒑2π‘ž ]
p
7
12
0.00
17
Synthetic noise
10.00
15.00
20.00
Mean Noise Sigma
True Noise Sigma
Clean image
5.00
Our method
Metric-Q
Best BM3D
Absolute Noise Level Estimation with Global Optimization
ο‚§ We estimate {πœŽπ‘– } conditioning on
ο‚§
2
πœŽπ‘–
2
𝑖≠𝑗 𝑀𝑖𝑗 πœŽπ‘–
βˆ—
{πœŒπ‘–π‘— }
2
πœŽπ‘—
=argmin
βˆ’
βˆ’
𝑀𝑖𝑗 : similarity between two faces
βˆ— 2
πœŒπ‘–π‘—
User Study
2
πœŽπ‘–
βˆ—
πœŒπ‘–π‘—
ο‚§ Solving a linear system
2
πœŽπ‘—
ο‚§ The system is under-determined, up to adding
a constant number.
- option 1: assign some images to be zero noise
- option 2: assuming the collection contains clean
images, assign the least noisy one to be zero. We use this one for evaluations
ο‚§ Based on BM3D denoised result, decide which
one is preferable
ο‚§ Ran on 71 images, each is evaluated by 3 users
Input image
Our method
Metric-Q
35%
24%
41%
Οƒ=46.3
Οƒ=35
Metric-Q
Οƒ
PSNR
19.25
22.33
20.31
34.65
27
34.53
23
34.65
Selected References
C. Liu, R. Szeliski, S. Kang, C. Zitnick, and W. Freeman. Automatic estimation and removal of
noise from a single image. IEEE Transactions on Pattern Analysis and Machine Intelligence,
30(2), 2008
X. Zhu and P. Milanfar. Automatic parameter selection for denoising algorithms using a noreference measure of image content. IEEE Transactions on Image Processing, 19(12), 2010.
Acknowledgements
Ours
Equally good
Οƒ=9.1
Οƒ=27
We thank MIT Graphics and Vision group for helpful discussion. We would like to thank the
volunteers who participated in the user study.