Thesis Presentation Single Image Denoising
Thesis Presentation Single Image Denoising
Denoising.
Jubyrea
MCE 077 05513
Supervisor: Md. Tauhid Bin Iqbal
August 30, 2023
Department of Computer Science and Engineering
Stamford University Bangladesh
1
Table of contents
1. Introduction
2. Background Studies
Traditional Denoiser
Learning-based Denoising Methods
3. Limitations of Existing Works
4. Proposed Solutions
Cut2Self
Experimental Result
Cut2Self-DDM
Experimental Results
2
Introduction
Image Denoising?
y x n
Figure 1: Visual representation of a noisy image (y), a clean counterpart (x), and noise (n). Noise
(n) can be signal independent, signal dependent, or both.
y=x+n (1)
4
Background Studies
Denoising Methods
Denoising Methods
Un/Semi-
Supervised Self-supervised
supervised
5
Traditional Denoiser
6
Learning-based Denoising Methods
Supervised Learning
DnCNN [5], U-Net [6], FFDNet [7], RIDNet [8], are some supervised learning models
with impressive performance.
7
Learning-based Denoising Methods (contd.)
Self-supervised Learning
• denoiser train from the body of input image
• no external data other than input image
• applicable in areas only noisy image available
8
Limitations of Existing Works
Challenges and Strategies of Self-supervised Learning
• trains the denoising network by creating training pairs from input noisy image
• pairs are generated by randomly dropping pixels from the input image using Bernoulli
distribution
• try to address: identity mapping, overfitting, variance
• assumption → noise is pixel independent
• real world noise is pixel dependent [11] [26] [25]
9
Self2Self-Limitations
11
Cut2Self Discussion
Figure 3: The proposed framework for denoising. Our input to the network is a noisy image (a),
(b) is the noisy input to the network after masking out square regions as depicted in the figure, (c)
represents the denoising network depicting the encoder and decoder blocks. Final output is the
denoised image (d); for ease, area under red box is shown after zooming in for a clear view.
12
Cut2Self (Training Schema)
◦ We create training pairs from the noisy input by masking out square regions.
First sample of the training pair: ŷ = c ⊙ y
Second sample of the training pair: ȳ = (1 − c) ⊙ y
y c ŷ ȳ
Figure 4: Noisy input (y), proposed cutout mask (c), first and sample of training pair (ŷ & ȳ).
N
∑
min ∥Fθ (ŷn ) − ȳn ∥ℓcn , (2)
n=1
13
Cut2Self (Training Schema) (contd.)
We rewrite the loss function Eq. 2 for ℓ2 loss,
N
∑ 2
∥Fθ (ŷn ) − ȳn ∥c =
n
∑N
2
∥Fθ (ŷn ) − (xn + nn )∥c
n
Proposition
n=1 n=1 Training a denoiser with loss function (Eq. 2) is
N N
equivalent to training with ground truth loss and
∑ ∑
=
2
∥Fθ (ŷn ) − xn ∥c +
2
∥nn ∥c noise variance. The expectation of the loss
n=1
n
n=1
n function with respect to noise is similar to the
expectation of training with ground truth loss
N
−2
∑
(Fθ (ŷn ) − xn )((1 − cn ) ⊙ n)
T (3) and noise variance. Proof is shown in Eq. 4.
n=1 N N
∑ ℓ ∑ ℓ
N
∑ ∑N ∥Fθ (ŷ) − x∥c + ∥σ∥c (5)
2 2 n n
= ∥Fθ (ŷ) − x∥c + ∥n∥c n=1 n=1
n n
n=1 n=1
N Here σ represents the standard deviation of
T ∑
−2n ( (1 − cn ) ⊙ (Fθ (ŷn ) − xn )
n=1 noise. The proof above is shown using ℓ = 2. The
proof can however be generalized for other
Now if we take En on Eq. 3:
values of ℓ as well.
N
∑ 2
En [ ∥Fθ (ŷn ) − ȳn ∥c ]
n
n=1
N
∑ N
∑
2 2
= En [ ∥Fθ (ŷ) − x∥c ] + En [ ∥n∥c ]
n n
n=1 n=1
(4)
N
T ∑
−2En [n ( (1 − cn ) ⊙ (Fθ (ŷn ) − xn )]
n=1
N
∑ ∑N
2 2
≡ ∥Fθ (ŷ) − x∥c + ∥σ∥c
n n
n=1 n=1
14
Cut2Self (Denoising Schema)
Important takeaways:
15
Network Architecture
16
Parameter Setting: Cut2Self
◦ Cut2Self is tested for best hyper ◦ Best parameters setting:
parameter setting on Set-9 dataset. • Square shape of mask worked best
◦ Some of the key parameters are: • Dropping 0.3% area yields best denoising
• Shape of mask results
• Dropping percentage • Size of mask is set to 4 × 4
• Size of mask • After 1000 training iterations we averaged
• Number of training iteration add 100 images to generate a prediction
prediction • We have experimented with ℓ1 loss
• Choice of loss function function
Circle shaped mask Triangle shaped mask Ellipse shaped mask Square shaped mask
17
Parameter Setting: Cut2Self (contd.)
Area dropping: 0.1% Area dropping: 0.3% Area dropping: 1.0% Area dropping: 1.5% Area dropping: 3.0%
Figure 7: Visual representation of masked area for different dropping percentages, e.g., 0.1%,
0.3%,1.0%, 1.5%, and 3.0% respectively.
30.8
30.8 30.71 30.71
30.71 30.71
30.7 30.63 30.7
30.6
PSNR(db)
30.6 30.5
PSNR(db)
PSNR(db)
30.62
PSNR(db)
PSNR(db)
(a) Shape of Mask (b) Dropping Percentage (c)Size of Mask (d) Iteration-Prediction (e) Loss function
18
Cut2Self (Quantitative Results - Additive noise)
Methods σ = 25 σ = 50
PSNR SSIM PSNR SSIM
†BM3D 27.96 7.74 28.71 0.670
‡DIP 28.55 0.782 25.04 6.45
‡N2V 25.34 6.81 23.85 6.18
‡K-SVD 28.42 7.96 25.08 6.53
†N2N 29.15 0.831 26.23 0.725
‡N2S 27.19 7.69 24.53 6.42
†IDR 29.20 0.835 26.25 0.726
‡Self2Self 28.70 8.03 25.92 6.99
Cut2Self 31.70 0.835 28.95 0.525
19
Cut2Self (Quantitative Results - Additive noise) (contd.)
20
Self2Self-Cutout (Quantitative Results - Real-World noise)
Table 3: Quantitative comparison of real world noise for different methods on PolyU dataset.
Comparisons are done using (PSNR(db)/SSIM) matrices, and the highest PSNR is marked in bold.
Results except ours are taken from [21].
Table 4: Quantitative comparison of real world noise for different methods for SSID+ validation
dataset. Comparisons are done using (PSNR(db)/SSIM) matrices, and the highest PSNR is marked
in bold. Results except ours are taken from [25].
Metric BM3D WNNM NC MCWNNM C2N R2R N2V N2S CVF-SID Cut2Self
PSNR 25.65 26.20 31.31 33.40 34.08 35.04 29.35 30.72 34.81 31.95
SSIM 0.475 0.693 0.725 0.815 - 0.844 0.651 0.787 0.944 0.642
21
Qualitative Results
GT Noisy Cut2Self
22
Qualitative Results (contd.)
GT Noisy Cut2Self
23
Qualitative Results (contd.)
GT Noisy BM3D
Figure 11: Visual comparisons of Set-14 with different models for varied σ ∈ [5, 50]. 24
Improvement Schopes of Cut2Self
Although Cut2Self achieves prominent result, there are scope for improvement.
GT Noisy Cut2Self
◦ The dropout rate is fixed which should be varied for better generalization.
◦ Trivial reconstruction errors (i.e., noise in output, blurriness) can be further improved.
◦ Method can be improved for faster convergence. 25
Cut2Self-DDM Discussion
Figure 13: Sample demonstration of masked input with a varied dropout rate during training
iterations. Moderately, the average dropout rate is half of the input image.
1
1 [29], [30], [31], [32], [33].
26
Cut2Self-DDM: Mask Generation
27
Cut2Self-DDM (Training Schema)
Mask
Set-9 Set-14
Orientation
PSNR SSIM PSNR SSIM
Vertical 32.605 0.79 32.15 0.855
Horizontal 32.295 0.754 31.832 0.812
(a) (b)
Table 6: Quantitative comparison between L1
and L2 regularization.
Table 7: Quantitative comparison on BSD68 dataset using (PSNR(db)/SSIM) matrices, and the
highest PSNR is marked in bold.
Methods σ = 25 σ = 50
PSNR SSIM PSNR SSIM
BM3D 27.96 7.74 28.71 0.670
DIP 28.55 0.782 25.04 6.45
N2V 25.34 6.81 23.85 6.18
K-SVD 28.42 7.96 25.08 6.53
DBSN 28.81 0.818 25.95 0.703
N2N 29.15 0.831 26.23 0.725
N2S 27.19 7.69 24.53 6.42
IDR 29.20 0.835 26.25 0.726
Self2Self 28.70 8.03 25.92 6.99
Cut2Self 31.70 0.835 28.95 0.525
Cut2Self-DDM 33.25 0.871 31.96 0.794
30
Cut2Self-DDM (Quantitative Results) (contd.)
31
Cut2Self-DDM (Quantitative Results) (contd.)
Set-9 K-SVD CBM3D CDnCNN N2V N2S DIP Self2Self Cut2Self Cut2Self-DDM
σ = 25 30.00 31.67 31.42 28.12 29.30 30.77 31.74 30.71 32.70
σ = 50 26.50 28.95 28.84 26.01 27.25 28.23 29.25 28.88 32.01
Table 10: Quantitative comparison of real world noise for different methods on PolyU dataset.
Comparisons are done using (PSNR(db)/SSIM) matrices, and the highest PSNR is marked in bold.
2
2ℓ loss performed poorly on PolyU dataset; result is generated using ℓ2 loss.
1
32
Qualitative Results
GT Noisy
GT Noisy
Figure 17: Visual comparisons of Set-14 with different models for varied σ ∈ [5, 50]. 34
Qualitative Results (contd.)
GT Noisy Cut2Self-DDM
35
Comparison Between Methods on
Higher Noise Level
Comparison between methods
GT Noisy
GT Noisy
Figure 20: Qualitative demonstration between Self2Self, Cut2Self, and Cut2Self-DDM for σ = 50
on Set-9 datasets.
37
Conclusion
Conclusion
• Future Works:
• The proposed work can be further used to domain-specific applications such as
medical imaging, satellite and surveillance imaging, and space imaging.
• The methods can be integrated with image superresolution tasks, as in many cases
denoising is a prior work of superresolution. The proposed work can be modified for
inpainting as well.
38
Publication List
39
Thank You
Questions Please
40
Appendix
(a) (b)
Figure 21: Effect of mean and standard deviation (std) on data driven dropout.