Submitted by CurrentlyJoblessFML t3_10g1tni in MachineLearning
Hi all,
I am trying to see if I can use DDPM (Denoising Diffusion Probabilistic Model) to denoise images using a supervised learning approach. However, I've learned that DDPM is only for unconditional image generation. Has anyone had experience using conditional DDPM and could help me out with some conceptual questions?
Here's what I'm trying to understand:
-
Say I have a pair of noisy and clean ground truth images.
-
Should I take my clean image and gradually corrupt it by adding gaussian noise in the forward diffusion (FD) process?
-
Could I get the network to learn the reverse diffusion process by giving it the noisy input, the FD noisy image, and positional embeddings? I was planning on concatenating the noisy input with the FD noisy image.
-
During training, the network learns to predict noise at t-1 given the image at t conditioned on the input noisy source image.
Here is an image showing you what I mean. Any thoughts or suggestions would be greatly appreciated. DDPM for image denoising
LanverYT t1_j501vdn wrote
That's a really interesting question, and I've been wondering about the same thing. I've never been able to figure it out, but I would love to see what others have to say about it. It sounds like you have a solid approach and understanding of the concept, so I'm curious to see how it turns out. Good luck with your experimentation and let us know how it goes