7998912fdff762214713f89880f29228.ppt
- Количество слайдов: 29
Active Random Fields Adrian Barbu 1
The MAP Estimation Problem n Estimation problem: Given input data y, solve n Example: Image denoising n n Given noisy image y, find denoised image x Noisy image y Issues n n Modeling: How to approximate ? Computing: How to find x fast? Denoised image x 2 FSU
MAP Estimation Issues n Popular approach: n n n Problems with this approach n n n Find a very accurate model Find best optimum x of that model Hard to obtain good Desired solution needs to be at global maximum For many models , the global maximum cannot be obtained in any reasonable time. Using suboptimal algorithms to find the maximum leads to suboptimal solutions E. g. Markov Random Fields 3 FSU
Markov Random Fields n Bayesian Models: n n Markov Random Field (MRF) prior E. g. Image Denoising model n Gaussian Likelihood n Fields of Experts MRF prior n Image Filters Ji Differential Lorentzian Roth and Black, 2005 4 FSU
MAP Estimation (Inference) in MRF n Exact inference is too hard n For the Potts model, one of the simplest MRFs it is already NP hard (Boykov et al, 2001) n Approximate inference is suboptimal n n n Gradient descent Iterated Conditional Modes (Besag 1986) Belief Propagation (Yedidia et al, 2001) Graph Cuts (Boykov et al, 2001) Tree-Reweighted Message Passing (Wainwright et al, 2003) 5 FSU
Gradient Descent for Fields of Experts n Energy function: n Analytic gradient (Roth & Black, 2005) FOE filters n Gradient descent iterations n 3000 iterations with small n Takes more than 30 min per image on a modern computer 6 FSU
Training the MRF n Gradient update in model parameters n n Gradient ascent in log-likelihood n Need to know Normalization Constant Z n EX from training data n Z and Ep obtained by MCMC n n Minimize KL divergence between learned prior and true probability Slow to train Training the FOE prior n Contrastive divergence (Hinton) n n n An approximate ML technique Initialize at data points and run a fixed number of iterations Takes about two days 7 FSU
Going to Real-Time Performance n Wainwright (2006) n n In computation-limited settings, MAP estimation is not the best choice Some biased models could compensate for the fast inference algorithm How much can we gain from biased models? Proposed denoising approach: n n 1 -4 gradient descent iterations (not 3000) Takes less than a second per image 1000 -3000 times speedup vs MAP estimation Better accuracy than FOE model 8 FSU
Active Random Field n Active Random Field = A pair (M, A) of n n a MRF model M, with parameters M a fast and suboptimal inference algorithm A with parameters A They cannot be separated since they are trained together E. g. Active FOE for image denoising n Fields of Experts model n Algorithm: 1 -4 iterations of gradient descent n Parameters: 9 FSU
Training the Active Random Field n n Discriminative training Training examples = pairs n inputs yi+ desired outputs ti n Training=optimization n Loss function L n n n Aka benchmark measure Evaluates accuracy on training set End-to-end training: n covers entire process from input image to final result 10 FSU
Related Work n Energy Based Models (Le. Cun & Huang, 2005) n n n Train a MRF energy model to have minima close to desired locations Assumes exact inference (slow) Shape Regression Machine (Zhou & Comaniciu, 2007) n n n Train a regressor to find an object Uses a classifier to clean up result Aimed for object detection, not MRFs 11 FSU
Related Work n Training model-algorithm combinations n n CRF based on pairwise potential trained for object classification Torralba et al, 2004 Auto. Context: Sequence of CRF-like boosted classifiers for object segmentation, Tu 2008 Both minimize a loss function and report results on another loss function (suboptimal) Both train iterative classifiers that are more and more complex at each iteration – speed degrades quickly for improving accuracy 12 FSU
Related Work n Training model-algorithm combinations and reporting results on the same loss function for image denoising n n Tappen, & Orlando, 2007 - Use same type of training for obtaining a stronger MAP optimum in image denoising Gaussian Conditional Random Fields: Tappen et al, 2007 – exact MAP but hundreds of times slower. Results comparable with 2 -iteration ARF Common theme: trying to obtain a strong MAP optimum This work: fast and suboptimal estimator balanced by a complex model and appropriate training 13 FSU
Training Active Fields of Experts n Training set n n 40 images from the Berkeley dataset (Martin 2001) Same as Roth and Black 2005 n Separate training for each noise level n Loss function L = PSNR n Same measure used for reporting results Trained Active FOE filters, niter=1 is the standard deviation of 14 FSU
Training 1 -Iteration ARF, =25 Follow Marginal Space Learning n Consider a sequence of subspaces Represent marginals by propagating particles between subspaces Propagate only one particle (mode) 1. Start with one filter, size 3 x 3 n n n 2. Add another filter initialized with zeros n 3. 4. Train until no improvement We found the particle in this subspace PSNR training (blue), testing (red) while training the 1 -iteration ARF, =25 filte Retrain to find the new mode Repeat step 2 until there are 5 filters Increase filters to 5 x 5 15 FSU
Training other ARFs n Other levels initialized as n n n Start with one iteration, =25 Each arrow takes about one day on a 8 core machine 3 -iteration ARFs can also perform 4 iterations 16 FSU
Concerns about Active FOE n Avoid overfitting n Use large patches to avoid boundary effect n n Totally 6 million nodes n n n Full size images instead of smaller patches Lots of training data Use a validation set to detect overfitting Long training time n n n Easily parallelizable 1 -3 days on a 8 core PC Good news: CPU power increases exponentially (Moore’s law) 17 FSU
Results Original Image 4 -iteration ARF, Corrupted with Gaussian noise, =25, PSNR=20. 17 PSNR=28. 94, t=0. 6 s 3000 -iteration FOE, PSNR=28. 67, t=2250 s 18 FSU
Standard Test Images Lena House Barbara Peppers Boats 19 FSU
Evaluation, Standard Test Images Lena Barbara FOE (Roth & Black, 2005) 30. 82 27. 04 Active FOE, 1 iteration 30. 15 Active FOE, 2 iterations Boat s Hous e Pepper s Averag e 28. 72 31. 11 29. 20 29. 38 27. 10 28. 66 30. 14 28. 90 28. 99 30. 66 27. 49 28. 99 30. 80 29. 31 29. 45 Active FOE, 3 iterations 30. 76 27. 57 29. 08 31. 04 29. 45 29. 58 Active FOE, 4 iterations 30. 86 27. 59 29. 14 31. 18 29. 51 29. 66 Wavelet Denoising (Portilla et al, 2003) 31. 69 29. 13 29. 37 31. 40 29. 21 30. 16 Overcomplete DCT (Elad et al, 2006) 30. 89 28. 65 28. 78 31. 03 29. 01 29. 67 Globally Trained Dictionary (Elad et al, 2006) 31. 20 27. 57 29. 17 31. 82 29. 84 29. 92 KSVD (Elad et al, 2006) 31. 32 29. 60 29. 28 32. 15 29. 73 30. 42 BM 3 D (Dabov et al, 2007) noise=25 32. 08 30. 72 29. 91 32. 86 30. 16 31. 15 20 FSU
Evaluation, Berkeley Dataset 68 images from the Berkeley dataset n Not used for training, not overfitted by other methods. n Roth & Black ‘ 05 also evaluated on them. n A more realistic evaluation than on 5 images. 21 FSU
Evaluation, Berkeley Dataset Average PSNR on 68 images from the Berkeley dataset, not used for training. 1: Wiener Filter 2: Nonlinear diffusion 3: Non-local means (Buades et al, 2005) 4: FOE model, 3000 iterations, 5, 6, 7, 8: Our algorithm with 1, 2, 3 and 4 iterations 9: Wavelet based denoising (Portilla et al, 2003) 10: Overcomplete DCT (Elad et al, 2006) 11: KSVD (Elad et al, 2006) 22 12: BM 3 D (Dabov et al, 2007) FSU
Speed-Performance Comparison noise=2 23 FSU
Performance on Different Levels of Noise n n n Trained for a specific noise level No data term Band-pass behavior 24 FSU
Adding a Data Term n Active FOE n 1 -iteration version has no data term n Modification with data term n Equivalent 25 FSU
Performance with Data Term n n Data term removes band-pass behavior 1 -iteration ARF as good as 3000 -iteration FOE for a range of noises 26 FSU
Conclusion n An Active Random Field is a pair of n n n Training = optimization of the MRF and algorithm parameters using n n n A benchmark measure on which the results will be reported Training data as pairs of input and desired output Pros n n n A Markov Random Field based model A fast, approximate inference algorithm (estimator) Great speed and accuracy Good control of overfitting using a validation set Cons n Slow to train 27 FSU
Future Work n Extending image denoising n n Learning filters over multiple channels Learning the robust function Learn filters for image sequences using temporal coherence Other applications n Computer Vision: n n n Edge and Road detection, Image segmentation Stereo matching, motion, tracking etc Medical Imaging n Learning a Discriminative Anatomical Network of Organ and Landmark Detectors 28 FSU
References n n n n n A. Barbu. Training an Active Random Field for Real-Time Image Denoising. IEEE Trans. Image Processing, 18, November 2009. Y. Boykov, O. Veksler, and R. Zabih. Fast approximate energy minimization via graph cuts. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 23(11): 1222– 1239, 2001. A. Buades, B. Coll, and J. M. Morel. A Non-Local Algorithm for Image Denoising. Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE Computer Society Conference on, 2, 2005. K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian. Image Denoising by Sparse 3 -D Transform-Domain Collaborative Filtering. Image Processing, IEEE Transactions on, 16(8): 2080– 2095, 2007. M. Elad and M. Aharon. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process, 15(12): 3736– 3745, 2006. G. E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence. Neural Computation, 14(8): 1771– 1800, 2002. Y. Le. Cun and F. J. Huang. Loss functions for discriminative training of energy-based models. Proc. of the 10 -th. International Workshop on Artificial Intelligence and Statistics (AIStats 05), 3, 2005. D. Martin, C. Fowlkes, D. Tal, and J. Malik. A Database of Human Segmented Natural Images and its Application to Evaluating Segmentation Algorithms. Proc. of ICCV 01, 2: 416– 425. J. Portilla, V. Strela, MJ Wainwright, and EP Simoncelli. Image denoising using scale mixtures of Gaussians in the wavelet domain. Image Processing, IEEE Transactions on, 12(11): 1338– 1351, 2003. 29 S. Roth and M. J. Black. Fields of Experts. International Journal of Computer Vision, FSU
7998912fdff762214713f89880f29228.ppt