 Open Access
 Total Downloads : 384
 Authors : Miss. Supriya S. Meshram, Prof (Mrs.) P. U. Chati
 Paper ID : IJERTV2IS60824
 Volume & Issue : Volume 02, Issue 06 (June 2013)
 Published (First Online): 25062013
 ISSN (Online) : 22780181
 Publisher Name : IJERT
 License: This work is licensed under a Creative Commons Attribution 4.0 International License
Iterative RichardsonLucy Algorithm For The Modeling And Correction Of Blurred Image With A Known Point Spread Function
Miss. Supriya S. Meshram Priyadarshini college of Engineering, Nagpur, Maharashtra,India
Prof (Mrs.) P. U. Chati
Priyadarshini College of Engineering Nagpur, Maharashtra, India
Abstract
This paper addresses how to model and correct the image blur that arises due to defocusing, handshaking, object motion (motion blur) while capturing an image. Nonblind deconvolution is a key component in an image deblurring system . The deconvolution tries to invert the blurring of an image that is modeled by the convolution g = f h. It is the process that obtains a sharp latent image from a blurred image when a point spread function (PSF) is known. However, ringing and noise amplification are inevitable artifacts in image deconvolution since perfect PSF estimation is impossible. The convolutional regularization to reduce these artifacts cannot preserve image details in the deconvolved image when PSF estimation error is large. Hence strong regularization is needed. A simple iterative technique has been developed for nonblind deconvolution using Richardson Lucy Algorithm of two convolved functions. The method is described as follows along with some experimental results. The goal of this paper is to preserve the details of the image, while suppressing ringing and noise artifacts by controlling regularization strength according to the local characteristics of the image.
Key words: PSF, Motion blur, Spatially varying motion blur, Ego motion.

INTRODUCTION
Restoration of digital images from their degraded measurement has always been a problem of great interest. A specific solution to the problem of image restoration is generally determined by the nature of degradation phenomena. So it is highly dependent on the nature of the noise present there. Given the noise function, one can use the RichardsonLucy Algorithm to restore the degraded image. This algorithm was introduced by W.H. Richardson (1972) and L.B. Lucy (1974).
Single image deblurring has gained considerable attention in recent years. Given a blurry input image b, existing deblurring approaches model the image degradation as:
b = l + n; (1)
where k is the blur kernel l refers to the underlying sharp latent image, n is noise, and is the convolution operator. The task for blind image deconvolution is to infer both k and l from a single input b, which is severely illposed. Blind deconvolution approaches intensively use nonblind deconvolution, the process of estimating the latent image l given the blurred image b and the estimated blur kernel k, while inferring k and for generating the final output l. This makes nonblind deconvolution a key component in the deblurring pipeline. Various nonblind deconvolution approaches have been proposed in the literature, ranging from classic Wiener filter to modern optimization approaches with image priors. However, in practice these methods often produce severe ringing artifacts even when the blur kernel is known or well estimated

What is an Image?
An image is nothing but a huge collection of numbers known as pixels. In particular a gray image is an image in which the value of each pixel is a single sample i.e. it carries only intensity information. So a pixel in a given image is just the intensity at that particular point. The pixel value is a number between 0 and 1 (both inclusive). 0 denotes the total absence (i.e. black) and 1 denotes the total presence (i.e. white).

Point Spread Function
The Point Spread Function describes the response of an imaging system to a point source or point object. Following is an example of a PSF:
Figure 1: Image formation in a Confocal Microscope: Central longitudinal (XZ) slice, the 3D acquired distribution arises from the convolution of the real light sources with the PSF.

Poisson Distribution
The Poisson Distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time and/or space if these events occur with a known average rate and independently of the time since the last event. Poisson Distribution can also be used for the number of events in other specified intervals such as distance, area and volume.
Fig 2. The horizontal axis is the index k, the number of occurrences. The function is only defined at integer values of k. The connecting lines are only guides for the eye.
Aim of our project can be very well explained from the flow diagram of figure 3, Which follows some steps are:

Step 1: Load a clear image

Step 2: Blur the loaded clear image

Step 3: Obtain clear image by using Richardson Lucy Algorithm

Step 4: Damping (i.e. Compare initially loaded clear image and step 3 clear image)
Fig 3. Flow of Project



RELATED WORK
Existing work targeting image blur due to camera ego motion has assumed a global PSF for the entire image. When the blur PSF is known, or can be estimated, well known debluring algorithms, such as RichardsonLucy and Wiener filter , can be applied to deblur the image. Due to poor kernel estimation or convolution with PSFs that result in unrecoverable frequencies, these conventional deblurring algorithms can introduce undesirable artifacts in the deblurred result such as ringing and amplification of image noise.
Consequently, research addressing image deblurring, including camera shake and other types of blur, typically targets either blur kernel estimation or ways to regularize the final result, or both. For example, Dey et al. and Chan and Wong utilized total variation regularization to help ameliorate ringing and noise artifacts. Fergus et al. demonstrated how to use a variational Bayesian approach combined with
gradientdomain statistics to estimate a more accurate PSF. Raskar et al. coded the exposure to make the PSF more suitable for deconvolution. Jia demonstrated how to use an objects alpha matte to better compute the PSF. Levin et al. introduced a gradient sparsity prior to regularize results for images exhibiting defocus blur. This prior is also applicable to motion blurred images. Yuan et al. proposed a multiscale approach to progressively recover blurred details, while Shan et al. introduced regularization based on highorder partial derivatives to reduce image artifacts.
These previous approaches all work under the uniform PSF assumption. Camera ego motion causes a spatially varying motion blur that cannot be accurately modeled with a uniform PSF. Prior work has recognized the need to handle nonuniform motion blur for camera ego motion, moving objects, and defocus blur. For example, early work by Sawchuk addressed motion blur from a moving camera by first using a logpolar transform to transform the image such that the blur could be expressed as a spatially invariant PSF. The range of motion that could be addressed was limited to rotation and translation. When addressing moving objects, the input image can be segmented into multiple regions each with a constant PSF, as demonstrated by Levin , Bardsley et
al., Cho et al. , and Li et al. . Such segmented regions, however, should be small, to make the constant PSF
original data
image = image ————— PSF n+1 n image * PSF
n
In formal way one can express the output of the continuous linear system in the form of probabilities:
y t h t x d
p>Where is the unblurred image
is the PSF (the fraction of light coming from true location that get scattered into observed pixel t)
And y(t) is the blurry image. from this one can write
x q t y t dt
assumption valid for the spatially varying motion blur in camera shake motion. For example, Tai et al. extended the hybrid camera framework used by Ben
where q t
theorem
is related tho t
via Bayes
Ezra and Nayar to estimate a PSF per pixel using an auxiliary video camera. This need for a perpixel PSF
revealed the futility of relying on the conventional
q t
x
h t
y t
kernelbased PSF model for spatially varying blur due to ego motion.
q t is calculated using an iteration approach
h t

RICHARDSON LUCY DECONVOLUTION ALGORITHM (Nonblind)
q(n)
and
t x(n)
y (n) t
h t
The reason for the popularity of the Richardson Lucy algorithm is its implementation of maximum
x(n)
q(n1) t y(0) t dt
x(n1)
y(n1) t
y(0) t dt
likelihood and its apperent ability to produce reconstructed image of good quality in the presence
with y(0) t being the measured data
y t and
of high noise levels.The RichardsonLucy algorithm generates a restored image through an iterative
y(n1) t
h t x(n1) d
method. The essence of the iteration is as follows: the (n+1)th estimate of the restored image is given by the
finally in discrete form we have
nth estimate of the restored image multiplied by a correction image. That is,
x(n ) i x(n1) i h j,i
M 1
y j
,
N 1
N 1
j 0
h j,k x(n1) k
k 0
i 0,M 1
The result of each iteration is used to make a prior for the next one,the RichardsonLucy iteration converges to the maximum likelihood solution for Poisson statistics in the data.

Advantages Of RL algorithm

Iterative RichardsonLucy algorithm gives an estimation x(k) of the true image x

The restored images are robust against small errors in the used point spread function.

Easy to implement (no nonlinear optimisation is needed)

The RL iteration converges to the maximum likelihood solution for Poisson statistics in the data (Shepp and Vardi 1982), which is appropriate for optical data with noise from counting statistics.

The RL method forces the restored image to be nonnegative and conserves ux both globally and locally at each iteration.



FLOW OF WORK
Fig4(a). Flow Of work
Fig4(b). Flow Of Work
Fig4(c). Flow Of Work

IMPLEMENTATION WORK
Step 1: Load a clear image
From fig 3 our first step is loading of clear image. The basic objective of step 1 is: An image must come on our Form in any case from Harddisk and while coming it must come with the whole properties of image such as size of image , name of image, pixel size, size of histogram, location of image etc
Methodology to achieve step 1 follows the following steps:

Creation Of Form

File open, file save, RGB, lightness, blur, diff, add, deblur button Event Creation.

Open Dialog Box

Dialog box showing filters with selected images

Loading the Clone of selected image on our form
Fig5. Creation of form Image Magic and file open, file save, RGB, lightness, blur, diff, add, deblur button event creation.
Fig6. Click event on: File Open button
Fig7. Selecting the image from open dialogue box, which is showing filters such as jpg., fpg., bmp., etc (file of type) along with the selected image and its properties.
Fig8. Loaded clone of selected image on our form.
Step2: Blur the loaded clear image.
From fig3 our next step is to obtain the blurred image from the loaded clone of image, which we have obtained from step1. Step 2 follows the following steps for obtaining the blur image:

Save the clone of image.

Click event on File save button

Click event on Lightness button
By clicking on lightness button we can obtain grays scale image. Here we require the gray scale image because in image processing whenever there is a need of prominent edge detection the RGB image has to be converted into gray scale image so that the edge can be detected with accuracy and the it is converted into editable mode using BMS engine.

Click event on Blur button
For obtaining the blurry image or the synthetic image there are so many techniques some of them are: linear blurring technique, in which horizontal and vertical fixed number of pixel. Block blurring technique where averaging a small block of pixels by propagating a fixed sized window through the entire image. Gaussian blurring technique where the convolution of image with two dimensional Gaussian functions has taken place. But in our project we are using linear blurring technique because of its advantages as follows:

A linear blur helps us to add motion to our images. It blurs all of our image or a portion of our image we select.

A linear blur is a blur that goes in a straight line. It can originate in any direction, running vertically, horizontally or at an angle.

Linear blur shouldn't be confused with linear sampling. Linear sampling is a type of Gaussian blur, not a true linear blur. A Gaussian blur creates soft edges in an image, making it appear blurry.


Obtaining the blur port.

Obtaining the blurred image or synthetic image on our form.
Fig9.Click event on File Save button
Fig10. Save the selected file with another name (here e.g. Testing 2013)
Fig11 . Click event on lightness button
Fig12. Obtained gray scale image from loaded clear image.
Fig13. Click event on blur button
Fig
14. Obtaining the bur port where we can adjust the blurring ratio (here e.g. blur ratio=10) also we can see the image by zooming it.
Fig15. Obtained blurred image or synthetic image.
Step 3: Obtaining the clear image by using Richardson Lucy Deblurring algorithm.
From fig3. Our next step is to obtained a deblurred image from the blurry image, which governs the following steps:

Click event on diff button

Obtaining the difference port

Obtaining the resultant image of differnce port on our form.

Click event on add button

Obtaining the add port

Obtaing the resultant image of add port on our form.

Click event on deblurr button (here we are using Richardson Lucy Deblurring algorithm (non blind) as we have mentioned in this paper)

Obtaining the resultant deblurred/ deconvolved image.
Fig16. Click event on difference button
Fig17(a). Obtaining the diff port (here Display Brightness = 10 Threshold = 15 Edge Enhancement= Lighten and Darken)
Fig17(b).Obtaining the diff port(here Display Brightness = 10 Threshold = 15 Edge Enhancement= Lighten only )
Fig17(c). Obtaining the blur port(here Display Brightness = 10 Threshold = 15 Edge Enhancement= Darken only)
Fig18. Obtaining the resultant image of diff port (here the resultant image is having edge enhancement
of lighten and darken both).
Fig19. Click event on Add button
Fig20. Obtaining the add port where we can adjust the value of alpha (e.g. non blind deconvolution where the value of PSF is known here PSF is nothing
but alpha) (here value of alpha =20)
Fig21. Obtaining the resultant image of add port.
Fig22. Click event on deblurr buttonFig23. Obtaining the resultant deblurred/ deconvolved image.
Step 4: Damping
So from fig3 our last step is damping that is comparison initially loaded image on our form and the finally obtained resultant image. In fig24 we can see the damping.
Fig24. Damping
process
Fig25. Experiment with different alpha values with edge enhancement as lighten and darken only.
Fig26. Experiment with different alpha values with edge enhancement as lighten only.
Fig27. Experiment with different alpha values with edge enhancement as darken only.

CONCLUSION
We introduced a powerful framework for space variant non blinddeconvolution technique. Our current implementation takes less than 1minute to obtain the clear image from the input blurry image. The running time of our algorithm depends on several factors, including image size , number of discrete sampling N, and the number of iterations. We have used dotnet framework 4.0 (version 2010) to obtain these results.

FUTURE DIRECTIONS
There are several future directions of this work. One is to explore other existing algorithms and hardware (e.g., coded exposure and coded aperture) that can
use projective motion path blur formulation. Another important direction is to consider how our framework can be used to perform blinddeconvolution, where the cameras motion path is unknown. One can also increase the size of loaded image, since in our experiment we can take input image which is not more than 85 kb.

REFERENCES

T.F. Chan and C.K. Wong, Total Variation Blind Deconvolutiontion, IEEE Trans. Image Processing, vol. 7, no. 3, pp. 370375, Mar1998

YuWing Tai, Member, IEEE, Ping Tan, Member, IEEE, and Michael S. Brown, Member, IEEE, RichardsonLucy Deblurring for Scenes under a Projective Motion Path IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 33, NO. 8, AUGUST 2011.

G. R. Ayers and J. C. Dainty, ÂªIterative blind deconvolution method and its applications,Âº Opt. Lett. 13, 547Â±549 (1988).

J. R. Feinup, ÂªPhase retrieval algorithms: a comparison,Âº Appl. Opt. 21, 2758Â±2769 (1982).

L.A. Shepp and Y. Vardi, Maximum Likelihood Reconstruction for Emission Tomography, IEEE Trans. Medical Imaging, vol. 1, no. 2, pp. 113122, Oct. 1982.

Y. Vardi, Nonlinear Programming. Prentice Hall, 1969.

A.A. Sawchuk, SpaceVariant Image Restoration by Coordinate Transformations, J. Optical Soc. Am., vol. 64, no. 2, pp. 138144, 1974.

W. H. Richardson, ÂªBayesianbased iterative method of image restoration,Âº J. Opt. Soc. Am. 62, 55Â±59 (1972).

L. B. Lucy, ÂªAn iterative technique for the rectification of observed distributions,Âº Astron. J. 79, 745Â±754 (1974).

10) Lucy L.B., A.J. 79 (1974) 745.

Richardson W.H., J. Opt. Soc. Am. 62 (1972) 55.

T. J. Holmes, ÂªBlind deconvolution of quantumlimited incoherent imagery: maximum likelihood approach,Âº J. Opt. Soc. Am. A 9, 1052Â±1061 (1992).
Books

Dot Net Programming Anetshel By Aslibilgin and A. Sybex

Mastering Dot Net By Evangelos

Digital Image Processing By Rafael C. Gonzalez And Richard E. Wood