In a conventional single-exposure photograph, moving objects or moving cameras cause motion blur. The exposure time defines a temporal box[via Behdad]
filter that smears the moving object across the image by convolution. This box filter destroys important high-frequency spatial details
so that deblurring via deconvolution becomes an ill-posed problem.
Rather than leaving the shutter open for the entire exposure duration, we ``flutter'' the camera's shutter open and closed during the
chosen exposure time with a binary pseudo-random sequence. The flutter changes the box filter to a broad-band filter that preserves
high-frequency spatial details in the blurred image and the corresponding deconvolution becomes a well-posed problem. We demonstrate that
manually-specified point spread functions are sufficient for several challenging cases of motion-blur removal including extremely large
motions, textured backgrounds and partial occluders.
coded exposure photography
I was tickled to discover another IBM developerworks article on one of my abandoned hacks and that both it and its predecessor have been translated…
I've been pretty busy with work lately, so I may as well dump this on the internet before it gets too dusty. Though I think I understand Git decently…