|
![](images/seal_bl_1inch_72dpi.gif) |
UNIVERSITY
OF CALIFORNIA,
SANTA CRUZ |
|
|
|
|
|
|
|
Space-Time
Steering Kernel Regression for Video Upscaling |
|
|
Hiroyuki
Takeda, Prof. Peyman Milanfar$
|
|
|
Abstract
The
need for precise (subpixel accuracy) motion estimates in conventional
super-resolution has limited its applicability to only video sequences
with relatively simple motions such as global translational or affine
displacements. In this work, we introduce a novel framework for adaptive
enhancement and spatiotemporal upscaling of videos containing complex
transitions without explicit need for subpixel motion estimation.
Using
some examples shown below, we illustrate that our algorithm has resolution
enhancement capabilities that provide improved optical resolution in
the output, while being able to work on general input video with essentially
arbitrary motions.
Related
publications:
H.
Takeda and P. Milanfar,"Locally
Adaptive Kernel Regression for Space-Time Super-Resolution",
To appear as a Chapter in Super-resolution
Imaging Edited by P. Milanfar, Available September 27,
2010.
Takeda, H., P. Milanfar, M. Protter, and M. Elad, "Superresolution
without Explicit Subpixel Motion Estimation", IEEE
Transactions on Image Processing, Vol. 18, No. 9, September 2009.
Takeda,
H., S. Farsiu, and P. Milanfar,
"Kernel
Regression for Image Processing and Reconstruction",
IEEE Transactions on Image Processing, Vol. 16, No. 2, pp. 349-366,
February 2007.
|
|
|
|
|
|
Space-Time
Video Upscaling
3-D
Steering Kernel Regression
Our approach
is based on multi-dimensional (3-D) kernel regression, where each pixel
in the video is approximated with a 3-D local polynomial (Taylor) series,
captureing the essential local behavior of its spatiotemporal neighborhood.
The coefficients of this series are estimated by solving a local weighted
least-squares problem, where the weights are a function of the 3-D space-time
orientation in the local analysis cubicle. As this method is fundamentally
based upon the comparison of neighboring pixels in both space and time,
it implicitly contains information about not only the spatial orientation
structures but also the local motion trajectories across time (see Figure
1 below). Therefore rendering unnecessary an explicit computation of motions
of modes size
![](images/MotionComp03.png)
(a) A local region |
![](images/SK3D_weights_tube.png)
(b) Steering kernel weights |
Figure 1: Suppose
we estimate pixel at the center of the middle frame in a local region of
the input video where a small circle moves upward frame-by-frame as illustrated
in (a), our 3-D steering kernel weights spread along the space-time orientation
structures as in (b).
Implementation
When the speed of the local motions is relatively large,
a mechanism for neutralizing such displacements roughly (to within
a whole pixel accuracy) is introduced, which may be called coarse
motion compensation. Within this framework, the remaning fine scale
(subpixel) motions are then captured using the 3-D kernel regression.
Figure 2 shows the block diagram of the 3-D SKR method with motion compensation
and an iteartive scheme. The proposed approach not only significantly
widens the applicability of super-resolution methods to a variety of video
sequences contaning complex motions, but also yields improved overall
perfomance.
![](images/SKR3D_diagram_initialization.png)
(a) Initialization |
![](images/SKR3D_diagram_iteration.png)
(b) Iteration |
Figure 2:
The block diagram representation of the iterative 3-D SKR with motion compensation:
(a) In the initialization step, first, we estimate block motions roughly
and compensate the large motions block-by-block, then analyze the space-time
local orientation structures from the gradients by 3-D classic kernel regression
in order to obtain steering matrices. Once the steering matrices are available,
we upscale the motion-compensated video in space and time by the 3-D SKR.
(b) Using the gradients estimated by the 3-D SKR, we refine the steering
matrices for the next iteration. When the input video carries severe noise,
the iteration can reduce the noise further. Finally, we perform the deblurring
in order to recover more detailes of the objects in each frame.
|
|
|
|
|
|
Examples
1.
Spatial upscaling example
![](SKR3Dvideos/ForemanUpscalingExample_Input_Lanczos_NLSR_3DISKR_F06_23.png)
![](SKR3D/Foreman_SpaceTimeUpscale_SKR3D.gif)
(a)
Input frames |
(b)
Lanczos |
|
(d)
3-D SKR |
A video upscaling
example using Foreman sequence: (a) the degraded frames at time
t = 6 and 23, (b) the upscaled frames by Lanczos interpolation,
(c) the upscaled frames by NL-means based SR, and (d) the upscaled
frames by 3-D SKR. |
![](images/ForemanUpscale_F06_Lanczos_NLM_SKR_zoom.png)
Selected regions
of the upscaled frames at time t = 6 by Lancoz, NL-means based
SR, and (d) the upscaled frames by 3-D SKR.
The output video is also available here.
The video shows the input frames in the left, the upscaled frames
by Lanczos, NLM-based SR, and 3-D SKR in the second left, the
second right, and the right, respectively.
|
2. Spatiotemporal
upscaling example
(e)
The input low resolution video |
(f)
3-D bilinear interpolation |
(g)
3-D SKR with motion compensation |
|
A
video upscaling example of Stefan sequence
(a) original
video frames at time t = 4 to 5 (cropped 8 frames in CIF format).
(b)(e) the input videos generated by blurred with a 2x2 uniform
PSF, adding white Gaussian noise with standard deviation 2,
and spatially downsampling with the factor of 2 : 1.
(c)(f) upscaled by 3-D bilinear interoplation and deblurred
by BTV.
(d)(g) upscaled by 3-D SKR with motion compensation and deblurred
by BTV.
The results
are also available in video (AVI) format here.
The
video shows the original frames ( left), the upscaled frames by
3-D linear interpolation (middle), and the upscaled frames by
3-D SKR with motion compensation (right). For this video, we used
the spatial upscaling factor 1 : 2 and temporal upscaling factor
1 : 10, and deblurred all the frames one-by-one with the 2x2 uniform
PSF and BTV
regularization. |
|
|
|
|
|
|
Software
The software
package of the space-time steering kernel regression for MATLAB is downloadable
from here. The user's manual
is also available as either PPT
or PDF.
This is
a package of the experimental codes. It is provided for noncommercial
research purposes only. Use at your own risk. No warranty is implied by
this distribution. Copyright © 2010 by University of California,
Santa Cruz. |
|
|
|
|
|
Acknowledgment
This
work was supported in part by AFOSR Grant FA9550-07-1-0365.
top
|
|
|
|
|
|
last update on Apr 4th, 2011 |
|