Blur Robust Optical Flow using Motion Channel

12
Blur Robust Optical Flow using Motion Channel Wenbin Li a,e , Yang Chen b , JeeHang Lee c , Gang Ren d , Darren Cosker e a Department of Computer Science, University College London, UK b Hamlyn Centre, Imperial College London, UK c Department of Computer Science, University of Bath, UK d School of Digital Art, Xiamen University of Technology, China e Centre for the Analysis of Motion, Entertainment Research and Applications (CAMERA), University of Bath, UK Abstract It is hard to estimate optical flow given a realworld video sequence with camera shake and other motion blur. In this paper, we first investigate the blur parameterization for video footage using near linear motion elements. We then combine a commercial 3D pose sensor with an RGB camera, in order to film video footage of interest together with the camera motion. We illustrates that this additional camera motion/trajectory channel can be embedded into a hybrid framework by interleaving an iterative blind deconvolution and warping based optical flow scheme. Our method yields improved accuracy within three other state-of-the-art baselines given our proposed ground truth blurry sequences; and several other realworld sequences filmed by our imaging system. Keywords: Optical Flow, Computer Vision, Image Deblurring, Directional Filtering, RGB-Motion Imaging 1. Introduction Optical flow estimation has been widely applied to computer vision applications, e.g. segmentation, image deblurring and stabilization, etc. In many cases, optical flow is often estimated on the videos captured by a shaking camera. Those footages may contain a significant amount of camera blur that bring ad- ditional difficulties into the traditional variational optical flow framework. It is because such blur scenes often lead to a fact that a pixel may match multiple pixels between image pair. It further violates the basic assumption – intensity constancy – of the optical flow framework. In this paper, we investigate the issue of how to precisely esti- mate optical flow from a blurry video footage. We observe that the blur kernel between neighboring frames may be near linear, which can be parameterized using linear elements of the camera motion. In this case, the camera trajectory can be informatic to enhance the image deblurring within a variational optical flow framework. Based on this observation, our major contribution in this paper is to utilise an RGB-Motion Imaging System – an RGB sensor combined with a 3D pose&position tracker – in or- der to propose: (A) an iterative enhancement process for camera shake blur estimation which encompasses the tracked camera motion (Sec. 3) and a Directional High-pass Filter (Sec. 4 and Sec. 7.2); (B) a Blur-Robust Optical Flow Energy formulation (Sec. 6); and (C) a hybrid coarse-to-fine framework (Sec. 7) for computing optical flow in blur scenes by interleaving an itera- tive blind deconvolution process and a warping based minimisa- tion scheme. In the evaluation section, we compare our method to three existing state-of-the-art optical flow approaches on our proposed ground truth sequences (Fig. 1, blur and baseline blur- free equivalents) and also illustrate the practical benefit of our algorithm given realworld cases. 0 3 6 (Pix.) 1 I 2 I Ours Error Map, Ours Portz et al. Error Map, Portz et al. Figure 1: Visual comparison of our method to Portz et al. [1] on our ground truth benchmark Grove2 with synthetic camera shake blur. First Column: the input images; Second Column: the optical flow fields calculated by our method and the baseline; Third Column: the RMS error maps against the ground truth. 2. Related Work Camera shake blur often occurs during fast camera move- ment in low-light conditions due to the requirement of adopt- ing a longer exposure. Recovering both the blur kernel and the latent image from a single blurred image is known as Blind Deconvolution which is an inherently ill-posed problem. Cho and Lee [2] propose a fast deblurring process within a coarse- to-fine framework (Cho&Lee) using a predicted edge map as a prior. To reduce the noise effect in this framework, Zhong et al. [3] introduce a pre-filtering process which reduces the noise along a specific direction and preserves the image information in other directions. Their improved framework provides high quality kernel estimation with a low run-time but shows diffi- culties given combined object and camera shake blur. To obtain higher performance, a handful of combined hard- ware and software-based approaches have also been proposed Preprint submitted to Neurocomputing March 8, 2016 arXiv:1603.02253v1 [cs.CV] 7 Mar 2016

Transcript of Blur Robust Optical Flow using Motion Channel

Page 1: Blur Robust Optical Flow using Motion Channel

Blur Robust Optical Flow using Motion Channel

Wenbin Lia,e, Yang Chenb, JeeHang Leec, Gang Rend, Darren Coskere

aDepartment of Computer Science, University College London, UKbHamlyn Centre, Imperial College London, UK

cDepartment of Computer Science, University of Bath, UKdSchool of Digital Art, Xiamen University of Technology, China

eCentre for the Analysis of Motion, Entertainment Research and Applications (CAMERA), University of Bath, UK

Abstract

It is hard to estimate optical flow given a realworld video sequence with camera shake and other motion blur. In this paper, wefirst investigate the blur parameterization for video footage using near linear motion elements. We then combine a commercial3D pose sensor with an RGB camera, in order to film video footage of interest together with the camera motion. We illustratesthat this additional camera motion/trajectory channel can be embedded into a hybrid framework by interleaving an iterative blinddeconvolution and warping based optical flow scheme. Our method yields improved accuracy within three other state-of-the-artbaselines given our proposed ground truth blurry sequences; and several other realworld sequences filmed by our imaging system.

Keywords: Optical Flow, Computer Vision, Image Deblurring, Directional Filtering, RGB-Motion Imaging

1. Introduction

Optical flow estimation has been widely applied to computervision applications, e.g. segmentation, image deblurring andstabilization, etc. In many cases, optical flow is often estimatedon the videos captured by a shaking camera. Those footagesmay contain a significant amount of camera blur that bring ad-ditional difficulties into the traditional variational optical flowframework. It is because such blur scenes often lead to a factthat a pixel may match multiple pixels between image pair. Itfurther violates the basic assumption – intensity constancy – ofthe optical flow framework.

In this paper, we investigate the issue of how to precisely esti-mate optical flow from a blurry video footage. We observe thatthe blur kernel between neighboring frames may be near linear,which can be parameterized using linear elements of the cameramotion. In this case, the camera trajectory can be informatic toenhance the image deblurring within a variational optical flowframework. Based on this observation, our major contributionin this paper is to utilise an RGB-Motion Imaging System – anRGB sensor combined with a 3D pose&position tracker – in or-der to propose: (A) an iterative enhancement process for camerashake blur estimation which encompasses the tracked cameramotion (Sec. 3) and a Directional High-pass Filter (Sec. 4 andSec. 7.2); (B) a Blur-Robust Optical Flow Energy formulation(Sec. 6); and (C) a hybrid coarse-to-fine framework (Sec. 7) forcomputing optical flow in blur scenes by interleaving an itera-tive blind deconvolution process and a warping based minimisa-tion scheme. In the evaluation section, we compare our methodto three existing state-of-the-art optical flow approaches on ourproposed ground truth sequences (Fig. 1, blur and baseline blur-free equivalents) and also illustrate the practical benefit of ouralgorithm given realworld cases.

0

3

6

(Pix.)

1I

2I

Ours Error Map, Ours

Portz et al. Error Map, Portz et al.

Figure 1: Visual comparison of our method to Portz et al. [1] on our groundtruth benchmark Grove2 with synthetic camera shake blur. First Column: theinput images; Second Column: the optical flow fields calculated by our methodand the baseline; Third Column: the RMS error maps against the ground truth.

2. Related Work

Camera shake blur often occurs during fast camera move-ment in low-light conditions due to the requirement of adopt-ing a longer exposure. Recovering both the blur kernel andthe latent image from a single blurred image is known as BlindDeconvolution which is an inherently ill-posed problem. Choand Lee [2] propose a fast deblurring process within a coarse-to-fine framework (Cho&Lee) using a predicted edge map as aprior. To reduce the noise effect in this framework, Zhong etal. [3] introduce a pre-filtering process which reduces the noisealong a specific direction and preserves the image informationin other directions. Their improved framework provides highquality kernel estimation with a low run-time but shows diffi-culties given combined object and camera shake blur.

To obtain higher performance, a handful of combined hard-ware and software-based approaches have also been proposed

Preprint submitted to Neurocomputing March 8, 2016

arX

iv:1

603.

0225

3v1

[cs

.CV

] 7

Mar

201

6

Page 2: Blur Robust Optical Flow using Motion Channel

Blur Kernel

GT Camera Trajectory

Time

Tracked Camera Motion

t t t t

0I

2I

3I

0I

1I

2I

3I

1M 2M

1I12M

1I 2I

1M 2M

(c) Image Channel and Close Up

(b) Tracked Camera Motion and Close Up

(d) RGB-Motion Data Capture Process(a) System Setup

RGB Sensor

3D Pose&Position Tracker

Figure 2: RGB-Motion Imaging System. (a): Our system setup using a combined RGB sensor and 3D Pose&Position Tracker. (b): The tracked 3D camera motionin relative frames. The top-right box is the average motion vector – which has similar direction to the blur kernel. (c): Images captured from our system. Thetop-right box presents the blur kernel estimated using [2]. (d): The internal process of our system where the ∆t presents the exposure time.

for image deblurring. Tai et al. [4] introduce a hybrid imag-ing system that is able to capture both video at high frame rateand a blurry image. The optical flow fields between the videoframes are utilised to guide blur kernel estimation. Levin etal. [5] propose to capture a uniformly blurred image by con-trolling the camera motion along a parabolic arc. Such uniformblur can then be removed based on the speed or direction ofthe known arc motion. As a complement to Levin el al.’s [5]hardware-based deblurring algorithm, Joshi et al. [6] apply in-ertial sensors to capture the acceleration and angular velocityof a camera over the course of a single exposure. This extrainformation is introduced as a constraint in their energy optimi-sation scheme for recovering the blur kernel. All the hardware-assisted solutions described provide extra information in addi-tion to the blurry image, which significantly improves overallperformance. However, the methods require complex electronicsetups and the precise calibration.

Optical flow techniques are widely studied and adoptedacross computer vision because of dense image correspon-dences they provide. Such dense tracking is important forother fundamental research topics e.g. 3D reconstruction [7]and visual effects [8, 9], etc. In the last two decades, theoptical flow model has evolved extensively – one landmarkwork being the variational model of Horn and Schunck [10]where the concept of Brightness Constancy is proposed. Un-der this assumption, pixel intensity does not change spatio-temporally, which is, however, often weakened in realworldimages because of natural noise. To address this issue, somecomplementary concepts have been developed to improve per-formance given large displacements [11], taking advantage offeature-rich surfaces [12] and adapting to nonrigid deforma-tion in scenes [13, 14, 15, 16, 17, 18]. However, flow ap-proaches that can perform well given blurred scenes – where theBrightness Constancy is usually violated – are less common.Of the approaches that do exist, Schoueri et al. [19] performa linear deblurring filter before optical flow estimation whilePortz et al. [1] attempt to match un-uniform camera motionbetween neighbouring input images. Whereas the former ap-proach may be limited given nonlinear blur in realworld scenes;the latter requires two extra frames to parameterise the motion-induced blur. Regarding non optical-flow based methods, Yuanet al. [20] align a blurred image to a sharp one by predefining

an affine image transform with a blur kernel. Similarly HaCo-hen et al. [21] achieve alignment between a blurred image anda sharp one by embedding deblurring into the correspondenceestimation. Li et al. [16] present an approach to solve the imagedeblurring and optical flow simultaneously by using the RGB-Motion imaging.

3. RGB-Motion Imaging System

Camera shake blur within video footage is typically due tofast camera motion and/or long exposure time. In particular,such blur can be considered as a function of the camera trajec-tory supplied to image space during the exposure time ∆t. Ittherefore follows that knowledge of the actual camera motionbetween image pairs can provide significant information whenperforming image deblurring [6, 5].

In this paper, we propose a simple and portable setup(Fig. 2(a)), combining an RGB sensor and a 3D pose&positiontracker (SmartNav by NaturalPoint Inc.) in order to capturecontinuous scenes (video footage) along with real-time camerapose&position information. Note that the RGB sensor could beany camera or a Kinect sensor – A Canon EOS 60D is applied inour implementation to capture 1920 × 1080 video at frame rateof 24 FPS. Furthermore, our tracker is proposed to provide therotation (yaw, pitch and roll), translation and zoom informationwithin a reasonable error range (2 mm). To synchronise thistracker data and the image recording, a real time collaboration(RTC) server [22] is built using the instant messaging protocolXMPP (also known as Jabber1) which is designed for message-oriented communication based on XML, and allows real-timeresponses between different messaging channels or any sig-nal channels that can be transmitted and received in messageform. In this case, a time stamp is assigned to the received mes-sage package by the central timer of the server. Those messagepackages are synchronised if they contain nearly the same timestamp. We consider the Jabber for synchronisation because ofits opensource nature and the low respond delay (around 10ms).

Assuming objects have similar depth within the same scene(a common assumption in image deblurring which will be dis-

1http://www.jabber.org/

2

Page 3: Blur Robust Optical Flow using Motion Channel

cussed in our future work), the tracked 3D camera motion inimage coordinates can be formulated as:

M j =1n

∑x

K([R|T ] X j+1 − X j

)(1)

where M j represents the average of the camera motion vec-tors from the image j to image j + 1. X denotes the 3D positionof the camera while x = (x, y)T is a pixel location and n rep-resents the number of pixels in an image. K represents the 3Dprojection matrix while R and T denote the rotation and transla-tion matrices respectively of tracked camera motion in the im-age domain. All these information K, R and T is computedusing Optitrack’s Camera SDK2 (version 1.2.1). Fig 2(b,c)shows sample data (video frames and camera motion) capturedfrom our imaging system. It is observed that blur from the re-alworld video is near linear due to the relatively high samplingrate of the camera. The blur direction can therefore be approx-imately described using the tracked camera motion. Let thetracked camera motion M j = (r j, θ j)T be represented in polarcoordinates where r j and θ j denote the magnitude and direc-tional component respectively. j is a sharing index betweentracked camera motion and frame number. In addition, we alsoconsider the combined camera motion vector of neighbouringimages as shown in Fig 2(d), e.g. M12 = M1 + M2 whereM12 = (r12, θ12) denotes the combined camera motion vectorfrom image 1 to image 3. As one of our main contributions,these real-time motion vectors are proposed to provide addi-tional constraints for blur kernel enhancement (Sec. 7) withinour framework.

4. Blind Deconvolution

The motion blur process can commonly be formulated:

I = k ⊗ l + n (2)

where I is a blurred image and k represents a blur kernel w.r.t.a specific Point Spread Function. l is the latent image of I; ⊗denotes the convolution operation and n represents spatial noisewithin the scene. In the blind deconvolution operation, both kand l are estimated from I, which is an ill-posed (but extensivelystudied) problem. A common approach for blind deconvolutionis to solve both k and l in an iterative framework using a coarse-to-fine strategy:

k = argmink‖I − k ⊗ l‖ + ρ(k), (3)l = argminl‖I − k ⊗ l‖ + ρ(l). (4)

where ρ represents a regularization that penalizes spatialsmoothness with a sparsity prior [2], and is widely used in re-cent state-of-the-art work [23, 12]. Due to noise sensitivity,low-pass and bilateral filters [24] are typically employed beforedeconvolution. Eq. 5 denotes the common definition of an op-timal kernel from a filtered image.

2http://www.naturalpoint.com/optitrack

k f = argmink f∥∥∥(k ⊗ l + n) ⊗ f − k f ⊗ l

∥∥∥ + ρ(k f )

≈ argmink f

∥∥∥l ⊗ (k ⊗ f − k f )∥∥∥ = k ⊗ f (5)

where k represents the ground truth blur kernel, f is a fil-ter, and k f denotes the optimal blur kernel from the filtered im-age I ⊗ f . The low-pass filtering process improves deconvolu-tion computation by removing spatially-varying high frequencynoise but also results in the removal of useful information whichyields additional errors over object boundaries. To preserve thisuseful information, we introduce a directional high-pass filterthat utilises our tracked 3D camera motion.

5. Directional High-pass Filter

Detail enhancement using directional filters has been provedeffective in several areas of computer vision [3]. Here we definea directional high-pass filter as:

fθ ⊗ I(x) = m∫

g(t)I(x + tΘ)dt (6)

where x = (x, y)T represents a pixel position and g(t) =

1− exp−t2/2σ2 denotes a 1D Gaussian based high-pass func-tion. Θ = (cos θ, sin θ)T controls the filtering direction along θ.m is a normalization factor defined as m =

(∫g(t)dt

)−1. The

filter fθ is proposed to preserve overall high frequency detailsalong direction θ without affecting blur detail in orthogonal di-rections [25]. Given a directionally filtered image bθ = fθ⊗I(x),the optimal blur kernel is defined (Eq 5) as kθ = k ⊗ fθ. Fig. 3demonstrates that noise or object motion within a scene usu-ally results in low frequency noise in the estimated blur kernel(Cho&Lee [2]). This low frequency noise can be removed byour directional high-pass filter while preserving major blur de-tails. In our method, this directional high-pass filter is supple-mented into the Cho&Lee [2] framework using a coarse-to-finestrategy in order to recover high quality blur kernels for use inour optical flow estimation (Sec. 7.2).

6. Blur-Robust Optical Flow Energy

Within a blurry scene, a pair of adjacent natural images maycontain different blur kernels, further violating Brightness Con-stancy. This results in unpredictable flow error across the dif-ferent blur regions. To address this issue, Portz et al. pro-posed a modified Brightness Constancy term by matching theun-uniform blur between the input images. As one of our maincontributions, we extend this assumption to a novel Blur Gradi-ent Constancy term in order to provide extra robustness againstillumination change and outliers. Our main energy function isgiven as follows:

E(w) = EB(w) + γES (w) (7)

A pair of consecutively observed frames from an image se-quence is considered in our algorithm. I1(x) represents the cur-rent frame and its successor is denoted by I2(x) where I∗ =

3

Page 4: Blur Robust Optical Flow using Motion Channel

Blur Kernel

Blur Direction

Filter Direction

Blur PatternClear Pattern GT Kernel Cho&Lee Cho&Lee + Directional Filter

Figure 3: Directional high-pass filter for blur kernel enhancement. Given the blur direction θ, a directional high-pass filter along θ + π/2 is applied to preserve blurdetail in the estimated blur kernel.

k∗ ⊗ l∗ and I∗, l∗ : Ω ⊂ R3 → R represent rectangular im-ages in the RGB channel. Here l∗ is latent image and k∗ denotesthe relative blur kernel. The optical flow displacement betweenI1(x) and I2(x) is defined as w = (u, v)T . To match the un-uniform blur between input images, the blur kernel from eachinput image is applied to the other. We have new blur imagesb1 and b2 as follows:

b1 = k2 ⊗ I1 ≈ k2 ⊗ k1 ⊗ l1 (8)b2 = k1 ⊗ I2 ≈ k1 ⊗ k2 ⊗ l2 (9)

Our energy term encompassing Brightness and GradientConstancy relates to b1 and b2 as follows:

EB(w) =

∫Ω

φ(‖b2(x + w) − b1(x)‖2

+ α ‖∇b2(x + w) − ∇b1(x)‖2)dx (10)

The term ∇ = (∂xx, ∂yy)T presents a spatial gradient andα ∈ [0, 1] denotes a linear weight. The smoothness regulariserpenalizes global variation as follows:

ES (w) =

∫Ω

φ(‖∇u‖2 + ‖∇v‖2)dx (11)

where we apply the Lorentzian regularisation φ(s) = log(1 +

s2/2ε2) to both the data term and smoothness term. In our case,the image properties, e.g. small details and edges, are brokenby the camera blur, which leads to additional errors in thoseregions. We suppose to apply strong boundary preservationeven the non-convex Lorentzian regularisation may bring theextra difficulty to the energy optimisation (More analysis canbe found in Li et al. [14]). In the following section, our opticalflow framework is introduced in detail.

7. Optical Flow Framework

Our overall framework is outlined in Algorithm 1 based onan iterative top-down, coarse-to-fine strategy. Prior to minimiz-ing the Blur-Robust Optical Flow Energy (Sec. 7.4), a fast blinddeconvolution approach [2] is performed for pre-estimation ofthe blur kernel (Sec. 7.1), which is followed by kernel refine-ment using our Directional High-pass Filter (Sec. 7.2). Allthese steps are detailed in the following subsections.

Algorithm 1: Blur-Robust Optical Flow FrameworkInput : A image pair I1, I2 and camera motion θ1, θ2, θ12Output : Optimal optical flow field w1: A n-level top-down pyramid is built with the level index i2: i← 03: li1 ← Ii

1, li2 ← Ii2

4: ki1 ← 0, ki

2 ← 0, wi ← (0, 0)T

5: for coarse to fine do6: i← i + 17: Resize ki

1,2, li1,2, Ii

1,2 and wi with the ith scale8: foreach ∗ ∈ 1, 2 do9: ki

∗ ← IterBlindDeconv ( li∗, Ii∗ )

10: ki∗ ← DirectFilter ( ki

∗, θ1, θ2, θ12 )11: li∗ ← NonBlindDeconvolve ( ki

∗, Ii∗ )

12: endfor13: bi

1 ← Ii1 ⊗ ki

2, bi2 ← Ii

2 ⊗ ki1

14: dwi ← Energyoptimisation ( bi1, b

i2,w

i )15: wi ← wi + dwi

16: endfor

7.1. Iterative Blind DeconvolutionCho and Lee [2] describe a fast and accurate approach

(Cho&Lee) to recover the unique blur kernel. As shown in Al-gorithm 1, we perform a similar approach for the pre-estimationof the blur kernel k within our iterative process, which involvestwo steps of prediction and kernel estimation. Given the la-tent image l estimated from the consecutively coarser level, thegradient maps ∆l = ∂xl, ∂yl of l are calculated along the hor-izontal and vertical directions respectively in order to enhancesalient edges and reduce noise in featureless regions of l. Next,the predicted gradient maps ∆l as well as the gradient map ofthe blurry image I are utilised to compute the pre-estimated blurkernel by minimizing the energy function as follows:

k = argmink

∑I∗,l∗

ω∗ ‖I∗ − k ⊗ l∗‖2 + δ ‖k‖2

(I∗, l∗) ∈ (∂xI, ∂xl), (∂yI, ∂yl), (∂xxI, ∂xxl),(∂yyI, ∂yyl), (∂xyI, (∂x∂y + ∂y∂x)l/2) (12)

where δ denotes the weight of Tikhonov regularization andω∗ ∈ ω1, ω2 represents a linear weight for the derivativesin different directions. Both I and l are propagated from thenearest coarse level within the pyramid. To minimise this en-ergy Eq. (12), we follow the inner-iterative numerical schemeof [2] which yields a pre-estimated blur kernel k.

4

Page 5: Blur Robust Optical Flow using Motion Channel

7.2. Directional High-pass Filtering

Once the pre-estimated kernel k is obtained, our DirectionalHigh-pass Filters are applied to enhance the blur informationby reducing noise in the orthogonal direction of the trackedcamera motion. Although our RGB-Motion Imaging Systemprovides an intuitively accurate camera motion estimation, out-liers may still exist in the synchronisation. We take into ac-count the directional components θ1, θ2, θ12 of two consecu-tive camera motions M1 and M2 as well as their combinationM12 (Fig. 2(d)) for extra robustness. The pre-estimated blurkernel is filtered along its orthogonal direction as follows:

k =∑β∗,θ∗

β∗k ⊗ fθ∗+π/2 (13)

where β∗ ∈ 1/2, 1/3, 1/6 linearly weights the contri-bution of filtering in different directions. Note that twoconsecutive images I1 and I2 are involved in our frame-work where the former accepts the weight set (β∗, θ∗) ∈(1/2, θ1), (1/3, θ2), (1/6, θ12) while the other weight set(β∗, θ∗) ∈ (1/3, θ1), (1/2, θ2), (1/6, θ12) is performed for thelatter. This filtering process yields an updated blur kernel kwhich is used to update the latent image l within a non-blinddeconvolution [3]. Note that the convolution operation is com-putationally expensive in the spatial domain, we consider anequivalent filtering scheme in the frequency domain in the fol-lowing subsection.

7.3. Convolution for Directional Filtering

Our proposed directional filtering is performed as convolu-tion operation in the spatial domain, which is often highly ex-pensive in computation given large image resolutions. In ourimplementations, we consider a directional filtering scheme inthe frequency domain where we have the equivalent form offiltering model Eq. (6) as follows:

KΘ(u, v) = K(u, v)FΘ(u, v) (14)

where KΘ is the optimal blur kernel in the frequency domainwhile K and FΘ present the Fourier Transform of the blur ker-nel k and our directional filter fθ respectively. Thus, the opti-mal blur kernel kθ in the spatial domain can be calculated askθ = IDFT[KΘ] using Inverse Fourier Transform. In this case,the equivalent form of our directional high-pass filter in the fre-quency domain is defined as follows:

FΘ(u, v) = 1 − exp−L2(u, v)/2σ2

(15)

where the line function L(u, v) = u cos θ + v sin θ controlsthe filtering process along the direction θ while σ is the stan-dard deviation for controlling the strength of the filter. Pleasenote that other more sophisticated high-pass filters could alsobe employed using this directional substitution L. Even thoughthis consumes a reasonable proportion of computer memory,convolution in the frequency domain O(N log2 N) is faster thanequivalent computation in the spatial domain O(N2).

Having performed blind deconvolution and directional filter-ing (Sec. 7.1, 7.2 and 7.3), two updated blur kernels ki

1 and ki2

on the ith level of the pyramid are obtained from input imagesIi1 and Ii

2 respectively, which is followed by the uniform blurimage bi

1 and bi2 computation using Eq. (9). In the following

subsection, Blur-Robust Optical Flow Energy optimisation onbi

1 and bi1 is introduced in detail.

7.4. Optical Flow Energy optimisationAs mentioned in Sec. 6, our blur-robust energy is continu-

ous but highly nonlinear. minimisation of such energy functionis extensively studied in the optical flow community. In thissection, a numerical scheme combining Euler-Lagrange Equa-tions and Nested Fixed Point Iterations is applied [11] to solveour main energy function Eq. 7. For clarity of presentation, wedefine the following mathematical abbreviations:

bx = ∂xb2(x + w) byy = ∂yyb2(x + w)by = ∂yb2(x + w) bz = b2(x + w) − b1(x)bxx = ∂xxb2(x + w) bxz = ∂xb2(x + w) − ∂xb1(x)bxy = ∂xyb2(x + w) byz = ∂yb2(x + w) − ∂yb1(x)

At the first phase of energy minimization, a system is builtbased on Eq. 7 where Euler-Lagrange is employed as follows:

φ′b2z + α(b2

xz + b2yz) · bxbz + α(bxxbxz + bxybyz)

−γφ′(‖∇u‖2 + ‖∇v‖2) · ∇u = 0 (16)

φ′b2z + α(b2

xz + b2yz) · bybz + α(byybyz + bxybxz)

−γφ′(‖∇u‖2 + ‖∇v‖2) · ∇v = 0 (17)

An n-level image pyramid is then constructed from the topcoarsest level to the bottom finest level. The flow field is initial-ized as w0 = (0, 0)T on the top level and the outer fixed pointiterations are applied on w. We assume that the solution wi+1

converges on the i + 1 level. We have:

φ′(bi+1z )2 + α(bi+1

xz )2 + α(bi+1yz )2

·bixbi+1

z + α(bixxbi+1

xz + bixybi+1

yz )

−γφ′(∥∥∥∇ui+1

∥∥∥2+

∥∥∥∇vi+1∥∥∥2

) · ∇ui+1 = 0 (18)

φ′(bi+1z )2 + α(bi+1

xz )2 + α(bi+1yz )2

·biybi+1

z + α(biyybi+1

yz + bixybi+1

xz )

−γφ′(∥∥∥∇ui+1

∥∥∥2+

∥∥∥∇vi+1∥∥∥2

) · ∇vi+1 = 0 (19)

Because of the nonlinearity in terms of φ′, bi+1∗ , the system

(Eqs. 18, 19) is difficult to solve by linear numerical methods.We apply the first order Taylor expansions to remove these non-linearity in bi+1

∗ , which results in:

bi+1z ≈ bi

z + bixdui + bi

ydvi,

bi+1xz ≈ bk

xz + bixxdui + bi

xydvi,

bi+1yz ≈ bk

yz + bixydui + bi

yydvi.

5

Page 6: Blur Robust Optical Flow using Motion Channel

Based on the coarse-to-fine flow assumption of Brox etal. [11] w.r.t. ui+1 ≈ ui + dui and vi+1 ≈ vi + dvi where theunknown flow field on the next level i + 1 can be obtained usingthe flow field and its incremental from the current level i. Thenew system can be presented as follows:

(φ′)iB · b

ix(bi

z + bixdui + bi

ydvi)

+αbixx(bi

xz + bixxdui + bi

xydvi)

+αbixy(bi

yz + bixydui + bi

yydvi)

−γ(φ′)iS · ∇(ui + dui) = 0 (20)

(φ′)iB · b

iy(bi

z + bixdui + bi

ydvi)

+αbiyy(bi

yz + bixydui + bi

yydvi)

+αbixy(bi

xz + bixxdui + bi

xydvi)

−γ(φ′)iS · ∇(vi + dvi) = 0 (21)

where the terms (φ′)iB and (φ′)i

S contained φ provide robust-ness to flow discontinuity on the object boundary. In addi-tion, (φ′)i

S is also regularizer for a gradient constraint in motionspace. Although we fixed wi in Eqs. 20 and 21, the nonlinear-ity in φ′ leads to the difficulty of solving the system. The innerfixed point iterations are applied to remove this nonlinearity:dui, j and dvi, j are assumed to converge within j iterations byinitializing dui,0 = 0 and dvi,0 = 0. Finally, we have the linearsystem in dui, j+1 and dvi, j+1 as follows:

(φ′)i, jB · b

ix(bi

z + bixdui, j+1 + bi

ydvi, j+1)

+αbixx(bi

xz + bixxdui, j+1 + bi

xydvi, j+1)

+αbixy(bi

yz + bixydui, j+1 + bi

yydvi, j+1)

−γ(φ′)i, jS · ∇(ui + dui, j+1) = 0 (22)

(φ′)i, jB · b

iy(bi

z + bixdui, j+1 + bi

ydvi, j+1)

+αbiyy(bi

yz + bixydui, j+1 + bi

yydvi, j+1)

+αbixy(bi

xz + bixxdui, j+1 + bi

xydvi, j+1)

−γ(φ′)i, jS · ∇(vi + dvi, j+1) = 0 (23)

where (φ′)i, jB denotes a robustness factor against flow discon-

tinuity and occlusion on the object boundaries. (φ′)i, jS represents

the diffusivity of the smoothness regularization.

(φ′)i, jB = φ′(bi

z + bixdui, j + bi, j

y dvi, j)2

+ α(bixz + bi

xxdui, j + bixydvi, j)2

+ α(biyz + bi

xydui, j + biyydvi, j)2

(φ′)i, jS = φ′

∥∥∥∇(ui + dui, j)∥∥∥2

+∥∥∥∇(vi + dvi, j)

∥∥∥2

In our implementation, the image pyramid is constructedwith a downsampling factor of 0.75. The final linear systemin Eq. (22,23) is solved using Conjugate Gradients within 45iterations.

Algorithm 2: Auto Blur-Robust Optical Flow FrameworkInput : A image pair I1, I2 Without camera motionOutput : Optimal optical flow field w1: A n-level top-down pyramid is built with the level index i2: i← 03: li1 ← Ii

1, li2 ← Ii2

4: ki1 ← 0, ki

2 ← 0, wi ← (0, 0)T , θi = 05: for coarse to fine do6: i← i + 17: Resize ki

1,2, li1,2, Ii

1,2 and wi with the ith scale8: foreach ∗ ∈ 1, 2 do9: ki

∗ ← IterBlindDeconv ( li∗, Ii∗ )

10: ki∗ ← DirectFilter ( ki

∗, θi )

11: li∗ ← NonBlindDeconvolve ( ki∗, I

i∗ )

12: endfor13: bi

1 ← Ii1 ⊗ ki

2, bi2 ← Ii

2 ⊗ ki1

14: dwi ← EnergyOptimisation ( bi1, b

i2,w

i )15: wi ← wi + dwi

16: θi ← CameraMotionEstimation(wi)17: endfor

7.5. Alternative Implementation with Automatic Camera Mo-tion θ∗ Estimation

Alternative to using our assisted tracker, we also provide anadditional implementation by using the camera motion θ∗ esti-mated generically from the flow field. As shown in Algorithm2, the system does not take the camera motion (θ∗) as inputbut computes it (CameraMotionEstimation) genericallyat every level of the image pyramid.

Ai← AffineEstimation(x, x + wi)

θi ← AffineToMotionAngle(Ai) (24)

On each level, we calculate the Affine Matrix from Ii1 to Ii

2 us-ing the correspondences x → x + wi and RANSAC. The trans-lation information from Ai is then normalized and converted tothe angle format θi. In this case, our DirectionalFilter is alsodowngraded to consider one direction θi for each level. In thenext section, we quantitatively compare our method to otherpopular baselines.

8. Evaluation

In this section, we evaluate our method on both synthetic andrealworld sequences and compare its performance against threeexisting state-of-the-art optical flow approaches of Xu et al.’sMDP [12], Portz et al.’s [1] and Brox et al.’s [11] (an imple-mentation of [26]). MDP is one of the best performing opti-cal flow methods given blur-free scenes, and is one of the top3 approaches in the Middlebury benchmark [27]. Portz et al.’smethod represents the current state-of-the-art in optical flow es-timation given object blur scenes while Brox et al.’s contains asimilar optimisation framework and numerical scheme to Portzet al.’s, and ranks in the midfield of the Middlebury benchmarksbased on overall average. Note that all three baseline methods

6

Page 7: Blur Robust Optical Flow using Motion Channel

1I0I 2I 3I

1I0I 2I 3I

1I0I 2I 3I

1I0I 2I 3I

Figure 4: The synthetic blur sequences with the blur kernel, tracked camera motion direction and ground truth flow fields. From Top To Bottom: sequences ofRubberWhale, Urban2, Hydrangea and Urban2.

are evaluated using their default parameters setting; all exper-iments are performed using a 2.9Ghz Xeon 8-cores, NVIDIAQuadro FX 580, 16Gb memory computer.

In the following subsections, we compare our algorithm(moBlur) and four different implementations (auto, nonGC,nonDF and nonGCDF) against the baseline methods. autodenotes the implementation using the automatic camera mo-tion estimation scheme (Algorithm 2); nonGC represents theimplementation without the Gradient Constancy term whilenonDF denotes an implementation without the directional fil-tering process. nonGCDF is the implementation with neither ofthese features. The results show that our Blur-Robust OpticalFlow Energy and Directional High-pass Filter significantly im-prove algorithm performance for blur scenes in both syntheticand realworld cases.

8.1. Middlebury Dataset with camera shake blurOne advance for evaluating optical flow given scenes with

object blur is proposed by Portz et al. [1] where syntheticGround Truth (GT) scenes are rendered with blurry moving ob-jects against a blur-free static/fixed background. However, theiruse of synthetic images and controlled object trajectories lead toa lack of global camera shake blur, natural photographic proper-ties and real camera motion behaviour. To overcome these lim-itations, we render four sequences with camera shake blur andcorresponding GT flow-fields by combining sequences from theMiddlebury dataset [27] with blur kernels estimated using oursystem.

In our experiments we select the sequences Grove2, Hy-drangea, RubberWhale and Urban2 from the Middleburydataset. For each of them, four adjacent frames are selectedas latent images along with the GT flow field wgt (supplied by

Middlebury) for the middle pair. 40 × 40 blur kernels are thenestimated [2] from realworld video streams captured using ourRGB-Motion Imaging System. As shown in Fig. 4, those ker-nels are applied to generate blurry images denoted by I0, I1, I2and I3 while the camera motion direction is set for each framebased on the 3D motion data. Although the wgt between latentimages can be utilised for the evaluation on relative blur im-ages I∗ [28, 29], strong blur can significantly violate the origi-nal image intensity, which leads to a multiple correspondencesproblem: a point in the current image corresponds to multi-ple points in the consecutive image. To remove such multi-ple correspondences, we sample reasonable correspondence setw | w ⊂ wgt, |I2(x + w) − I1(x)| < ε to use as the GT for theblur images I∗ where ε denotes a predefined threshold. Oncewe obtain w, both Average Endpoint Error (AEE) and AverageAngle Error (AAE) tests [27] are considered in our evaluation.The computation is formulated as follows:

AEE =1n

∑x

√(u − u)2 + (v − v)2 (25)

AAE =1n

∑x

cos−1(

1.0 + u × u + v × v√

1.0 + u2 + v2√

1.0 + u2 + v2

)(26)

where w = (u, v)T and w = (u, v)T denotes the baselineflow field and the ground truth flow field (by removing multiplecorrespondences) respectively while n presents the number ofground truth vectors in w. The factor 1.0 in AAE is an arbitraryscaling constant to convert the units from pixels to degrees [27].Fig. 5(a) Left shows AEE (in pixel) and AAE (in degree) testson our four synthetic sequences. moBlur and nonGC lead bothAEE and AAE tests in all the trials. Both Brox et al. and MDPyield significant error in Hydrangea, RubberWhale and Urban2

7

Page 8: Blur Robust Optical Flow using Motion Channel

AEE Test on RubberWhale Seq.

Salt&Pepper Noise Level (%)

AEE

(pix

.)

0 5 10 15 20 25 30 35 40 45

1

2

3

4

5

6

7

moBlurnonDFnonGCnonDFGCPortz et al.Brox et al.MDP

Grove2AEE

Hydrangea Rub.Whale Urban2AAE AEE AAE AEE AAE AEE AAE

AEE/AAE

Portz et al.

nonDFGC

1.14 4 4.11 4

moBlur-nonGC

moBlur-nonDF

Brox et al.

1.62 7 5.14 7

0.49 2 2.38 2

1.52 6 4.96 6

1.24 5 4.53 5

2.62 6 3.55 6

2.28 5 3.21 4

0.95 2 2.23 2

1.83 3 3.00 3

2.26 4 3.47 5

3.12 6 8.18 6

1.25 4 7.71 4

0.64 2 3.71 2

1.12 3 6.45 3

2.44 5 7.98 5

3.44 6 5.10 4

2.98 5 5.44 6

1.54 2 3.03 2

2.50 3 5.19 5

2.92 4 4.60 3

Ours, moBlur 0.47 1 2.34 1 0.67 1 2.19 1 0.62 1 3.67 1 1.36 1 2.87 1

Xu et al., MDP 1.06 3 3.46 3 3.40 7 3.55 6 3.70 6 8.21 7 5.62 7 6.82 7

TimeCost

85

28

33

39

27

45

466

(a) Left: Quantitative Average Endpoint Error (AEE), Average Angle Error (AAE) and Time Cost (in second) comparisons on our synthetic sequenceswhere the subscripts show the rank in relative terms. Right: AEE measure on RubberWhale by ramping up the noise distribution.

0

3

6

(Pix.)

moBlur nonGC nonDF nonDFGC Portz et al. Brox et al. Xu et al. MDP

0

3

6

(Pix.)Grove2GT Flow Field

Rub.WhaleGT Flow Field

0

3

6

(Pix.)

0

3

6

(Pix.)

Urban2GT Flow Field

HydrangeaGT Flow Field

(b) Visual comparison on sequences RubberWhale, Urban2, Hydrangea and Urban2 by varying baseline methods. For each sequence, First Row: opticalflow fields from different methods. Second Row: the error maps against the ground truth.

Figure 5: Quantitative evaluation on four synthetic blur sequences with both camera motion and ground truth.

because those sequences contain large textureless regions withblur, which in turn weakens the inner motion estimation pro-cess as shown in Fig. 5(b). Fig. 5(a) also illustrates the aver-age time cost (second per frame) of the baseline methods. Ourmethod gives reasonable performance (45 sec. per frame) com-paring to the state-of-the-art Portz et al. and MDP even an in-ner image deblurring process is involved. Furthermore, Fig 5(a)Right shows the AEE metric for RubberWhale by varying the

distribution of Salt&Pepper noise. It is observed that a highernoise level leads to additional errors for all the baseline meth-ods. Both moBlur and nonGC yield the best performance whilePortz et al. and Brox et al. show a similar rising AEE trendwhen the noise increases.

Fig. 6 shows our quantitative measure by comparing our twoimplementations which use the RGB-Motion Imaging (moBlur)and automatic camera motion estimation scheme (auto, see

8

Page 9: Blur Robust Optical Flow using Motion Channel

0

1

2

3

4

AEE

(pix

.)

AEE Test on All Seqs.moBlurautoPortz

0

2

4

6

8

10

AAE

(deg

rees

)

AAE Test on All Seqs.moBlurautoPortz

Grove2 Hydra. Rub.Wh. Urban2 Grove2 Hydra. Rub.Wh. Urban20 1 2 3 4 5 6 7 8 9 10

Index of Pyramidal Level

0

3

6

9

12

15

18

Angu

lar E

rror (

degr

ees)

Auto Camera Motion ErrorGrove2HydrangeaRub.WhaleUrban2

Figure 6: Quantitative comparison between our implementations using RGB-Motion Imaging (moBlur); and automatic camera motion estimation scheme (auto, seeSec. 7.5). From Left To Right: AEE and AAE tests on all sequences respectively; the angular error of camera motion estimated by auto by varying the pyramidallevels of the input images.

0 10 20 30 40 50 60 70 80 90

AEE Test of moBlur by Varying λ

AEE

(pix

.)

Varying Angle Di. λ (°)0 10 20 30 40 50 60 70 80 90

Varying Angle Di. λ (°)

GT Blur Direction

Input Direction

0

3

6

(Pix.)

= 0 = 60 = 90o o o

AEE Comparison by Varying λ

0.5

1

1.5

2

2.5

3

3.5

0.5

1

1.5

2

2.5

3

3.5 moBlur.Gro2moBlur.HydrmoBlur.RubbmoBlur.Urb2nonDF.Gro2nonDF.HydrnonDF.RubbnonDF.Urb2Portz.Gro2Portz.HydrPortz.RubbPortz.Urb2

(a) sample sequence and error maps (b) Comparion of on/off filter (c) Comparion to Portz et al.A

EE (p

ix.)

Figure 7: AEE measure of our method (moBlur) by varying the input motion directions. (a): the overall measure strategy and error maps of moBlur on sequenceUrban2. (b): the quantitative comparison of moBlur against nonDF by ramping up the angle difference λ. (c): the measure of moBlur against Portz et al. [1].

0I 1I 2I 3I

0I 1I 2I 3I

0I 1I 2I 3I

0I 1I 2I 3I

Figure 8: The realworld sequences captured along the tracked camera motion. From Top To Bottom: sequences of warrior, chessboard, LabDesk and shoes.

Sec. 7.5) respectively. For better observation, we also give thePortz et al. in this measure. We observe that both our im-

plementations outperform Portz et al. in the AEE and AAEtests. Especially the moBlur gives the best accuracy in all tri-

9

Page 10: Blur Robust Optical Flow using Motion Channel

als. The implementation auto yields the less accurate resultsthan the moBlur. It may be because the auto camera motionestimation is affected by ambiguous blur that often caused bymultiple moving objects. To investigate this issue, we plot theangular error by comparing the auto-estimated camera motionto the ground truth on all the sequences (Fig. 6, right end). Weobserve that our automatic camera motion estimation schemeleads to higher errors on the upper/coarser level of the imagepyramid. Even the accuracy is improved on the finer levels butthe error may be accumulated and affect the final result.

In practice, the system may be used in some challengescenes, e.g. fast camera shaking, super high frame rate capture,or even infrared interference, etc. In those cases, the wrongtracked camera motion may be given to some specific frames.To investigate how the tracked camera motion affects the ac-curacy of our algorithm, we compare moBlur to nonDF (ourmethod without directional filtering) and Portz et al. by vary-ing the direction of input camera motion. As shown in Fig. 7(a),we rotate the input camera motion vector with respect to theGT blur direction by an angle of λ degrees. Here λ = 0 rep-resents the ideal situation where the input camera motion hasthe same direction as the blur direction. The increasing λ sim-ulates more errors in the camera motion estimation. Fig. 7(b,c)shows the AEE metric by increasing the λ. We observe thatthe AEE increases during this test. moBlur outperforms thenonDF (moBlur without the directional filter) in both Grove2and RubberWhale while nonDF provides higher performancein Hydrangea when λ is larger than 50. In addition, moBluroutperforms Portz et al. in all trials except Hydrangea wherePortz et al. shows a minor advantage (AEE 0.05) when λ = 90.The rationale behind this experiment is that the wrong cameramotion may yield significant information loss in the directionalhigh-pass filtering. Such information loss harms the deblurringprocess and consequently leads to errors in the optical flow esti-mation. Thus, obtaining precise camera motion is the essentialpart of this system, as well as a potential future research.

8.2. Realworld Dataset

To evaluate our method in the realworld scenes, we capturefour sequences warrior, chessboard, LabDesk and shoes withtracked camera motion using our RGB-Motion Imaging Sys-tem. As shown in Fig. 8, both warrior and chessboard con-tain occlusions, large displacements and depth change while thesequences of LabDesk and shoes embodies the object motionblur and large textureless regions within the same scene. Fig. 9shows visual comparison of our method moBlur against Portz etal. on these realworld sequences. It is observed that our methodpreserves appearance details on the object surface and reduceboundary distortion after warping using the flow field. In addi-tion, our method shows robustness given cases where multipletypes of blur exist in the same scene (Fig.9(b), sequence shoes).

9. Conclusion

In this paper, we introduce a novel dense tracking frameworkwhich interleaves both a popular iterative blind deconvolution;

as well as a warping based optical flow scheme. We also in-vestigate the blur papameterization for the video footages. Inour evaluation, we highlight the advantages of using both theextra motion channel and the directional filtering in the opticalflow estimation for the blurry video footages. Our experimentsalso demonstrated the improved accuracy of our method againstlarge camera shake blur in both noisy synthetic and realworldcases. One limitation in our method is that the spatial invarianceassumption for the blur is not valid in some realworld scenes,which may reduce accuracy in the case where the object depthsignificantly changes. Finding a depth-dependent deconvolu-tion and deep data-driven model would be a challenge for futurework as well.

10. Acknowledgements

We thank Ravi Garg and Lourdes Agapito for providingtheir GT datasets. We also thank Gabriel Brostow and theUCL Vision Group for their helpful comments. The authorsare supported by the EPSRC CDE EP/L016540/1 and CAM-ERA EP/M023281/1; and EPSRC projects EP/K023578/1 andEP/K02339X/1.

References

References

[1] T. Portz, L. Zhang, H. Jiang, Optical flow in the presence of spatially-varying motion blur, in: IEEE Conference on Computer Vision and Pat-tern Recognition (CVPR’12), 2012, pp. 1752–1759.

[2] S. Cho, S. Lee, Fast motion deblurring, ACM Transactions on Graphics(TOG’09) 28 (5) (2009) 145.

[3] L. Zhong, S. Cho, D. Metaxas, S. Paris, J. Wang, Handling noise in singleimage deblurring using directional filters, in: IEEE Conference on Com-puter Vision and Pattern Recognition (CVPR’13), 2013, pp. 612–619.

[4] Y.-W. Tai, H. Du, M. S. Brown, S. Lin, Image/video deblurring usinga hybrid camera, in: IEEE Conference on Computer Vision and PatternRecognition (CVPR’08), 2008, pp. 1–8.

[5] A. Levin, P. Sand, T. S. Cho, F. Durand, W. T. Freeman, Motion-invariantphotography, ACM Transactions on Graphics (TOG’08) 27 (3) (2008) 71.

[6] N. Joshi, S. B. Kang, C. L. Zitnick, R. Szeliski, Image deblurring usinginertial measurement sensors, ACM Transactions on Graphics (TOG’10)29 (4) (2010) 30.

[7] C. Godard, P. Hedman, W. Li, G. J. Brostow, Multi-view reconstructionof highly specular surfaces in uncontrolled environments, in: 3D Vision(3DV), 2015 International Conference on, IEEE, 2015, pp. 19–27.

[8] Z. Lv, A. Tek, F. Da Silva, C. Empereur-Mot, M. Chavent, M. Baaden,Game on, science-how video game technology may help biologists tacklevisualization challenges, PloS one 8 (3) (2013) e57990.

[9] Z. Lv, A. Halawani, S. Feng, H. Li, S. U. Rehman, Multimodal hand andfoot gesture interaction for handheld devices, ACM Transactions on Mul-timedia Computing, Communications, and Applications (TOMM) 11 (1s)(2014) 10.

[10] B. Horn, B. Schunck, Determining optical flow, Artificial intelligence17 (1-3) (1981) 185–203.

[11] T. Brox, A. Bruhn, N. Papenberg, J. Weickert, High accuracy optical flowestimation based on a theory for warping, in: European Conference onComputer Vision (ECCV’04), 2004, pp. 25–36.

[12] L. Xu, S. Zheng, J. Jia, Unnatural l0 sparse representation for naturalimage deblurring, in: IEEE Conference on Computer Vision and PatternRecognition (CVPR’13), 2013, pp. 1107–1114.

[13] W. Li, D. Cosker, M. Brown, An anchor patch based optimisation frame-work for reducing optical flow drift in long image sequences, in: AsianConference on Computer Vision (ACCV’12), Springer, 2012, pp. 112–125.

10

Page 11: Blur Robust Optical Flow using Motion Channel

1I

2I

1I

2I

mo

Blu

rP

ort

z et

al.

Warping Direction

Flow Field Closeups of Warping Results Flow Field Closeups of Warping Results

Warping Direction

(a) Visual comparison on realworld sequences of warrior and chessboard.

Closeups of Warping Results Flow Field Closeups of Warping Results

1I

2I

mo

Blu

rP

ort

z et

al.

Flow Field

Warping Direction

1I

2I

Warping Direction

(b) Visual comparison on realworld sequences of LabDesk and shoes.

Figure 9: Visual comparison of image warping on realworld sequences of warrior, chessboard, LabDesk and shoes, captured by our RGB-Motion Imaging System.

[14] W. Li, D. Cosker, M. Brown, R. Tang, Optical flow estimation using lapla-cian mesh energy, in: IEEE Conference on Computer Vision and PatternRecognition (CVPR’13), IEEE, 2013, pp. 2435–2442.

[15] W. Li, D. Cosker, M. Brown, Drift robust non-rigid optical flow enhance-ment for long sequences, Journal of Intelligent and Fuzzy Systems 0 (0)(2016) 12.

[16] W. Li, Y. Chen, J. Lee, G. Ren, D. Cosker, Robust optical flow estimationfor continuous blurred scenes using rgb-motion imaging and directionalfiltering, in: IEEE Winter Conference on Application of Computer Vision(WACV’14), IEEE, 2014, pp. 792–799.

[17] R. Tang, D. Cosker, W. Li, Global alignment for dynamic 3d morphablemodel construction, in: Workshop on Vision and Language (V&LW’12),2012.

[18] W. Li, Nonrigid surface tracking, analysis and evaluation, Ph.D. thesis,University of Bath (2013).

[19] Y. Schoueri, M. Scaccia, I. Rekleitis, Optical flow from motion blurredcolor images, in: Canadian Conference on Computer and Robot Vision,2009.

[20] L. Yuan, J. Sun, L. Quan, H.-Y. Shum, Progressive inter-scale and intra-scale non-blind image deconvolution 27 (3) (2008) 74.

[21] Y. HaCohen, E. Shechtman, D. B. Goldman, D. Lischinski, Non-rigiddense correspondence with applications for image enhancement, ACMTransactions on Graphics (TOG’11) 30 (4) (2011) 70.

[22] J. Lee, V. Baines, J. Padget, Decoupling cognitive agents and virtual en-vironments, in: Cognitive Agents for Virtual Environments, 2013, pp.17–36.

[23] Q. Shan, J. Jia, A. Agarwala, High-quality motion deblurring from a sin-gle image, ACM Transactions on Graphics (TOG’08) 27 (3) (2008) 73.

[24] Y.-W. Tai, S. Lin, Motion-aware noise filtering for deblurring of noisyand blurry images, in: IEEE Conference on Computer Vision and PatternRecognition (CVPR’12), 2012, pp. 17–24.

[25] X. Chen, J. Yang, Q. Wu, J. Zhao, X. He, Directional high-pass filterfor blurry image analysis, Signal Processing: Image Communication 27(2012) 760–771.

[26] C. Liu, Beyond pixels: exploring new representations and applicationsfor motion analysis, Ph.D. thesis, Massachusetts Institute of Technology(2009).

[27] S. Baker, D. Scharstein, J. Lewis, S. Roth, M. Black, R. Szeliski, Adatabase and evaluation methodology for optical flow, International Jour-nal of Computer Vision (IJCV’11) 92 (2011) 1–31.

[28] D. J. Butler, J. Wulff, G. B. Stanley, M. J. Black, A naturalistic opensource movie for optical flow evaluation, in: European Conference onComputer Vision (ECCV’12), 2012, pp. 611–625.

[29] J. Wulff, D. J. Butler, G. B. Stanley, M. J. Black, Lessons and insightsfrom creating a synthetic optical flow benchmark, in: ECCV Workshop onUnsolved Problems in Optical Flow and Stereo Estimation (ECCVW’12),

11

Page 12: Blur Robust Optical Flow using Motion Channel

2012, pp. 168–177.

12