Robust Direct Visual Odometry using Mutual Information Kumar Shaurya Shankar and Nathan Michael The Robotics Institute Carnegie Mellon University Pittsburgh, PA, 15213 USA Email: {kshaurya,nmichael}@cmu.edu

Abstract—Robust vision-based state estimation in real-world indoor and outdoor environments is a challenging problem due to the combination of drastic lighting changes and limited dynamic range of commodity cameras. This limitation is at odds with the fundamental constancy of illumination assumption made in image-intensity based tracking methodologies. We present and experimentally validate a Mutual Information (MI) based dense rigid body tracking algorithm that is demonstrably robust to drastic illumination changes, and compare the performance of this algorithm to a canonical Sum of Squared Differences based Lucas-Kanade tracking formulation. Further, we propose a novel approach that combines the robustness benefits of information-based measures and the speed of traditional intensity based Lucas-Kanade tracking for robust state estimation in real-time.

I. M OTIVATION AND R ELATED W ORK Vision-based state estimation is an essential component for autonomous vehicles operating in GPS-denied environments. Significant progress in vision-based SLAM has been achieved in the past few years as evident by popular benchmarks [1], [2]. Nonetheless, consistently robust performance over extended sequences with uncontrolled illumination conditions remains a challenging and unsolved problem. Examples include, but are not limited to: stark contrast between adjacent image areas, changes in global lighting, and transitions between well-lit and shaded areas. Inevitable control delays in camera auto-exposure gains further exacerbate their adverse impact. Our goal in this work is to develop a robust visual odometry algorithm capable of addressing the plethora of nonlinear and nonlocal deformations of intensities commonly found in real world search and rescue environments. A first step towards robustness is to ensure the use of as much information from a single image as possible. Dense direct methods [3] have demonstrated more reliable camera tracking results [4], [5] in comparison to their sparse counterparts [6]. This robustness has been primarily demonstrated with respect to blur and fast camera motions [4], which are known to degrade the extraction and precise localization of sparse keypoints [7]. While dense direct methods have been shown to be robust against image blur, they lack this robustness in face of challenging illumination. This is due to the reliance on

Fig. 1. The proposed method is able to estimate odometry even in extreme illumination changes where conventional tracking methods fail. Here, we track the camera from a brightly lit region measured at 61.0 Lux (Top Left) to the same static scene over a linear ramp in intensity measured at 9.8 Lux (Top Right). The algorithm tries to warp the current image (Turquoise border) to the reference image (Pink Border) such that it minimizes cost. We demonstrate performance using residual images that visualize the absolute difference between a reference and warped image. When using Sum of Squared Differences, the tracking fails (Bottom Left) due to underexposure. Our algorithm (Bottom Right) maintains tracking due to it comparing the distribution of intensities, rather than their absolute values.

the brightness constancy assumption, which requires image intensities to remain constant as a function of geometric transformations. In variable lighting scenarios, however, the brightness constancy assumption is seldom satisfied [8], thereby limiting the applicability of direct methods. A strategy previously presented in literature to tackle this issue is to expand the state to include an affine model that attempts to model a global and local intensity shift for a region with prior modelled appearance [9]. It is unclear however as to whether the optimization process is influenced adversely by the extra degrees of freedom offered by the expanded parameterization. Further, this enforced structure on the nature of the illumination model may not be wholly generalizable to real-world variations. Information theoretic measures of similarity are inherently invariant under smooth and uniquely invertible transformations (homeomorphisms) [10], which is a desirable property

when attempting to establish correspondences across sensing modalities. The use of MI has been particularly successful in medical image registration between intramodal data sources, such as that obtained from MRI and CT scans [11]. In robotics literature MI and its variants have been explored for template tracking and visual servoing [12], [13], 3D object tracking [14], and in robot localization given prior appearance [15], [16]. Contributions: Our objectives in this work are to: • Investigate the formulation of direct VO as the task of maximizing information content between images using Mutual Information (MI) • Evaluate the performance of MI on a range of synthetic and real datasets with challenging illumination conditions • Compare the performances of MI and SSD based rigid body tracking and outline their respective strengths and weaknesses, and • Propose a hybrid tracking methodology to achieve realtime performance. To the best of our knowledge, this is the first work to investigate employing MI for the task for visual odometry without a priori data using commodity hardware. II. T ECHNICAL A PPROACH In this section we first summarize the Lucas Kanade Tracking formulation and then the corresponding substitution of the robust Mutual Information (MI) based cost function in place of the canonical Sum of Squared Differences (SSD) photometric cost function. A. Lucas-Kanade Tracking with SSD The seminal work of Lucas and Kanade (LK) [17] has been applied for various dense correspondence estimation tasks. Given a fixed (reference) template image Ir , a moving (live) image Il , and geometric motion model (warp) w(x; θ) parameterized by the vector θ and pixel coordinates x, the LK algorithm estimates the parameters such that X θ ∗ = argmin kIr (w(x; θ)) − Il (x)k2 (1) θ

x∈Il

Note that the cost function used here is minimizing the photometric error using SSD on the pixel intensities. The optimization problem in Eq. 1 is nonlinear as for nontrivial warps there is no linear relationship between the warp parameters and image intensities. A local minima is usually obtained by utilizing standard nonlinear optimization methods such as Gauss-Newton or Levenberg-Marquardt. A commonly used variant to classical LK is the inverse compositional algorithm (IC) by Baker and Matthews [17], which achieves vast computational savings. We present it in Sec. II-E B. Warp Function Given pixel coordinates x = [x, y]T ∈ R2 with their associated disparities d, the warp function is defined as w(x; θ) = π (T(θ)X(x; d))

where π is the image projection function, T is the mapping from the vector parameterization to the transformation T matrix, and X(x; d) = [X, Y, Z] is the corresponding 3D point in the camera frame. This point can be obtained from a registered depth sensor, triangulated by stereo, or raycast using a registered precomputed mesh. For computational simplicity while obtaining the Hessian, we parameterize the delta transformations by a 6 element vector consisting of the translation and the Euler rotation anT gles [x, y, z, ψ, φ, θ] . This approximation is valid given the assumption that each incremental update considers images that are proximal (a common assumption in visual odometry). C. MI as a Better Cost Function In the traditional LK formulation, raw pixel intensities are used as the sole indicator of alignment accuracy, which has been shown to be a good predicate for tracking of natural images under consistent illumination [18]. However, the slightest deviation from the brightness constancy assumption usually biases the solution, which results in more drift, and more commonly causes unrecoverable tracking failures [19]. In contrast, distributions of intensities provide a more robust measure for comparing images instead of absolute intensities. For instance, sensor saturation adversely impacts the computation of residuals, but leaves the relative distribution of most pixel intensities unchanged. Information-theoretic tracking works by maximizing this similarity between reference and warped image distributions, and Shannon’s MI has been used in literature to compute this measure [20]. Shannon’s MI is the Kullback-Liebler divergence between the joint distribution of two random variables and the product of the two individual distributions. Intuitively, this measure is minimized when the two distributions are independent and therefore contribute no information about each other, and maximized when they are identical. Due to practical considerations we discretize the distributions and express it as   X prl (r, l; θ) (2) MI(Ir , Il ) = prl (r, l; θ) log pr (r; θ)pl (l; θ) r,l

Here the image is arbitrarily discretized into r and l bins. This is stored internally as a histogram, where the joint distribution is computed as prl (r, l; θ) =

1 X β(l − Il (w(x; θ))) · β (r − Ir (x)) Nx x (3)

where β(.) denotes a histogram occupancy function, Nx is the number of samples used to compute the intensity histogram, and Il and Ir are the live and reference images respectively. Note that this formulation implicitly takes into account that images are a joint distribution of intensities and pixel locations; spatial information is encoded by comparing intensities at the same pixel location in the live and reference images.

D. Optimization Algorithm

C. Pixel Subselection

Maximizing the objective in Eq. 2 requires a nonlinear optimization procedure. As demonstrated in [13], [15], a reliable optimization procedure must make use of second-order information. In this work, we employ the full Hessian of the cost function [21] in a second order iterative Newton descent which has been shown to be the theoretically correct choice as opposed to using an approximation of the Hessian [12].

For the MI cost functional, the Jacobian and the Hessian are directly proportional to the magnitude of the image gradient for a particular pixel [24]. Thus, significant computational savings are achieved by only computing the Jacobian and Hessian for pixels that have high gradient magnitudes. We further perform a Non-Max Suppression (NMS) on the pixels to avoid redundancy in the gradient data and only choose the most salient pixels, thereby reducing the Jacobian and Hessian computation to approximately 5% of image pixels with the same performance (Fig. 3) as compared to using all the pixels. For the highest pyramid levels, we simply use all of the pixels to avoid data deprivation.

E. Inverse Compositional Form The optimization objective is of the form ∆θ ∗ = argmax MI(Ir (w(x; ∆θ)), Il (w(x; θk ))

(4)

∆θ

where θk is the estimate of the warp parameters at the k-th iteration. At the next iteration, the parameters are updated using inverse composition w(x, θk+1 ) ← w(x; θk ) ◦ w(x; ∆θ)−1

(5)

This method can be thought of as iteratively updating the delta warp from the reference and composing it with the warp that brings the live image to the reference. Casting it in this form enables computation of the Hessian and the Jacobian with respect to the reference image. Note that for the MI cost function, the Jacobian and Hessian still require terms from the live image, and thus require computation per iteration. However, the expensive second order terms can be precomputed for increased efficiency [22]. III. ROBUST D IRECT V ISUAL O DOMETRY VIA H IERARCHICAL T RACKING We now detail extensions and heuristics required to enable robust real-time performance for the proposed approach. A. Computing the Joint Distribution The majority of MI tracking formulations use normalized joint intensity probability distributions to compute the entropy terms. The choice of how to populate histograms to represent this distribution is extensively discussed in previous work [10], [23]. We employ the common approach of using B-splines in conjunction with Parzen windowing [12], [13]. The primary appeal of this Parzen kernel choice is in its easy to compute analytic derivative and small support. B. Hierarchical Tracking We perform standard pyramidal tracking where Gaussian pyramids are generated from the reference and live images to be able to track larger warps. The intuition behind this is that the higher level pyramids perform a coarse alignment that is progressively refined as the optimization proceeds towards the lower level, higher resolution images. Due to the scarcity of information at the higher level pyramids, the gradient steps can steer the optimization towards incorrect local minima. We perform Armijo line-search iterations to only update the warp parameters when the cost function is appropriately maximized (MI) or minimized (SSD). The initial search step for the linesearch is scaled appropriately with the pyramid level.

D. Hybrid Tracking Since the complexity of computing the joint histogram is O(N 2 ) and the number of pixels operated over increases by a factor of 4 at each level, at the lowest levels the time taken for computing the cost is prohibitively large. This is manifested in the line search iterations in Table I (shown in bold) forming the bulk of the consumed time for MI. Further, as the MI computation effectively blurs the pixel intensities when populating the joint histogram using the Bspline kernel, we observe that the lowermost pyramid levels do not add significantly more refinement to the pose estimate provided by the higher pyramid levels. Therefore, we exploit the robustness of MI at higher pyramid levels and use that to seed the SSD tracking at lower pyramid levels for fast refinement. Algorithm 1 Hybrid MI-SSD Cost Tracking 1: procedure H YBRID T RACKER (Ir , Il , θ) 2: for each level l do 3: H ← ComputeHessian(Ir ) 4: Jiw ← ComputeImageW arpJacobian(Ir ) 5: repeat 6: Iw ← W arpImage(Il , θ) 7: if level < switch then T 8: J ← Jiw (Ir − Iw ) 9: else 10: G ← ComputeM IGradient(Iw , Ir ) 11: α ← ArmijoLineSearch(G, Ir , Il ) 12: J ← αG 13: end if 14: ∆θ ← −H −1 J 15: θ ← θ∆θ −1 16: until ∆Cost < cost or ∆θ < theta 17: end for 18: return θ 19: end procedure IV. E XPERIMENTS The aim of the experiments are to evaluate: (i) the correctness of the MI formulation in idealized environments, and (ii) assess the performance of MI in challenging scenarios. To do so, we compare the performance against SSD tracking [25].

TABLE I T IME BREAKDOWN IN ms PER ITERATION ON A DESKTOP I NTEL I 7 Level 4 3 2 1 0

SSD

MI Line Search

0.5 ± 1.4 0.2 ± 0.2 1.8 ± 5.4 2.2 ± 1.0 1.2 ± 2.0 6.2 ± 2.5 2.4 ± 2.4 22.6 ± 9.0 6.2 ± 2.8 134.3 ± 13.0

Total 13.7 ± 7.6

Total

Hybrid Line Search Total

3.7 ± 4.7 0.2 ± 0.2 3.9 ± 5.8 11.1 ± 2.3 1.8 ± 1.3 9.0 ± 2.8 16.1 ± 1.7 4.8 ± 1.4 12.7 ± 0.6 84.5 ± 6.6 NA 1.8 ± 0.6 169.7 ± 13.9 NA 5.6 ± 1.1

287.3 ± 16.7

35.0 ± 6.6

A. Error Criterion Ei = log(δmocap,i δodom,i )

(6)

i+1 i i where δodom,i = Todom Todom , δmocap,i = Tmocap i+1 Tmocap , and is the inverse composition operator [26]. The rotation and translation error are the L2 norms of the corresponding three elements of the twist vector Ei respectively.

B. Benchmarking Accuracy on a Synthetic Dataset 1) Algorithm performance under ideal conditions: We evaluate the performance of the algorithm using the fluorescent lighting images of the new Tsukuba dataset [27]. As this is a dataset with known ground truth, we plot the distribution of tracking errors, and observe that the majority of the errors are small in magnitude (Fig. 2).

Fig. 3. Comparison of the robustness of tracking to different initial conditions. Uniformly sampled initial starting warps (1000) are generated in 6 Degrees of Freedom of the indicated magnitude. The amount of successfully tracked warps are plotted. Three observations are noted: (1) the convergence basin for the MI algorithm is narrower than the SSD algorithm; (2) increasing number of bins adversely affects convergence for MI; and (3) pixel subselection does not influence algorithm performance.

are successfully tracked by the algorithm. A tracking iteration is considered successful if the norm of the Relative Pose Error (RPE) translation and rotation components between the generated warp and the computed warp is less than 0.01 in magnitude. As reported in prior literature, we observe a decline in the performance of the MI-based cost function as the number of bins used to determine MI is increased (Fig. 3). This degradation occurs as the MI cost function becomes less smooth, causing the Newton descent algorithm to converge to a local minima [13]. 3) Comparison of Optimization Algorithms: Prior literature has also proposed using quasi-Newton optimization for this problem due to issues with the Hessian. In our case, we empirically observe that L-BFGS optimization (using the implementation in the ceres1 library) is sub-optimal both in terms of accuracy and computation time taken. The MI optimization, being a pure Newton descent method, is expected to have a slightly narrower convergence basin than the Gauss Newton optimization used for SSD [17], which is evident in Fig. 3. C. Benchmarking Accuracy on Real Sensor Data

Fig. 2. A still from tracking on the Tsukuba dataset [27]. Clockwise from top left: reference image, stereo depth image, subselected salient features, and the distribution of pose errors for MI tracking versus ground truth.

2) Sensitivity to Initial Conditions: The use of a rigid body warp imposes practical constraints on the photometric quality of the warped images and limits the successful convergence basin for the optimization. To investigate the sensitivity to initial conditions, we generate uniform random twists from a given reference image and attempt to track this synthetically generated image to the reference. Figure 3 compares the percentage of warps at a given twist norm that

1) Comparisons Given Idealized Real-world Environments: Data is collected using an off-the-shelf Asus Xtion Pro Live RGB-D camera with a resolution of 640 × 480 running at 30 Hz. This particular choice of sensor is made so that the depth estimation is not significantly affected by the varying light intensity. We evaluate both the algorithms on a sequence of trajectories as specified in Table II 2) Comparisons under Controlled Variable Lighting Conditions: In order to characterize the performance of the algorithm to track motion across dynamic lighting changes, inspired by [9], we consider the following scenarios: 1 Ceres Solver: A Nonlinear Least Squares Minimizer http://ceres-solver. org/

TABLE II R ELATIVE P OSE E RROR OF A LGORITHMS VERSUS M OTION C APTURE Algorithm MI SSD Hybrid

Translations

Rotations

Nominal

Varying

Nominal

Varying

(7.791 ± 5.232) × 10−3 (9.598 ± 6.254) × 10−3 (7.792 ± 5.220) × 10−3 (9.598 ± 6.253) × 10−3 (7.792 ± 5.220) × 10−3 (9.599 ± 6.253) × 10−3

(1.413 ± 1.072) × 10−2 (1.040 ± 0.661) × 10−2 (1.439 ± 1.314) × 10−2 (1.073 ± 0.825) × 10−2 (1.382 ± 0.977) × 10−2 (1.036 ± 0.608) × 10−2

(6.506 ± 3.566) × 10−3 (2.397 ± 1.459) × 10−2 (6.531 ± 3.545) × 10−3 (2.397 ± 1.457) × 10−2 (6.531 ± 3.543) × 10−3 (2.397 ± 1.458) × 10−2

(1.172 ± 1.287) × 10−2 (3.630 ± 2.606) × 10−2 (1.265 ± 2.929) × 10−2 (3.760 ± 3.007) × 10−2 (1.060 ± 0.776) × 10−2 (3.622 ± 2.558) × 10−2

TABLE III R ELATIVE P OSE E RROR FOR STATIC TESTS Algorithm

Globally Varying Light

Moving Light

MI

(1.4 ± 1.3) × 10−3 (0.48 ± 0.52) × 10−3 (7.3 ± 9.0) × 10−1 (2.6 ± 3.2) × 10−1 (11.7 ± 12.9) × 10−3 (2.34 ± 2.72) × 10−3

(1.3 ± 0.7) × 10−3 (0.8 ± 0.5) × 10−3 (3.14 ± 3.97) × 10−2 (8.0 ± 10.3) × 10−2 (3.24 ± 3.20) × 10−3 (1.62 ± 1.63) × 10−3

SSD Hybrid

m rad m rad m rad

Fig. 4. Sample screenshot on a dataset collected for a figure of eight pattern. Clockwise from left: reference image, registered depth image from the sensor (note the rectification offsets that reduce the effective match region), live image to be tracked, and the residual image after tracking. Note how both the reference image and the live image are blurred in different directions. Conventional sparse feature based tracking is not expected to work in such a situation and this is a major advantage of direct tracking methods.

S1: Global ramp and step change in intensity. The former is seen for instance when moving away from a single point source of light, and the latter during an exposure change due to transitions between dim and well-lit environments S2: Local ramp and step in intensity, such as observed from slow moving light sources, and S3: Motion in oscillating lighting environments. The experimental setup consists of four electronically dimmable high power LED lamps whose brightness can be controlled individually in real time. To measure the luminance, we use an off the shelf sensor directed at the scene. S1: Static Sensor Test with Varying Illumination: We set up the sensor in front of a static indoor scene, and slowly ramp the incident light on the scene from maximum to minimum

m rad m rad m rad

intensity, and then a jump to maximum intensity again. The two extremes are shown in Fig. 1. Instead of performing pure visual odometry between consecutive frames, we compute the transform from the well lit image to the current live image. An ideal tracking algorithm would report the resulting transformations to be at identity, and thus this is a test for both the robustness of the algorithm, and its drift over time. As shown in Table III, the MI tracking drifts only a small distance while maintaining reasonable warps even at the darkest image. We attribute the error primarily to the imprecision of the registration of the depth image to the camera frame on the sensor. SSD tracking, on the other hand, is driven to local minima and loses tracking very early (Fig. 1).

Fig. 5. Plot of the RPE translation error norm over time for the two static tests. (Left) Static test with varying global illumination (Right) Moving light test. Note that SSD fails drastically while MI continues to track.

S2: Static Sensor Test with Moving Light Source For this experiment we move a bright lamp around the scene faced by the sensor and even introduce it into the field of view of the sensor (Fig. 6) S3: Moving Sensor Test with Varying Illumination Here we translate and rotate the sensor along the same trajectories as measured under nominal lighting within a motion capture arena in the presence of oscillating lighting (varying from 7.9 to 32.7 Lux). Results are combined with nominal lighting results in Table II. V. C ONCLUSION AND F UTURE W ORK We propose an optimization based approach for visual odometry based on the principle of maximizing Mutual Information. We evaluate the performance of the proposed tracking algorithm under nominal lighting conditions and illustrate examples of the tracking algorithm working under extreme lighting changes where it significantly outperforms conventional Sum of Squared Differences tracking. A hybrid

Fig. 6. Screenshots for the static sensor test with a moving lightsource (S2). Note the inclusion of a local ramp and a step ramp when the lamp enters the field of view of the sensor. The reference image is only illuminated from the lamp on the left (not pictured).

algorithm is proposed that exploits the robustness against lighting variations at the higher pyramid levels, and utilizes the speed of SSD refinement at the lower levels. In the future we intend to characterize the performance change introduced by this switching algorithm to gain further theoretical insights into the tradeoff between robustness and speed. Although these results serve as a validation of the feasibility of using MI-based tracking for visual odometry, many other possible improvements exist. From a theoretical standpoint, Shannon’s MI is not a true metric, and there are variants of the cost function in the literature that claim to perform better [28], or for instance incorporate spatial information into the MI cost function [29] that could be employed. For high speed tracking on mobile robots, it is important to devise efficient computation strategies to calculate the cost function faster, and further parallelization of the code would be beneficial. Finally, for more robust state estimation, we intend to tightly couple inertial data and incorporate this within a SLAM framweork. ACKNOWLEDGMENT We gratefully acknowledge support from ARL grant W911NF-08-2-0004 and ONR grant N00014-15-1-2929. The authors would also like to thank Hatem Alismail for his valuable insights and discussions. R EFERENCES [1] A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for autonomous driving? The KITTI vision benchmark suite,” in Proc. of IEEE Conference on Computer Vision and Pattern Recognition, June 2012, pp. 3354–3361. [2] J. Sturm, N. Engelhard, F. Endres, W. Burgard, and D. Cremers, “A benchmark for the evaluation of RGB-D SLAM systems,” in Proc. of IEEE International Conference on Intelligent Robots and Systems, Oct 2012, pp. 573–580. [3] M. Irani and P. Anandan, “About Direct Methods,” in Vision Algorithms: Theory and Practice, 2000, pp. 267–277. [4] R. Newcombe, S. Lovegrove, and A. Davison, “DTAM: Dense tracking and mapping in real-time,” in IEEE International Conference on Computer Vision, Nov 2011, pp. 2320–2327.

[5] J. Engel, T. Sch¨ops, and D. Cremers, LSD-SLAM: Large-Scale Direct Monocular SLAM, 2014, pp. 834–849. [6] G. Klein and D. Murray, “Parallel Tracking and Mapping for Small AR Workspaces,” in Proc. of IEEE and ACM International Symposium on Mixed and Augmented Reality, Nov 2007, pp. 225–234. [7] Z. Song and R. Klette, Robustness of Point Feature Detection. Berlin, Heidelberg: Springer Berlin Heidelberg, 2013, pp. 91–99. [8] S. Negahdaripour and C. H. Yu, “A generalized brightness change model for computing optical flow,” in Proc. of International Conference on Computer Vision, May 1993, pp. 2–11. [9] M. Meilland, A. Comport, and P. Rives, “Real-time dense visual tracking under large lighting variations,” in Proc. of the British Machine Vision Conference, 2011, pp. 45.1–45.11, http://dx.doi.org/10.5244/C.25.45. [10] A. Kraskov, H. St¨ogbauer, and P. Grassberger, “Estimating mutual information,” Phys. Rev. E, vol. 69, p. 066138, Jun 2004. [11] J. P. W. Pluim, J. B. A. Maintz, and M. A. Viergever, “Mutualinformation-based registration of medical images: a survey,” IEEE Transactions on Medical Imaging, vol. 22, no. 8, pp. 986–1004, Aug 2003. [12] N. Dowson and R. Bowden, A Unifying Framework for Mutual Information Methods for Use in Non-linear Optimisation, Berlin, Heidelberg, 2006, pp. 365–378. [13] A. Dame and E. Marchand, “Accurate real-time tracking using mutual information,” in Proc. of IEEE and ACM International Symposium on Mixed and Augmented Reality, Oct 2010, pp. 47–56. [14] G. Caron, A. Dame, and E. Marchand, “Direct model based visual tracking and pose estimation using mutual information,” Image and Vision Computing, vol. 32, no. 1, pp. 54 – 63, 2014. [15] A. D. Stewart, “Localisation using the appearance of prior structure,” Ph.D. dissertation, University of Oxford, 2014. [16] G. Pascoe, W. Maddern, and P. Newman, “Robust direct visual localisation using normalised information distance,” in Proc. of British Machine Vision Conference, vol. 3, 2015, p. 4. [17] S. Baker and I. Matthews, “Lucas-kanade 20 years on: A unifying framework,” International journal of computer vision, vol. 56, no. 3, pp. 221–255, 2004. [18] D. L. Ruderman, “The statistics of natural images,” Network: computation in neural systems, vol. 5, no. 4, pp. 517–548, 1994. [19] Y.-H. Kim, A. M. Mart´ınez, and A. C. Kak, “Robust motion estimation under varying illumination,” Image and Vision Computing, vol. 23, no. 4, pp. 365–375, 2005. [20] P. Viola and W. M. Wells III, “Alignment by maximization of mutual information,” International Journal of Computer Vision, vol. 24, no. 2, pp. 137–154, 1997. [21] A. Dame, “A unified direct approach for visual servoing and visual tracking using mutual information,” Theses, Universit´e Rennes 1, Dec. 2010. [Online]. Available: https://tel.archives-ouvertes.fr/tel-00558196 [22] A. Dame and E. Marchand, “Second-order optimization of mutual information for real-time image registration,” Image Processing, IEEE Transactions on, vol. 21, no. 9, pp. 4190–4203, 2012. [23] P. Thevenaz and M. Unser, “Optimization of mutual information for multiresolution image registration,” IEEE Transactions on Image Processing, vol. 9, no. 12, pp. 2083–2099, Dec 2000. [24] N. Dowson and R. Bowden, “Mutual information for lucas-kanade tracking (MILK): An inverse compositional formulation,” IEEE Transactions on Pattern Analysis & Machine Intelligence, no. 1, pp. 180– 185, 2007. [25] H. Alismail and B. Browning, “Direct disparity space: Robust and real-time visual odometry,” 2014. [26] R. Smith, M. Self, and P. Cheeseman, “Estimating uncertain spatial relationships in robotics,” in Autonomous robot vehicles. Springer, 1990, pp. 167–193. [27] S. Martull, M. Peris, and K. Fukui, “Realistic cg stereo image dataset with ground truth disparity maps,” in Proc. of ICPR workshop TrakMark, 2012. [28] N. X. Vinh, J. Epps, and J. Bailey, “Information theoretic measures for clusterings comparison: Variants, properties, normalization and correction for chance,” Journal of Machine Learning Research, vol. 11, no. Oct, pp. 2837–2854, 2010. [29] D. B. Russakoff, C. Tomasi, T. Rohlfing, and C. R. Maurer Jr, “Image similarity using mutual information of regions,” in Computer VisionECCV 2004. Springer, 2004, pp. 596–607.

Robust Direct Visual Odometry using Mutual Information

Differences based Lucas-Kanade tracking formulation. Further, we propose a novel approach that combines the robustness benefits of information-based measures and the speed of tra- ditional intensity based Lucas-Kanade tracking for robust state estimation in real-time. I. MOTIVATION AND RELATED WORK.

5MB Sizes 4 Downloads 205 Views

Recommend Documents

Interactive Visual Object Search through Mutual Information ...
Oct 29, 2010 - Figure 1: Illustration of our discriminative mutual infor- mation score ... query object class. ... one nearest neighbor is considered in both classes.

Visual Servoing from Robust Direct Color Image Registration
as an image registration problem. ... article on direct registration methods of color images and ..... (either in the calibrated domain or in the uncalibrated case).

Visual Servoing from Robust Direct Color Image Registration
article on direct registration methods of color images and their integration in ..... related to surfaces. (either in the calibrated domain or in the uncalibrated case).

Maximum Mutual Information Multi-Phone Units in Direct Modeling
of state transitions given acoustic features, in an HMM that still encodes conventional sequencing constraints from the lexicon and decision tree. In the CRF approach, the total path probabil- ity is factored as the product of state-state transition

On Combining Visual SLAM and Visual Odometry - The University of ...
monocular SLAM system which combines the benefits of these two techniques. ... that end we recast the usual world-centric EKF implementation of visual SLAM ...

Milk: Monocular Visual Odometry with Motion Constraints
Abstract - This video introduces about monocular visual odometry based on planar and circular motion assump- tion. Many indoor and on-road vehicles show planar mo- tion, and each small piece of motion can approximate to circular motion. We presents t

Mutual Information Statistics and Beamforming ... - CiteSeerX
Engineering, Aristotle University of Thessaloniki, 54 124, Thessaloniki,. Greece (e-mail: ...... Bretagne, France and obtained a diploma degree in electrical ...

Reconsidering Mutual Information Based Feature Selection: A ...
Abstract. Mutual information (MI) based approaches are a popu- lar feature selection paradigm. Although the stated goal of MI-based feature selection is to identify a subset of features that share the highest mutual information with the class variabl

Mutual Information Statistics and Beamforming ... - CiteSeerX
G. K. Karagiannidis is with the Department of Electrical and Computer. Engineering, Aristotle ... The most general approach so far has been reported in [16] ...... His research mainly focuses on transmission in multiple antenna systems and includes p

Experian Proprietary Information Agreement (Mutual).pdf ...
1.01.09 3 of 3 ProprietaryInformationAgreement(Mutual). Page 3 of 3. Experian Proprietary Information Agreement (Mutual).pdf. Experian Proprietary Information ...

Weighted Average Pointwise Mutual Information for ... - CiteSeerX
We strip all HTML tags and use only words and numbers as tokens, after converting to .... C.M., Frey, B.J., eds.: AI & Statistics 2003: Proceedings of the Ninth.

Robust Information Extraction with Perceptrons
First, we define a new large-margin. Perceptron algorithm tailored for class- unbalanced data which dynamically ad- justs its margins, according to the gener-.

Robust audio watermarking using perceptual masking - CiteSeerX
Digital watermarking has been proposed as a means to identify the owner or ... frequency bands are replaced with spectral components from a signature.

Robust Information Extraction with Perceptrons
... by “building” is the mention of an entity of type FACILITY and sub- ..... We call this algo- rithm the ..... 24. 90. 116. 5.6. 38.5. 2.4. 53.5. 88.0. 59.1. 70.7. PHYS. 428. 76. 298. 113. 8.7. 69.1. 6.2 .... C-SVC SVM type) takes over 15 hours

Using Visual C ++ - Extras Springer
5. Click on Add. Set the Project Properties. Use the following steps to set a project's properties: 1. In the Solution Explorer tree control, right-click on Example1 and select Properties. 2. In the Property Pages dialog box, select All Configuration

Robust cross-media transfer for visual event detection
ferent types of noisy social multimedia data as input and conduct robust event ... H.3.1 [Information Storage and Retrieval]: Content. Analysis and Indexing.

Fusion with Diffusion for Robust Visual Tracking
A weighted graph is used as an underlying structure of many algorithms like ..... an efficient algorithm with time complexity O(n2), which can be used in real time.

Species independence of mutual information in coding and noncoding ...
5624. ©2000 The American Physical Society .... on Eq. 2, we are able to express, for each single DNA sequence, the maxima and .... Also the variances of log 10. I¯ are almost the same ... This finding leads us to the conclusion that there must.

k-ANMI: A mutual information based clustering ...
Available online 10 July 2006. Abstract ... the class label of each data object to improve the value of object function. That is ... Available online at www.sciencedirect.com ...... high-dimensional data mining, Ph.D. thesis, The University of Texas.

On Hypercontractivity and the Mutual Information ...
sd + (1 − s)(1 − c). If c and d are in (0, 1) the first derivative is −∞ at p(W =1)=0 and it is +∞ at p(W =1)=1. When c or d is in {0, 1} we can use continuity to ...

AMIFS: Adaptive Feature Selection by Using Mutual ...
small as possible, to avoid increasing the computational cost of the learning algorithm as well as the classifier complexity, and in many cases degrading the ...

Robust Visual Tracking with Double Bounding Box Model
As many conventional tracking methods describe the target with a single bounding box, the configuration at time t is typically represented by a three-dimensional.

G-ANMI: A mutual information based genetic clustering ...
better clustering results than the algorithms in [4,2]. Meanwhile,. G-ANMI has the .... http://www.cs.umb.edu/~dana/GAClust/index.html. 2 We use a data set that is ...