IEICE TRANS. INF. & SYST., VOL.E95–D, NO.11 NOVEMBER 2012

2693

PAPER

Enhancing Memory-Based Particle Filter with Detection-Based Memory Acquisition for Robustness under Severe Occlusion Dan MIKAMI†a) , Kazuhiro OTSUKA† , Shiro KUMANO† , and Junji YAMATO† , Members

SUMMARY A novel enhancement for the memory-based particle filter is proposed for visual pose tracking under severe occlusions. The enhancement is the addition of a detection-based memory acquisition mechanism. The memory-based particle filter, called M-PF, is a particle filter that predicts prior distributions from past history of target state stored in memory. It can achieve high robustness against abrupt changes in movement direction and quick recovery from target loss due to occlusions. Such high performance requires sufficient past history stored in the memory. Conventionally, M-PF conducts online memory acquisition which assumes simple target dynamics without occlusions for guaranteeing high-quality histories of the target track. The requirement of memory acquisition narrows the coverage of M-PF in practice. In this paper, we propose a new memory acquisition mechanism for M-PF that well supports application in practical conditions including complex dynamics and severe occlusions. The key idea is to use a target detector that can produce additional prior distribution of the target state. We call it M-PFDMA for M-PF with detection-based memory acquisition. The detection-based prior distribution well predicts possible target position/pose even in limited-visibility conditions caused by occlusions. Such better prior distributions contribute to stable estimation of target state, which is then added to memorized data. As a result, MPFDMA can start with no memory entries but soon achieve stable tracking even in severe conditions. Experiments confirm M-PFDMA’s good performance in such conditions. key words: M-PF, visual object tracking, severe occlusion, detection

1.

Introduction

Visual object tracking has been acknowledged as one of the most important techniques in computer vision [1], and is required for a wide range of applications such as automatic surveillance, man-machine interfaces [2], [3], and communication scene analysis [4]. For visual object tracking, Bayesian filter-based trackers have been acknowledged as a promising approach; they represent a unified probabilistic framework for sequentially estimating the target state from an observed data stream [5]. At each time step, the Bayesian filter computes the posterior distribution of the target state by using observation likelihood and the prior distribution. One implementation, the particle filter [6], has been widely used for target tracking. It represents probability distributions of the target state by a set of samples, called particles. A particle filter (PF) can potentially handle non-Gaussian, nonlinear dynamics/observation processes; this contributes to robust tracking. However, most particle filter-based visual trackers are Manuscript received October 28, 2011. Manuscript revised April 27, 2012. † The authors are with NTT Communication Science Laboratories, NTT Corporation, Atsugi-shi, 243–0198 Japan. a) E-mail: [email protected] DOI: 10.1587/transinf.E95.D.2693

rather constrained since they employ linear, Gaussian, and time invariant dynamics for simplicity. We focused on the issue that the simplicity degrades PF’s robustness in real-world situations including abrupt movements and occlusions. To deal with the target’s nonMarkov, non-Gaussian, and time-varying dynamics, We proposed a memory-based particle filter, called M-PF, as an extension of the particle filter [7]. M-PF eases the Markov assumption of PF and uses past history of the target’s states to predict the prior distribution on the basis of the target’s long-term dynamics. M-PF offers robustness against abrupt object movements and quick recovery from tracking failure. However, such high performance can be achieved only if target history in memory is voluminous and high quality. The first implementation of M-PF in [7] includes an online memory acquisition period which requires the capture of simple dynamics without occlusions for assuring stable tracking. Demanding memory acquisition in this manner narrows the coverage of M-PF. To acquire memory more stably in a wider range of real-world situations, this paper proposes a new memoryacquisition mechanism for the M-PF-based tracker. The key idea is combining M-PF with a target detector. The target detector can find the target position/pose even in cluttered conditions, and the detection result is used for creating an additional prior distribution of target position/pose. This detection-based prior distribution and the original memorybased prior distribution are combined and provided to the posterior distribution estimation step. Such combined prior distribution prediction contributes to more stable estimation of the target state even in limited visibility conditions. This estimated target state is then added to memory and is used for creating the memory-based prior distribution in future steps. This cycle of detection, combined prior distribution prediction, posterior distribution estimation, and memory accumulation is highly synergistic in terms of boosting MPF performance in real-world environments. We name it M-PFDMA for M-PF with detection-based memory acquisition. M-PFDMA has the following advantages; stable initial tracking without memory, quick recovery after occlusion, and wider pose range recoverability. We have proposed M-PF with appearance prediction (M-PFAP) [8] as another enhancement of M-PF. M-PFAP enhances robustness against appearance changes due to pose changes. However, it also fails to obtain history under severe occlusions. M-PFDMA offers the memory acquisition ability, and it can be applied to M-PFAP in the same manner.

c 2012 The Institute of Electronics, Information and Communication Engineers Copyright 

IEICE TRANS. INF. & SYST., VOL.E95–D, NO.11 NOVEMBER 2012

2694

To verify the effectiveness of M-PFDMA, we implement a facial pose tracker. Facial pose tracking should yield attributes of the face such as position and rotation. As the object detector, we use the “joint probabilistic increment sign correlation face detector,” in short JPrISC face detector, the multi-view face detector proposed by Tian et al. [9]. The JPrISC face detector can detect faces from frontal view to near profile view, and can output ten face pose classes. Facial pose tracking experiments verify that M-PFDMA can acquire target’s state history even under severe occlusion and achieves accurate tracking and high recoverability. In the context of human pose tracking, the idea of tracking-by-detection has been gaining attention in recent years as a possible alternative to the traditional target tracking approach [10]–[12]; it reflects the rapid progress in object detectors. However, the current human facial pose detectors are inadequate for realizing mature tracking-bydetection since they are not fast enough for real-time tracking, not accurate enough for determining precise target position/pose, and not able to well handle target dynamics. Rather than relying on just the detector, combining the detector with a tracker has been seen as a reasonable solution. One example was proposed by [13]–[15]. They combined a PF-based tracker with a face detector for observation [13] and prior distribution prediction [14], [15]. In [13], a multipose class face detector provides multiple choices with regard to the observation function, but the tracking pose resolution is limited to detectable pose classes. In [14], a head position detector yields a uniform prior distribution over all pose ranges, which limits pose accuracy in tracking. Their usage of detectors improved PF-based tracking, but they targeted only a simple environment with no occlusions. In [16], the environmental existence map (EEM) obtained from observation of human actions for a long period of time is used for stabilizing human tracking. It challenged stable tracking by employing past trajectory of human motions. It only estimates three-dimensional positions and assumes that a lost target can be rediscovered within the basic PF framework, though it requires rather long time steps. In the context of high-dimensional state estimation, a target once lost due to occlusions is difficult to be rediscovered in the basic PF framework. Unlike the above previous works, our goal is the estimation of high-dimensional states, i.e., position and rotation, with high recoverability from severe occlusions. For that purpose, the M-PFDMA integrates the object detector into the M-PF framework. The most important contribution of the M-PFDMA is the use of M-PF. Unlike Okuma et al. [15], who reported a combination of PF and a detector, M-PFDMA uses detection results not only for prior distribution prediction: It stores the detection results in history and reuses them for prior distribution prediction at latter time steps, regardless of the detection result at the time. Generally speaking, a detector requires much more processing time than a tracker. The detection delay deteriorates the detection-based prior distribution. In addition, motion blur makes target detection difficult. By storing and

reusing detection-based tracking results, M-PF framework within M-PFDMA overcomes the above detector’s shortcomings. The synergetic effects of M-PFDMA separate it from previous combination methods. This paper is based on the conference proceeding [17] and adds new analyses and discussions of determining mixing weight in Sect. 5. The remainder of this paper is organized as follows; Sect. 2 briefly reviews memory-based particle filter and its recoverability from tracking failure. Section 3 proposes our new facial pose tracker. Section 4, details the experimental environment and results. Section 5 discusses mixing weight of memory-based and detectionbased prior distributions for further development. Finally, Sect. 6 concludes and discusses our proposal. 2.

Memory-Based Particle Filter and Its Tracking Recoverability

This section overviews the memory-based particle filter and addresses the recovery problem from tracking failure, which has, up to now, been an unsolved problem in the field of tracking. 2.1

Memory-Based Particle Filter

M-PF [7] realizes robust target tracking without explicit modeling of the target’s dynamics even when the target moves quickly. Figure 1 outlines M-PF. M-PF keeps the T } 1:T = { x1 , · · · x temporal sequence of past state estimates x 1:T denotes a sequence of state estimates in memory. Here, x t denotes a pose estimate at time from time 1 to time T , and x t. M-PF assumes that the subsequent parts of past similar states provide good estimates of the current future. M-PF introduced Temporal Recurrent Probability (TRP), which is a probability distribution defined in the temporal domain that indicates the possibility that a past state will reappear in the future. To predict the prior distribution, M-PF starts with TRP modeling. It then conducts temporal sampling on the basis of TRP. The sampled histories are denoted by blue dots in Fig. 1. It retrieves the corresponding past state estimates for each sampled time step, denoted by pink dots in Fig. 1. After that, considering the uncertainty in the state estimates, each referred past state is convoluted with kernel distributions (light green dist. in Fig. 1), and they are mixed together to generate the prior distribution (green dist. in Fig. 1). Finally, a set of particles is generated according to the prior distribution (blue dots in right part of Fig. 1). The M-PF-based face pose tracker in [7] estimates the position and rotation at each time step. M-PF uses the same observation process as traditional PF, which uses a single template built at initialization. This yields the 50 degree face rotation limit noted in [7]. 2.2

Recovery from Tracking Failure

The conventional visual trackers track a target with the assumptions of simple dynamics and excellent visibility. Se-

MIKAMI et al.: MEMORY-BASED PARTICLE FILTER WITH DETECTION-BASED MEMORY ACQUISITION

2695

Fig. 1 M-PF employs past state sequences to predict a future state. First, it calculates the reoccurrence possibility of past state estimates (TRP). Past time steps are then sampled on the basis of TRP. Past state estimates corresponding to the sampled time steps are combined to predict prior distribution. M-PF enables the implicit modeling of complex dynamics.

Fig. 2 Properties of facial pose tracker with regard to recovery speed and recoverable pose range; PF is only able to rediscover the target if it takes a pose similar to the pose prior to tracking loss. By integrating a detector, PF+Detector enables recovery if the target can be detected. M-PF enables recovery if the target takes a stored pose. M-PF-based recovery can find the target faster than the detector. M-PFDMA supports memory-based quick recovery and detection-based wide pose range recovery.

vere occlusions remained a challenging problem, i.e. how can a tracker rediscover the lost target under severe occlusion. This occlusion recovery problem can be viewed from two aspects; quickness of recovery and recoverable pose range. The quickness of recovery indicates how rapidly a tracker can rediscover a target after the target reappears after being lost due to occlusion. The recoverable pose range indicates the pose range within which the tracker can rediscover the face, we must expect the target pose to change significant during an occlusion. Conventional methods can be mapped as in Fig. 2 according to these aspects. The conventional PF-based tracker tries to rediscover a lost target by simply broadening prior distribution frame by frame according to random walk dynamics. It may be able to rediscover a lost target if the occlusion period is short and pose changes is small. However, as time passes and a pose

changes significantly, rediscovery probability falls dramatically. Original M-PF (with simple online memory acquisition) can rediscover the lost target if the target takes a pose that is stored in memory by memory-based prior distribution prediction. Such memory-based rediscovery is faster than detection-based rediscovery or PF-based rediscovery. This is because memory-based prior distribution can well predict possible poses/positions of the target after occlusion. Therefore, M-PF provides more rapid recovery and wider recoverable pose range. Combinations of particle filtering and detector, denoted by PF+Detector in Fig. 2, are able to rediscover the lost target if the target takes detectable poses. Though the detectable poses and the required time for detection depend on the detector’s performance, generally speaking, detectors have much higher computational costs than trackers. To detect targets that take a greater variety of poses, computational costs become even higher. This also means that the recovery speed tends to be rather slow. These trends are well confirmed by our experiments (see Figs. 7, 8 and 9). M-PFDMA aims at achieving faster recovery in wide pose ranges by integrating detection-based memory acquisition into the memory-based particle filter. Detection-based prior distribution helps rediscovery of targets that take previously unobserved poses. Rediscovered and tracked positions/poses are stored in history. Acquired history can be used to improve memory-based prior distribution prediction. This synergetic combination of detection-based memory acquisition and memory-based prior distribution prediction enables faster recovery in wide pose ranges. 3.

Memory-Based Particle Filter with Detection-Based Memory Acquisition: M-PFDMA

This section proposes an enhancement of M-PF for object tracking called M-PFDMA. It stands for M-PF with detection-based memory acquisition. It achieves quick recovery in wide pose range due to its synergistic combination of an object detector and tracker. As reviewed in Sect. 2.1, the basic assumption of MPF was that the target repeats similar movements again and again. On the basis of this assumption, M-PF introduced the temporal recurrent probability (TRP), which indicated the tendency of past similar states reappearing in the future. MPF replaced the simple dynamics model employed in conventional particle filters, such as the random walk model, by the temporal recurrent probability, for predicting prior distribution, called memory-based prior distribution. M-PFDMA yields detection-based prior distribution in addition to the memory-based prior distribution. The detection-based prior distribution is folded into memory-based prior distribution. This integration of an effective detector into a unified M-PF framework is the key contribution of M-PFDMA. The remainder of this section first overviews MPFDMA, and then, describes the prior distribution prediction formulation. Finally, we describe its implementation in

IEICE TRANS. INF. & SYST., VOL.E95–D, NO.11 NOVEMBER 2012

2696

tory is important for M-PF (DMA). Accumulation of history directly from a detector may deteriorate the quality of history. On the contrary, if the detection result is used for prior distribution prediction, subsequent process, i.e. likelihood calculation by sparse template matching, can eliminate the erroneous detection. From above two reasons, M-PFDMA framework employs a detector for improving prior distribution. 3.2 Fig. 3 Block diagram of our facial pose tracker. It has four main components; initialization, prior distribution prediction, posterior distribution prediction, and tracking result estimation. The key differences of M-PFDMA from M-PF are hatched, i.e. integration of detection-based prior distribution prediction into memory-based prior distribution prediction. It enables rediscovery of a lost target even if the target takes a position/pose which has not been stored in memory while occlusions. As the more position/pose are added into the memory, the better the memory-based prior distribution prediction becomes. The synergetic effect between detection and tracking is the key contribution of the proposed M-PFDMA.

a facial pose tracker. 3.1 System Overview M-PFDMA integrates an object detector into the M-PF framework. Figure 3 illustrates the block diagram of MPFDMA; the differences from M-PF are hatched. MPFDMA has four main components; initialization, prior distribution prediction, posterior distribution prediction, and tracking result estimation. In the initialization step, the tracker detects a target and makes a target model. In the prior distribution calculation step, it calculates prior distribution on the basis of two clues. One is memorybased prior distribution prediction, which is described in Sect. 2.1. The other is detection-based prior distribution estimation; detection-based probability distribution is generated on the basis of the detection results. The memory-based and detection-based prior distributions are then combined, which is described in Sect. 3.2 in more detail. The posterior distribution calculation step calculates observation likelihood, and then, by using likelihood and prior distribution, calculates posterior distribution. Finally, a pose estimate is obtained by weighted averaging and stored in memory. Here, the stability is automatically judged by thresholding of the maximum likelihood among particles. Only when a tracking is stable, the estimate is stored. The steps from determining the prior distribution to pose estimation are repeated in each frame. Within the framework of a combination of a detector and M-PF, using a detector to obtain the position and pose of a target directly might be considered. However, two drawbacks exist in the solution; first one is the low resolution of pose estimation. On the basis of [7], M-PF can estimate poses of a face with errors of about 3.0%. On the contrary, the resolution of pose estimation by JPrISC is only ten classes. Second one is the error detection. Quality of his-

Formulation of Prior Distribution Prediction of MPFDMA

M-PFDMA’s key extension from M-PF resides in its prior distribution prediction parts. Unlike M-PF, M-PFDMA employs an object detector and combines memory-based prior distribution and detection-based prior distribution. This section formulates the prior distribution of M-PFDMA. Bayesian filters, including particle filter, calculate prior distribution by multiplying the observation likelihood by the motion dynamics of target state as in  p(xt+1 |x1:t ) · p(x1:t |Z1:t )dx1:t , (1) p(xt+1 |Z1:t ) = where xt denotes the state vector indicating position and rotation, x1:t = {x1 , · · · , xt } denotes the state sequence of state vector from time 1 to t, and Z1:t = {Z1 , · · · , Zt } denotes the sequence of observations from time 1 to t. As the dynamics model, conventional particle filters assume a short term Markov model as in p(xt+1 |x1:t ) ≈ p(xt+1 |xt ).

(2)

The memory-based particle filter assumes the tendency of repeating past positions/poses, and introduced the temporal recurrent probability Φ(·). It replaced the dynamics model with the temporal recurrent probability given by p(xt+Δt |x1:t , Δt) =

t 

Φ(t| x1:t , Δτ) · K(xt+Δτ | xτ ),

(3)

τ=1

where Δt denotes the time offset between current time and  denotes a point estimate stored in prediction target and x memory. K() is the kernel distribution that represents uncertainty in the stored state estimate. M-PFDMA replaces the memory-based prior distribution by a combination of memory-based prior distribution and detection-based prior distribution calculated on the ba j (zt ) as sis of object detection results x p(xt+Δt |x1:t , Δt, zt+Δt ) t  Φ(t| x1:t , Δτ) · K(xt+Δτ | xτ ) = (1 − α) τ=1



Nd 

q(xt | x j (zt ), Σ),

(4)

j=1

where α denotes the mixing weight between memory-based

MIKAMI et al.: MEMORY-BASED PARTICLE FILTER WITH DETECTION-BASED MEMORY ACQUISITION

2697

prior distribution (first part of (4)) and detection-based prior distribution q() (latter part of (4)), and Nd denotes the number of detected objects. The detection result  x j (zt ) includes estimated position and pose, and q(x| x j (zt ), Σ) de j (zt ) and variancenotes Gaussian distributions with mean x covariance matrix Σ for position/pose. In this paper, a static predefined value, which is independent of the number of detected objects, is employed for the mixing weight α. The combined prior distribution realizes high recoverability and accuracy, which yields quick and stable memory acquisition. 3.3 Implementation in a Facial Pose Tracker We implement a facial pose tracker on the basis of MPFDMA. For detection, we used the multi-view face detector proposed by Tian et al. [9], which is called JPrISC† . Note that, we selected JPrISC as an example of implementation; other trackers, e.g., [19], can be applied as the detector for M-PFDMA. The JPrISC detector has certain advantages as a detector for M-PFDMA. One is its pose estimation capability in face detection scenarios, which is reported in [9]. The pose estimation greatly enhances the effectiveness of prior distribution. The other is scalability. The JPrISC is a trainingbased detector; it does not specialize in face detection and can be applied for the detection of various objects. 4.

Experiment

To confirm M-PFDMA’s performance, we focus on severe occlusion cases because stable tracking is already possible with conventional methods if the target is not occluded, even if it exhibits complex dynamics. 4.1 Experimental Settings Video capturing environment is as follows. We used PointGreyResearch’s FLEA, a digital color camera, to capture 1024 × 768 (pixels) images at 30 frames per second. Note that the tracking processes uses only grayscale images converted from the color images. The CPU of the PC used was an Intel Core2Extreme 3.0 GHz (Quad Core) and the GPU was NVIDIA GeForce GTX480. For all experiments in this section, 2000 particles were used except for creating ground truth data in Sect. 4.3 (quantitative evaluation). Our tracker was implemented on the basis of STCTracker [18], which accelerates particle filtering by GPU implementation. It can run at 30 fps, and our M-PFDMAbased tracker also can run at 30 fps. 4.2 Typical Example of Proposed Facial Pose Tracker in Action We compare M-PFDMA to two other methods. One is the memory-based particle filter (M-PF) [7], and the other is a combination of particle filtering and face detector called the

JPrISC face detector (PF+Detector). Note that the JPrISC face detector can output ten pose classes in addition to position. Therefore, PF+Detector is expected to show higher performance than similar method in [14] especially in terms of speed of recovery. We prepared two videos with severe occlusions. In the first video, objects horizontally and vertically cross at the camera’s centerline as in Fig. 10, and cause occlusions. Tracking starts in the top-right area, and the subject moves up-down and left-right while changing his pose. The second video simulates video conference situation; the subject makes a presentation in front of a camera, interacting with participants on the other side. Figure 4 shows snapshots of the tracking behavior of M-PF, PF+Detector, and M-PFDMA, from the first video. Only upper half images are shown. In Fig. 4, left column, middle column, and right column show the result of M-PF, PF+Detector, and M-PFDMA, respectively. In each column, figures are listed in time order. From first row to second row, the target face moved from right to left; and from second row to third row, it moved from left to right. Second row shows that M-PFDMA and PF+Detector successfully detected the target face while M-PF didn’t, because the target’s pose had not been stored. It is one example of improvement in recoverable pose range. This recovery confirms the effectiveness of the detector. However, these recoveries were achieved by the detector, so they took rather a long time. Third row shows that M-PF and M-PFDMA found the lost target, while PF+Detector took much longer to rediscover it. In this situation, past state history of right side had already been stored, so the memory-based methods quickly rediscovered the lost target. Figure 5 shows snapshots of the tracking behavior of M-PFDMA from the second video. The snapshots are listed in time order. The second video includes self-occlusion, e.g., turning back while moving left to right (Fig. 5 (b)-(f)) and right to left (Fig. 5 (h)-(i)). It also includes scale changes (Fig. 5 (f)-(g)) and non-rigid deformations (Fig. 5 (h)), e.g., facial expression changes. During natural behaviors as depicted in Fig. 5, our M-PFDMA successfully estimates the position and rotation of the target while the target is not occluded. The above comparisons verified that M-PFDMA supports both memory-based quick recovery and detectionbased recovery and so covers the cases in which the target deviates from stored poses. 4.3

Quantitative Evaluations

This section shows quantitative evaluations of tracking performance. For the quantitative evaluations, we prepared a new video. Originally, the new video did not include occlusions so that the M-PF with 15000 particles well estimates the positions and poses of the target face, and we used the † Though we used the JPrISC as the face pose detector, it is not specialized in face pose. By collecting training images, it can be applied for detecting other objects.

IEICE TRANS. INF. & SYST., VOL.E95–D, NO.11 NOVEMBER 2012

2698

Fig. 4 Tracking results for M-PF, PF+Detector, and M-PFDMA (proposed). White mesh denotes estimated position/pose. While tracking is unstable, the mesh turns gray (almost invisible). The left column shows M-PF’s output, the middle column shows PF+Detector’s output, and the right column shows M-PFDMA’s output. In each column, figures are listed in time order.

Fig. 5 Tracking results for M-PFDMA (proposed). White mesh denotes estimated position/pose. While tracking is unstable, the mesh turns gray to become inconspicuous. (a) Initialization. (b)-(f) The subject moves from right to left while rotating; (f)-(h) the subject gets close to and backs away from it; (h)-(i) the subject moves from left to right while rotating. During such natural behavior of the subject, the proposed M-PFDMA occasionally missed the target. However, it rediscovered the target soon after occlusions.

tracking result as the ground truth. After that, artificial occlusions were overlaid at the camera’s center lines as shown in Fig. 6 by image processing. This video was used as the tracking target. We compared the proposed M-PFDMA to M-PF and to PF+Detector. First, the tracking success ratio was examined. The

tracking success ratio is the ratio of frames in which the tracker estimated poses and positions correctly. The correctness is confirmed by absolute differences between positions and poses in the ground truth data and in estimation results; for calculating the tracking success ratio, we define a result whose estimation error is within 20 pixels in pose and 10

MIKAMI et al.: MEMORY-BASED PARTICLE FILTER WITH DETECTION-BASED MEMORY ACQUISITION

2699

degrees in pose is success. Though this video includes numerous occlusions, the tracking success ratio was calculated regardless of whether the target was occluded or not. Table 1 shows the result. M-PFDMA showed the best tracking success ratio. Figures 7 (a) and 7 (b) show the estimated horizontal position and pitch angle, respectively. The red line shows the ground truth. The other lines, i.e., the estimation results for M-PFDMA, of M-PF, and PF+Detector, are discontinuous. Breaks indicate that the tracking results in the corresponding frames are unstable.

Fig. 6 Evaluation target; white occluders are overlaid at the cameras’ center lines. A subject moves right-left and up-down across the white occluderes. Table 1 Method M-PFDMA M-PF PF+Detector

Tracking success ratio. Tracking Success Ratio [%] 67.4% 36.8% 44.6%

From frame 647 to 655, the tracking target was occluded. Therefore, tracking results for all methods became unstable. At frame 655, the occlusion cleared. Soon after the occlusion cleared, M-PFDMA rediscovered the target and tracking was restarted. On the contrary, PF+Detector did not restart tracking until frame 670. The restart of tracking in M-PFDMA was achieved by memory-based prediction using history. Note that the history was obtained because the detector successfully found the target in a previous time step. M-PF failed to restart tracking around frame 670 because it does not have detector. After that, the target was again occluded and the occlusion cleared at frame 705. At that time, M-PF and MPFDMA quickly restarted the tracking. In this case, both M-PF and M-PFDMA had accumulated the proper history of past states and memory-based prior distribution prediction yielded the proper prior distribution. PF+Detector took much time to restart tracking because its recovery is depending on the detection and detection-based prior distribution prediction. In addition, we employed a well-known commercial facial pose tracker, called FaceAPI [20]. It can detect a face in near frontal view, and then can sequentially estimate position and pose of the face. FaceAPI obtains a position of a face in the camera coordinate (x, y, z), whereas the MPFDMA obtains it in the pixel coordinate (u, v) and scale coefficient. These conversions are difficult and include ambiguity; therefore, we judged correctness of estimation by manual observation. The target videos were the same as in

(a) Horizontal position in video window

(b) Pitch angle Fig. 7 M-PFDMA rediscovered the lost target soon after occlusion cleared; M-PF failed to rediscover the target until stored position/pose appeared at frame 705; PF+Detector required much time than MPFDMA.

IEICE TRANS. INF. & SYST., VOL.E95–D, NO.11 NOVEMBER 2012

2700 Table 2 Comparisons of tracking successful ratio among M-PF, PF+Detector, FaceAPI, and M-PFDMA (proposed). When we calculated the tracking successful ratio, the frames which were not able to track due to occlusion were manually excluded. Method M-PF PF+Detector FaceAPI M-PFDMA (proposed)

successful tracking ratio video 1 video 2 37.5% 36.9% 52.4% 62.2% 64.2% 53.7% 81.5% 90.1% Fig. 9 Tracking results for M-PF, PF+Detector, FaceAPI, and MPFDMA are shown. M-PFDMA achieved faster recovery of lost target than the others. Compared with Fig. 8, a similar property is observed.

(a) M-PF

(b) M-PFDMA (proposed)

Fig. 10 Stored history by each method; (a) M-PF and (b) M-PFDMA (proposed method). Positions of stored history are shown by yellow lines. M-PF obtained history only in right-top area, where the target face was initialized. Contrary, M-PFDMA obtained history in wider area.

Fig. 8 Tracking results for M-PF, PF+Detector, FaceAPI, and MPFDMA are shown. M-PFDMA achieved faster recovery of lost target than the others. In frame 225, PF+Detector and M-PFDMA successfully estimated position/pose of the target, whereas M-PF didn’t because the history wasn’t stored close to the position/pose (left column). In frame 400, MPF and M-PFDMA successfully estimated the position/pose of the target, but PF+Detector wasn’t because detection requires large computation time (middle column). In frame 655, only M-PFDMA successfully estimated the position/pose of the target (right column).

Sect. 4.2. Tracking success ratio and the graphs of tracking results are follows. Tracking success ratios, shown in Table 2, confirm that M-PFDMA successfully estimated the target’s position and rotation in 81.5% / 90.1% (video 1/video 2) of frames, whereas M-PF, PF+Detector, and FaceAPI achieved only 37.5% / 36.9%, 52.4% / 62.2%, and 64.2% / 53.7%, respectively. Figure 8 shows a part of the tracking results for video 1 in more detail. In Fig. 8, the horizontal axis denotes frame number. Each line denotes a tracking result. The discontinuities indicate the frames in which target position/pose were not correctly estimated. Figure 9 shows tracking results for video 2. As shown in Figs. 8 and 9, M-PFDMA combined the frames correctly estimated by PF+Detector and M-PF while avoiding their deficiencies. A qualitative evaluation of PF-based trackers, i.e., PF+Detector, M-PF, and M-PFDMA, is described in

Sect. 4.2. Therefore, we briefly add qualitative comparisons with respect to FaceAPI. Note that because FaceAPI is a commercial facial pose tracker, the detailed algorithm is not apparent. FaceAPI is able to detect a face in near frontal view, so it successfully rediscovered the tracking target after occlusions if the target was near frontal view and was observed without blur. While a target is moving, the observed image becomes blurred. Therefore, FaceAPI failed to rediscover the lost target while it was moving even if it had not been occluded. On the contrary, M-PFDMA quickly rediscovered the lost target under blurred observation. For likelihood calculation, M-PFDMA employs sparse template matching. The sparse template matching has been verified to have robustness under small changes in appearance such as those caused by motion blur [7]. Therefore, M-PFDMA can rediscover the target if a good prior distribution is predicted, even in a blurred image. Thus, M-PFDMA outperformed other trackers including PF-based trackers and a commercial non-PF-based tracker. Next, the memory acquisition results of the previous M-PF and the proposed M-PFDMA from video 1 are shown in Fig. 10. Yellow lines denote the position of the history stored in memory. So, target’s movements along the yellow lines are supposed to be tracked stably on the basis of memory-based prior distribution prediction. The left figure shows the output of M-PF, and the right figure shows that of M-PFDMA. As shown in Fig. 10, M-PF stored history covered only the top-right area, where tracking started. This means that the tracker failed to rediscover target when the target moved to other areas after occlusions due to large

MIKAMI et al.: MEMORY-BASED PARTICLE FILTER WITH DETECTION-BASED MEMORY ACQUISITION

2701 Table 3

Successful tracking ratio (200 particles case).

M-PFDMA (static) M-PFDMA (dynamic)

Video 1 49.8% 60.7%

Video 2 76.9% 85.5%

changes in position/pose while occlusions. On the contrary, the stored memory of M-PFDMA covered the entire field of view. The numbers of stored memories of this sequence by M-PF and M-PFAP are 455 and 808, respectively. The memory acquisition performance of M-PFDMA under severe occlusion was confirmed. 5.

Discussion

In the above experiments, mixing weight α = 0.15 was employed. However, appropriate α may vary according to tracking conditions. For example, when tracking is stable, memory-based prior distribution well predicts prior distribution without using detection-based prior distribution. In addition, deploying more particles for memory-based prior distribution improves tracking stability. On the contrary, deploying more particles for the detection-based prior distribution may enhance recoverability of target while the current positions/poses weren’t stored in memory. However, deploying too many particles for detection-based prior distribution may deteriorate the tracking accuracy, especially when a target does not have repetitive movement. In this section, the mixing weight α is analyzed for further improvement. To verify the effectiveness of dynamic updating of α, we empirically employed ⎧ ⎪ 0 if d < 3 ⎪ ⎪ ⎪ ⎨ α=⎪ (5) 0.4 · (d − 3)/27 3 ≤ d ≤ 30 ⎪ ⎪ ⎪ ⎩0.4 otherwise, for updating α, where d denotes the number of frames whose tracking status is unstable; d becomes zero when the current tracking is stable. Hereafter, M-PFDMA that employs static α is denoted as M-PFDMA (static) and M-PFDMA that employs dynamic α updated by Eq. (5) is denoted as M-PFDMA (dynamic). First, successful tracking ratios of Video 1 and Video 2 were examined for M-PFDMA (static) and M-PFDMA (dynamic). There was no significant difference between them. The number of particles, i.e., 2000, was large enough for assigning 15% of them for detection-based prior distribution. This did not deteriorate the tracking-successful ratio. Therefore, assuming a condition that is severe in terms of computational cost, we conducted the same experiment with 200 particles. Table 3 shows the results, which indicate that dynamic update of α improved the tracking-successful ratio. To examine the effect of the dynamic update of α in more detail, we carried out the following two experiments. 5.1 Robustness under Abrupt Movement To examine the robustness under abrupt movements, we pre-

Fig. 11 Tracking results of horizontal position; tracking result of MPFDMA (static) becomes unstable when the target’s movement becomes faster.

Fig. 12

Elapsed time for restarting stable tracking.

pared a new video in which a subject repeatedly moved left and right. During the video the subject moved faster as the time passed. While the tracking status is stable, MPFDMA (dynamic) generates prior distribution based more on memory-based prior distribution. Therefore, it is expected that the M-PFDMA (dynamic) would realize more robust tracking under abrupt movement than M-PFDMA (static) does. The results are shown in Fig. 11, where the horizontal axis denotes time and the vertical axis denotes the horizontal position of the target. In addition to M-PFDMA (dynamic) and M-PFDMA (static), the tracking result for M-PF† and ground truth data are shown. Note that, the ground truth data is obtained by the stable tracking with 15000 particles. Whereas M-PFDMA (static) includes large estimation error after frame 325, M-PFDMA (dynamic) estimated as accurately as M-PF. Figure 11 suggests that M-PFDMA (dynamic) is robust against abrupt movements. 5.2

Recoverability

The next experiment examined recoverability. In this experiment, an artificial occlusion is overlaid by image processing and generated tracking unstable situation. We examined the elapsed time for restarting stable tracking. Various conditions, i.e., occlusion durations (5, 15, 30, and 100 frames) and poses when the occluder disappeared (0, 30, and 60 degrees in horizontal rotation), were examined. The results are shown in Fig. 12. In Fig. 12, results of 5, 15, 30, and 100 † This video does not include occlusion, so M-PF is able to track the target in this video.

IEICE TRANS. INF. & SYST., VOL.E95–D, NO.11 NOVEMBER 2012

2702

frames occlusion are listed from left to right. Each occlusion duration has three bars; each showing horizontal rotation when the occluder cleared; 0, 30, and 60 degrees. Additionally, averages for the horizontal rotations and for the four occlusion durations are shown. In the same conditions, M-PFDMA (dynamic) totally required a shorter time for stabilizing tracking. From the averages, more time was required for stabilizing tracking as the pose became large for both M-PFDMA (dynamic) and M-PFDMA (static). This is because the likelihood calculation becomes unstable as the pose becomes large. With increasing occlusion duration, the required time for stabilizing tracking in M-PFDMA (static) increased, but that in M-PFDMA (dynamic) stayed short. According to the equation for updating α, α becomes large when the occlusion duration becomes longer. As a result, in the M-PFDMA (dynamic), more particles are deployed close to the detected face and more rapid restart of stable tracking is achieved. The above experiments verify that the dynamic update of α enhances tracking stability against abrupt movement and recoverability from tracking failure. Further, the results suggest that occlusion duration and the pose when the occluder clears will be key information. 6.

Conclusion

A memory-based particle filter with detection-based memory acquisition, M-PFDMA, was proposed for vision-based object tracking. M-PFDMA offers robust memory acquisition under severe occlusion since it creates a synergistic combination of detection-based memory acquisition and the memory-based approach. M-PFDMA was shown to achieve high accuracy and quick recovery in real-world situations. We verified its effectiveness in facial pose tracking experiments. Future works include dynamic update of mixing weight α and memory management. We have already discussed a little about dynamic update of α in Sect. 5. We disclosed the duration of tracking instability and the pose of the lost face when it rediscovered may affect adequate mixing weight α. We plan to consider the strategy of determining α. Memory management is also important issue. M-PFDMA stores the correctly estimated target state in memory. The correctness is automatically judged by using the maximum likelihood among particles. Though it works well in most cases, the quality of stored data is very important for memory-based prior distribution prediction. Therefore, we will consider more precise ways of judging tracking correctness. References [1] D. Comaniciu, V. Ramesh, and P. Meer, “Kernel-based object tracking,” IEEE Trans. Pattern Anal. Mach. Intell., vol.25, no.5, pp.564– 577, 2003. [2] G.R. Bradski, “Computer vision face tracking for use in a perceptual user interface,” Proc. IEEE Workshop Applications of Computer Vision, pp.214–219, 1998. [3] J. Tua, H. Taob, and T. Huang, “Face as mouse through visual face

tracking,” CVIU, vol.108, no.1-2, pp.35–40, 2007. [4] K. Otsuka, S. Araki, K. Ishizuka, M. Fujimoto, M. Heinrich, and J. Yamato, “A realtime multimodal system for analyzing group meeting by combining face pose tracking and speaker diarization,” Proc. ACM ICMI, pp.257–264, 2008. [5] N. Gordon, D. Salmond, and A.F.M. Smith, “Novel approach to non-linear and non-Gaussian Bayesian state estimeation,” lEE Proc. F: Communications Rader and Signal Processing, vol.140, no.2, pp.107–113, 1993. [6] M. Isard and A. Blake, “Condensation - conditional density propagation for visual tracking,” IJCV, vol.29, no.1, pp.5–28, 1998. [7] D. Mikami, K. Otsuka, and J. Yamato, “Memory-based particle filter for face pose tracking robust under complex dynamics,” Proc. CVPR, pp.999–1006, 2009. [8] D. Mikami, K. Otsuka, and J. Yamato, “Memory-based particle filter for tracking objects with large variation in pose and appearance,” IEICE Trans. Inf. & Syst. (Japanese Edition), vol.J94-D, no.8, pp.1194–1205, Aug. 2011. [9] L. Tian, S. Ando, A. Suzuki, and H. Koike, “A probabilistic approach for fast and robust multi-view face detection using compact local patterns,” Proc. IIEEJ Image Electronics and Visual Computing Workshop, 2010. [10] E. Murphy-Chutorian and M.M. Trivedi, “Head pose estimation in computer vision: A survey,” IEEE Trans. Pattern Anal. Mach. Intell., Digital Library, 2008. [11] M. Ozuysal, V. Lepetit, F. Fleuret, and P. Fua, “Feature harvesting for tracking-by-detection,” LNCS 3953 (ECCV 2006), Part III, pp.592–605, 2006. [12] M. Andriluka, S. Roth, and B. Schiele, “Monocular 3D pose estimation and tracking by detection,” Proc. CVPR, pp.623–630, 2010. [13] Y. Kobayashi, D. Sugimura, Y. Sato, K. Hirasawa, N. Suzuki, H. Kage, and A. Sugimoto, “3D head tracking using the particle filter with cascaded classifiers,” Proc. BMVC2006, 2006. [14] S. Ba and J.-M. Odobez, “Probabilistic head pose tracking evaluation in single and multiple camera,” Multimodal Technologies for Perception of Humans, 4625/2008, pp.276–286, 2008. [15] K. Okuma, A. Taleghani, N. Freitas, O. Freitas, J. Little, and D. Lowe, “A boosted particle filter: Multi-target detection and tracking,” Proc. ECCV2004, pp.28–39, 2004. [16] D. Sugimura, Y. Kobayashi, Y. Sato, and A. Sugimoto, “Incorporating long-term observations of human actions for stable 3D people tracking,” Proc. IEEE Workshop on Motion and Video Computing (WMVC2008), 2008. [17] D. Mikami, K. Otsuka, S. Kumano, and J. Yamato, “Enhancing memory-based particle filter with detection-based memory acquisition for robustness under severe occlusion,” Proc. VISAPP2012, 2012. [18] O.M. Lozano and K. Otsuka, “Real-time visual tracker by stream processing,” J. VLSI Signal Processing Systems, vol.57, no.2, pp.285–295, 2008. [19] P. Viola and M.J. Jones, “Robust real-time face detection,” IJCV, vol.57, no.2, pp.137–154, 2004. [20] FaceAPI, “http://www.seeingmachines.com/product/faceapi/”

MIKAMI et al.: MEMORY-BASED PARTICLE FILTER WITH DETECTION-BASED MEMORY ACQUISITION

2703

Dan Mikami received his B.E. and M.E. degree from Keio University, Kanagawa, Japan, in 2000 and 2002, respectively. He has been working for Nippon Telegraph and Telephone Corporation, NTT from 2002. He received his Ph.D. in engineering from Tsukuba University in 2012. His current research activities are mainly focused on robust visual object tracking. He received the Meeting on Image Recognition and Understanding (MIRU) 2009 Excellent Paper Award, the IEICE Best Paper Award 2010, the IEICE KIYASU-Zen’iti Award 2010, and the MIRU2011 Interactive Session Award. He is a member of the IEEE.

Kazuhiro Otsuka received his B.E. and M.E. degrees in electrical and computer engineering from Yokohama National University in 1993 and 1995, respectively. He joined the NTT Human Interface Laboratories, Nippon Telegraph and Telephone Corporation in 1995. He received his Ph.D. in information science from Nagoya University in 2007. He is now a senior research scientist in the NTT Communication Science Laboratories and is entitled as a distinguished researcher in NTT. His current research interests include communication scene analysis, multimodal interactions, and computer vision. He was awarded the Best Paper Award of IPSJ National Convention in 1998, the IAPR Int. Conf. on Image Analysis and Processing Best Paper Award in 1999, the ACM Int. Conf. on Multimodal Interfaces 2007 Outstanding Paper Award, the Meeting on Image Recognition and Understanding (MIRU) 2009 Excellent Paper Award, the IEICE Best Paper Award 2010, the IEICE KIYASU-Zen’iti Award 2010, and the MIRU2011 Interactive Session Award. He is a member of the IEEE and the IPSJ.

Shiro Kumano received the Ph.D. degree in Information Science and Technology from the University of Tokyo in 2009. He is currently a researcher at NTT Communication Science Laboratories. His research interests include computer vision, human behavior analysis, and automatic meeting analysis, especially in facial expression recognition and emotion estimation. He received the Honorable Mention Award at the Asian Conference on Computer Vision (ACCV 2007) and the MIRU2011 Interactive Session Award. He is a Member of the IEEE and IPSJ.

Junji Yamato is the Executive Manager of Media Information Laboratory, NTT Communication Science Laboratories. He received the B.E., M.E., and Ph.D. degrees from the University of Tokyo in 1988, 1990, and 2000, respectively, and the S.M. degree in electrical engineering and computer science from Massachusetts Institute of Technology in 1998. His areas of expertise are computer vision, pattern recognition, human-robot interaction, and multiparty conversation analysis. He is a visiting professor of Hokkaido University and Tokyo DENKI University and a lecturer of Waseda University. He is a senior member of the IEEE, and a member of the Association for Computing Machinery.

Enhancing Memory-Based Particle Filter with Detection-Based ...

Nov 11, 2012 - The enhance- ment is the addition of a detection-based memory acquisition mechanism. The memory-based particle filter, called M-PF, is a ...

3MB Sizes 0 Downloads 331 Views

Recommend Documents

Boosting Target Tracking Using Particle Filter with Flow ...
Target Tracking Toolbox: An object-oriented software toolbox is developed for implementation ..... In data fusion problems it is sometimes easier to work with the log of a ..... [13] Gustafsson, F., “Particle filter theory and practice with positio

Particle Filter Integrating Asynchronous Observations ...
Position tracking of mobile robots has been, and currently ..... GPS. 1. 20 ot. 4. Camera Network. 1. 500. The experimental testbench was composed by two com- puters. .... Particle Filters for Online Nonlinear/Non-Gaussian Bayesian Tracking,”.

Convergence Results for the Particle PHD Filter - CiteSeerX
convergence of the empirical particle measure to the true PHD measure. The paper first ... tation, or Particle PHD Filter algorithm, is given in Section. Daniel Edward Clark ...... [Online]. Available: citeseer.ist.psu.edu/crisan00convergence.html. [

Convergence Results for the Particle PHD Filter
[2], the random finite set Θ was represented by a random counting measure nΘ ..... error of the PHD Particle filter at each stage of the algorithm. These depend on ..... t−1. ) 2. |Qt−1]), where the last equality holds because the particles are

Convergence Results for the Particle PHD Filter - CiteSeerX
distribution itself. It has been shown that the PHD is the best-fit ... Electrical and Computer Engineering, Heriot-Watt University, Edinburgh. [email protected] ... basic idea of point processes is to study collections of point occurrences, the .....

Particle PHD Filter Multiple Target Tracking in Sonar ...
The matrices in the dynamic model ... The PHD is approximated by a set of discrete samples, or ... observation covariance matrix R. The dot product of the.

Particle Filter based Multi-Camera Integration for ...
calibrated cameras via one color-based particle filter. The algorithm re- ... ensured using a multi-camera system, which guarantees broad view and informa-.

Probabilistic Multiple Cue Integration for Particle Filter ...
School of Computer Science, University of Adelaide, Adelaide, SA 5005, ..... in this procedure, no additional heavy computation is required to calculate the.

Filter with efficiently sealed end
Jan 30, 2009 - 7/1995 Paas. 5,484,466 A. 1/1996 Brown et al. 5,487,767 A. 1/1996 Brown. (Continued). FOREIGN PATENT DOCUMENTS. DE. 3001674.

Filter with efficiently sealed end
Jan 30, 2009 - 5/2002. 6/2002. 7/2002. 9/2002. 11/2002. 1/2003. Aimura et al. CoulonvauX ...... an outer perimeter, and has a second section extending radi.

Enhancing billing system efficiency with cloud computing
architecture-based billing system—including computing performance, ... with Intel Xeon process E7 family and cloud computing technology enables a reliable.

Semiconductor laser with integral spatial mode filter
Oct 15, 1999 - Mittelstein et al., “Broadband tunability of gain—?attened quantum Well .... 21, 1987. Surerte et al., “High—PoWer Ring Laser Using a Broad—Area ...... speed modulation and loWer modulation current require ments are thus ...

Display device with a transparent optical filter
Aug 19, 2010 - 5/1991. JP. A8-204595. 8/1996. JP. A 3-156420. 7/1991. JP. 8-234162. 9/1996 .... as a personal digital assistant (PDA), an electronic notebook.

Enhancing billing system efficiency with cloud computing
Adopt a cloud computing solution. Use Intel Xeon processor E7-8800/4800 product families to build an enhanced cloud computing platform that provides ...

Particle Filter Based Localization of the Nao Biped Robots
Moreover, kidnap scenarios which could not be considered and handled with the uni-modal Kalman .... Thereafter, by running a binary search on green color between the current pixel and the previous checked one, ... motions of the cameras installed on

Cheap Zomei 0.42X Fisheye Lens Filter Phone HD Filter with Macro ...
Cheap Zomei 0.42X Fisheye Lens Filter Phone HD Filt ... xy S7 Note 6 iPhone 6 6S Plus HTC Etc CellPhone.pdf. Cheap Zomei 0.42X Fisheye Lens Filter Phone ...

Computing Stable Skeletons with Particle Filters
including visual tracking, speech recognition, mobile robot localization, robot map building ..... There are two important differences in comparison to the standard .... filter can deal with the condition that the path between two endpoints are not.

Particle Concentration and Characteristics near a Major Freeway with ...
vehicles in the Los Angeles highway network. Particle concentration measurements were accompanied by measurements of black carbon, elemental and organic carbon, and gaseous species (CO, CO2). Using the incremental increase of CO2 over the background

Phishing Website Detection Using Random Forest with Particle Swarm ...
Phishing Website Detection Using Random Forest with Particle Swarm Optimization.pdf. Phishing Website Detection Using Random Forest with Particle Swarm ...

Particle-in-cell beam dynamics simulations with a ...
Mar 2, 2007 - simulations, a good initial approximation to the iterative solution is always readily .... applying it to two analytic potential-density pairs of inter-.

The Cantonese utterance particle gaa3 and particle ...
Metalanguage (NSM) framework and natural speech data from the Hong Kong. Cantonese Corpus to ..... 'But you need – no aa3, [to participate in] those you need to plan for the correct time. (4) gaa3. ..... Both try to back up their arguments ...

Real-Time Particle Filtering with Heuristics for 3D ...
3D motion capture with a consumer computer and webcam. I. INTRODUCTION ... 20 degrees of freedom for body poses) where multiple local minimums may ...