Vision-Based Pedestrian Detection for Rear-View Cameras Shai Silberstein, Dan Levi, Victoria Kogan and Ran Gazit

Abstract— We present a new vision-based pedestrian detection system for rear-view cameras which is robust to partial occlusions and non-upright poses. Detection is made using a single automotive rear-view fisheye lens camera. The system uses “Accelerated Feature Synthesis”, a multiple-part based detection method with state-of-the-art performance. In addition, we collected and annotated an extensive dataset of videos for this specific application which includes pedestrians in a wide range of environmental conditions. Using this dataset we demonstrate the benefits of using part-based detection for detecting people in various poses and under occlusions. We also show, using a measure developed specifically for videobased evaluation, the gain in detection accuracy compared with template-based detection.

(a) HoG

(b) Accelerated Feature Synthesis

(c) Detection examples

I. I NTRODUCTION According to the National Highway Traffic Safety Administration (NHTSA) about 18, 000 people a year are hurt in back-over accidents in the U.S. alone, with about 3, 000 suffering “incapacitating” injuries. 44 percent of the incidents involve children under age 5. In a related study [19], NHTSA found that no efficient sensing-based solution exists for preventing many of these accidents, and that camerabased solutions are the most promising technology. In the recent years, significant progress has been made in the field of camera-based pedestrian detection (e.g. [24]). However, current solutions focus on front applications and are designed for upright, fully visible pedestrians. Such solutions do not work for pedestrians in non-upright poses, and specifically children. They have also difficulties in detecting partially occluded pedestrians, and close-by pedestrians which are also not fully visible to the camera. Solving these issues is critical for reaching systems capable of preventing back-over accidents. Existing vision-based on-vehicle pedestrian detection systems mainly rely on so-called “template-based” appearance models. These models capture the pedestrian appearance using one rigid template matched to each image position. Template-based detection, such as the Histograms of Oriented Gradients (HoG) [5] shown in figure 1(a), can deal with small local deformations of the pedestrian shape, but cannot handle large ones caused for example by object articulations. In the recent years “part-based” appearance models such as the Deformable Part-based Model (DPM) [11] and Feature Synthesis (FS) [2] have shown capability to handle general object detection tasks and improve benchmark results [10], [7]. As illustrated in figure 1(b), part-based models capture Shai Silberstein, Dan Levi, Victoria Kogan and Ran Gazit are with the Advanced Technical Center - Israel, General Motors R&D, Hamada 7 Herzlyia, ISRAEL. [email protected],

{dan.levi,victoria.kogan,ran.gazit}@gm.com

Fig. 1. (a) Pedestrian HoG template (from [5]). (b) Examples of “localized part” features from the hundreds used in our appearance model. The rectangle surrounds the image fragment, used as a part model. The circle represents the possible spatial locations of the part which are modeled by a gaussian distribution (the circle marks its 1-std).(c) Examples of detections by our system under occlusions and with non-upright pose. Both examples were not detected by HoG-based system, with both systems tuned to 1 false alarm per second (FPAS) on the rear view pedestrian dataset. The images were cropped for clarity and appear in full in figure 7. See Section IV for further details.

appearance of single object parts and combine the parts using a geometric model of spatial part relations. This flexibility allows the model to inherently support large variations in object appearance caused for example by object articulation (i.e. different human poses) and partial occlusions [16]. Moreover, increasing the number of parts in the model can further improve robustness [2]. The main obstacle towards implementation of part-based detection algorithms remains their high computational cost. Most existing systems are therefore limited to only several parts [11], [21]. Pedestrian detection in the rear-view camera poses challenges stemming from two different factors: the camera characteristics and the application scenarios. Rear-view cameras will be mandated in all new vehicles in the U.S. in the next coming years. Their main purpose is driver display, and they are typically targeted to cover a wide angle, 180◦ FieldOf-View (FOV), at a low cost. A common solution uses a VGA resolution fisheye lens camera with substantial distortions especially in the periphery. The scenarios in which the application must operate are also different from front

applications. Backing over is done many times in parking lots, where pedestrians and in particular children may appear, in various activities and in many different postures. The proposed system addresses these challenges directly. This paper builds on a multiple-part based detection method, the Feature Synthesis (FS) which was first put forth in [2]. The FS uses hundreds of parts in its appearance model. A real time version, the Accelerated Feature Synthesis (AFS), was presented in [17], making the use of this model practical in terms of running time. The main contribution in this paper is a robust on-vehicle system for rear-view pedestrian detection using the AFS along-side additional algorithmic blocks. The proposed system includes three additional algorithmic modules: image warping that corrects the camera distortions, geometric reasoning for cropping irrelevant image regions which improves both running time and accuracy performance, and frame-to-frame target association which enables the use of video-based evaluation. The system is a first real-time on-vehicle system making use of multiple-part based object detection. We also provide an experimental video-based evaluation of our system compared to a template-based method on a new pedestrian dataset captured using the rear-view camera. The dataset contains over 200K annotated pedestrians in various real world scenarios. We use a video-based evaluation methodology which we developed in [1] to show the performance gained by using our part-based detection method compared to a standard template based method (HoG). We also show the advantage our system has in detecting partially occluded people and people in various poses. This paper is organized as follows. We present an overview of related works in the next section. We then describe the proposed system in Section III. In Section IV we introduce our new dataset as well as a method for evaluating detection in video data, and the experimental evaluation results. Conclusions and future work are presented last in Section IV-D. II. R ELATED W ORK Vision-based pedestrian detection systems based for Advanced Driver Assistance Systems (ADAS) have been extensively studied in the last decade [16]. Monocular visionbased systems usually include several stages: Foreground segmentation, in which candidate areas are extracted from the image; Object classification, which uses an appearance model to classify candidates; Verification in which candidates are eliminated based on information which is orthogonal to the classifier; And tracking, which exploits the continuity of the detection location over time. Different systems propose different implementations to each of these stages. This paper focuses on the object classification stage. The majority of the systems use template-based classifiers, either by matching the silhouette [3], [15] or the full appearance [20], [5]. Modern methods (e.g. [20], [5], [15]) use machine learning techniques such as Support Vector Machines (SVMs) in order to train the classifier. Several systems decompose the template into several individual parts (e.g. [22], [21]), but since the part location is fixed with respect to the template,

such models are not suitable for multi-pose detection. Most related to our work is the automotive pedestrian detection system presented in [4], which uses an optimized cascadeDPM based detector [12], but is limited to using several parts only to run in real time. For a in depth survey on ADAS pedestrian detection systems we refer the reader to [16]. Part-based methods [14], [11], [2] use object parts with a deformable configuration to model objects, increasing their ability to cope with partial occlusions and large appearance variations compared with template-based methods. Furthermore, using a large number of parts learned automatically with diverse appearances improves detection accuracy [2]. Evidently, part-based methods are highly ranked on large scale benchmarks [10], [7]. Such methods, however, are either limited in the number of parts modeled [11] to be able to run in reasonable time, or are impractical in terms of run time [2]. Accelerating part-based detection [12], [8], [6] efforts in the past mostly focused on methods relying on a small number of parts such as the DPM [11] since computation time increases linearly with the number of parts. In our detection system we use the Accelerated Feature Synthesis (AFS) algorithm, which we presented in [17]. The AFS uses hundreds of parts in its object model, and runs in real-time, maintaining performance comparable with the original FS algorithm. Acceleration is achieved by a series of approximations as following. A “coarse-to-fine” strategy for early window elimination and location refinement is used. In each image location, only the closest parts are compared, and for each part, only locally maximal-appearance positions are used for classification. In contrast, existing partbased methods (e.g. DPM,c-DPM) consider all parts in a dense grid of positions. Finally, to find the closest parts in each image positions a novel approximate nearest neighbor (ANN), search, termed “KD-Ferns”, has been developed. Several pedestrian detection datasets were collected in the past. The INRIA [5] and MIT [20] were among the first pedestrian detection datasets but are relatively of small scale. Recently a number of databases were collected specifically in driving scenarios with O(100K) pedestrians marked: the Caltech [7] and DaimlerDB [9]. These databases, however, are not directly applicable for testing rear-view camera pedestrian detection, since the rear-view camera imagery is very different in FOV and quality, and the scenarios and driving speeds are different as well. In [23] a dataset was specifically collected for evaluating rear-view pedestrian, but since it is currently not public we could not use it for our evaluation. In addition, the reported evaluation uses a cropped-window measure, which gives a poor prediction on full-frame performance [7]. In our work we collected a new large dataset specifically for the rear-view pedestrian detection application, containing more than 200, 000 marked pedestrians in full length video sequences. We measure performance on this dataset using a methodology we specifically developed for evaluating video-based detection [1].

(a)

(b)

Fig. 3. (a) Captured image example. (b) Warped image in virtual forwardlooking view. Notice the pedestrian becomes straight and upright.

Fig. 2.

Rear-view pedestrian detection system flowchart.

III. P EDESTRIAN D ETECTION S YSTEM The system hardware configuration is as following: a rearview camera (NTSC, 180◦ FOV) is connected to an i7 laptop via a frame grabber. The video is grabbed at 30 frames per second at VGA resolution (640 × 480). The detection system running on the laptop is composed of several algorithmic blocks as described in figure 2. In Stage A, image warping, we perform lens correction and virtual forward view rendering, which brings straight upright objects in the world to appear straight and vertical in the image. Stage B consists of geometry-based image scale-pyramid creation. Stage C consists of applying the accelerated feature synthesis detection method to the image pyramid, producing the detections in the current video frame. Finally using a simple association mechanism, targets are tracked across frames and are assigned the same ID. The output in each frame is a list of targets with the following properties: [x, y, w, h, S, ID] image position, width and height, classification score and ID. We next describe each step in more details. A. Image warping The first stage in the system consists of warping the image such that straight upright objects in the world appear straight and vertical in the image (Block A in figure 2). The purpose of this operation is to obtain a pedestrian appearance which is invariant to the camera lens distortion, the camera tilt and to the image position of the pedestrian. The warping transformation is automatically computed using the following steps. We compute the internal and external camera parameters using the Caltech Camera Calibration Toolbox. The tilt is obtained by including several checkerboards on the ground in the calibration process and averaging the normal direction

of the ground in camera reference coordinates. From this calibration we obtain two warping transformations: the radial lens distortion correction, and the Homography which transforms the image into a virtual forward-looking view. We then compute a scaling transformation that ensures that we can detect a pedestrian at minimal height h and maximal distance D, given that the system requires a target height of 90 pixels. The system was tuned to h = 170cm, D = 7m. Finally, we crop the image output size to cover 180◦ FOV in width and almost full image extent in height. We also developed a semiautomatic tool enabling the user to see the resulting warped image, with upright objects and distance markers enabling the user to fine-tune the warping parameters. Figure 3 shows a captured image (a) and the warped image (b). In the first image of figure 7 we show the extent of the warped image in the original image marked by a blue dotted line. B. Image Pyramid In the second stage (stage B in figure 2), the warped image is down-sampled at multiple resolutions creating an image scale pyramid. This is the standard technique for applying a fixed size detector to detect pedestrians at multiple sizes. The detector works in 4 full octaves (×0.5, ×0.25, . . .) and 4 sub-octaves, 16 scales in total. Using the known camera extrinsic parameters we use a geometrical model to crop out irrelevant regions in the image assuming a “roughly planar ground” as explained next. In each scale s of the image pyramid we use a fixed detector of size 128×64 pixels bounding a ∼ 90 pixels height pedestrian within it. The bounding box at scale s corresponds to a larger bounding box size with height hs pixels in the original image. We compute for a bounding box of height h = hs , what are the positions in the image in which it can appear assuming it tightly bounds a pedestrian with reasonable height located on a flat ground. These positions can be shown to lie between two horizontal rows in the image, defined by ymin , ymax , which are computed as following (Note we set (0, 0) as bottom left image pixel). Given a pedestrian bounding box with bottom part at pixel yb and top part at pixel yt , we can compute its height in the real world, H, as explained in figure 4. The result is a mapping M from each yb ,yt to the respective pedestrian height H. We then query M to obtain all entries Q in which yt − yb = h and H min < H < H max , where H min , H max are the parameters for minimal and maximal pedestrian height respectively. The

Fig. 4. Computing pedestrian height H from bounding box top (yt ) and bottom (yb ) in the image. Known camera parameters are: focal length (f ), image center row from bottom (vc ) and camera height in cm (Yc ). Since we render a virtual front view the camera is pointed towards the horizon. Distance to is given by: Z = Yc · ) the pedestrian ( b . Combining the two: Z = tan α. In addition tan π2 − α = vc −y f ( ) ′ , the point of b Yc · tan π2 − arctan vc −y . We similarly compute Z f intersection between the ground and the ray from camera center to pedestrian ′ top. Using similar triangles the height of an object is then: H = Yc · Z Z−Z ′ .

valid raw range is set accordingly to yˆmin = min{yb ∈ Q},ˆ ymax = max{yt ∈ Q}. This formulation makes two assumptions: a flat ground and known camera tilt. To relax these assumptions we allow for some uncertainty in the tilt (±3◦ ). Using the same computation above for the each of the two tilt values we obtain a wider range [ymin , ymax ]. Finally, the image in pyramid scale level s is cropped in the row domain by the range [ymin , ymax ]. C. Accelerated feature synthesis Accelerated Feature Synthesis (AFS) [17] is a sliding window object detection method, based on the Feature Synthesis (FS) [2] method. We developed the method to enable real-time part-based detection, and it is the core technology in the system (block C in figure 2). We start by describing the FS [2] method. In the FS, a part-based classifier model C discriminates sub-image windows Is of fixed size wx × wy as tightly containing the object or not. C is trained using a sequential feature selection method and a linear-SVM classifier. C is parameterized by F , a set of classifier features, R, a set of rectangular image fragments extracted from training images, and W = {Wf } the linear classifier weights. Computing C(Is ) ∈ R, the classification score of sub-image Is proceeds as follows. For each fragment r ∈ R the “fragment similarity map” ar (x, y) represents the appearance similarity of r to each (x, y) position in Is . ar (x, y) is computed as the innerproduct between the 128-dimension SIFT descriptor [18] of r and that of the image fragment in position (x, y). Subsequent stages use a list of spatially sparse fragment detection locations Lr = {lk = (xk , y k )}K k=1 computed by finding the K = 5 top local maxima in ar . The appearance score of each location l ∈ Lr is then ar (l). Each feature f ∈ F is a function f : Is 7→ R, computed using the fragment detections Lr of one or more fragments r. Each feature f represents different aspects of object-part detections. From the families of features suggested in [2], we use in the AFS only ones which significantly contribute

to performance: GlobalMax, Sigmoid, Localized, LDA and HoG component features. For example, a localized feature is computed as: f (Is ) = maxl∈Lr G(ar (l)) · N (l; µr , σI2×2 ) where N is a 2D Gaussian function of the detection location l and G, a learned sigmoid function on the appearance score. Such features represent location sensitive part detection, attaining a high value when both the appearance score is high and the position is close to a preferred part location µr , similar to parts in a star-like model [13]. Figure 1(b) shows 3 selected localized features. The final classification score is a linear combination of the feature values: C(Is ) = ∑ f ∈F (Wf ·f (Is )). We next describe briefly the AFS method for detecting objects in full images in real-time. AFS consists of 3 stages: a coarse detection stage which outputs candidate targets and a fine detection stage which scores all windows around the candidate targets. The input is the image pyramid and the output is the object detections represented by a set of bounding boxes at multiple locations and scales in the image and their classification scores. The AFS algorithm flow (see Figure 5) is composed of a twolevel coarse-to-fine cascade. The coarse level uses the sliding window methodology. It uses a trained coarse classifier C1 = (F1 , R1 , W1 ) to compute the classification score for a dense set of sub-windows sampled in scale and position space. For a specific scale, sub-windows are sampled on a regular grid with a spatial stride s = s1 pixels. Image locations which received a large enough score are then passed to the second level. Around each such location a local region is defined and sub-windows are sampled in that region on a finer grid with stride s = s2 and processed by the second level with classifier C2 = (F2 , R2 , W2 ) to produce the final classification score. Computing the classification score for each sampled sub-window is similar for both cascade levels. We refer to this procedure as a one-level detection (blue rectangles in Figure 5). The input to the first-level detection is the entire image pyramid and to the second level detection only the candidate image regions. Each image area is processed (either by coarse or fine detector) is in three stages, parameterized by the classifier model C(F, R, W ). The first stage computes local gradient orientation histograms, then fragment similarity maps are computed and finally the classification scores. Each of the computations has been optimized for speed with minimal accuracy loss. In the fragment similarity map computation we use a novel ANN algorithm, “KD-Ferns” developed specifically for this application. For further details on the computation of each stage in the AFS please refer to [17]. The output of this stage is a list of detections or bounding boxes, where each bounding box A is defined by its image position and classification score (S): A = [x, y, w, h, S]. D. Non-Maximal Suppression The next stage is the Non-maximal suppression (NMS) (Block D in figure 2), which removes redundant and inaccurate detections. We experimented with different variations of NMS, and found the following to maximize detection performance.

AFS algorithm Level 1 (coarse) Input image

Local gradient orientation histograms

Fragment similarity maps

Classification score Candidate object locations

Level 2 (fine) Local gradient orientation histograms

Fragment similarity maps

Classification score

Output: Final object detections

Fig. 5. Accelerated Feature Synthesis (AFS) detection algorithm flow. First level processes a full scale pyramid of the image while the second level processes only regions around candidate locations from level 1 and returns the final detections.

We compute a graph in which each detection is a represented by a node. Let A, B be two detections, and assume w.l.o.g. that s(A) < s(B). Then an edge connects two detections A and B if one of two following conditions is met: ∩ |A B| • √ > t1N M S |A|·|B| ∩ |A B| > t2N M S , |A| |A|

< |B|, t2N M S > 1 − δ Where |·| denotes area. In other words, either the bounding boxes sufficiently overlap, or A, the lower scoring bounding box, is almost fully contained by B. Once graph is created, any detection A with a graph neighbor B s.t. s(A) < s(B) is removed. In all the reported results the NMS parameters were set to: t1N M S = 0.75, t2N M S = 0.97. •

E. Target association The last stage of the detection is assigning a target ID to each bounding box (Block E in figure 2). This is done for the sole purpose of enabling video-based evaluation. In the video-based evaluation, explained in detail in section IV-B, real targets and false alarms on the same object are grouped to single detections across several frames for evaluation. Consequently, to obtain a good evaluation result, the detection system is encouraged to keep the target ID consistent across video frames. Our association mechanism follows a “tracking-by-detection” approach. In the first frame we assign arbitrary unique IDs to each target. We then associate targets in the current frame to close-by targets in the previous frame as following. Given a bounding box Bt in frame t, we look for the closest bounding box Bt−1 in frame t − 1, according to the mean squared distance between the upper left and lower right corners of the two bounding boxes. We then assign Bt the same ID as that of Bt−1 . In case multiple bounding boxes are assigned with the same ID, only the highest scoring one gets this ID, while the other are assigned with new IDs. Finally, we translate the bounding box to the original image by transforming the bounding box center to its coordinates in the original image and estimating the transformation of the bounding box width and height. IV. E XPERIMENTAL RESULTS We test our system on a new rear-view pedestrian dataset we collected. The dataset was collected in various condi-

tions and situations relevant to backing over scenarios and contains a significant amount of annotated pedestrians. We use a video-based evaluation [1], developed for measuring performance as perceived in a continuous visual display. We next describe the dataset, the evaluation criteria and the evaluation result. A. Rear-view pedestrian detection dataset To collect the dataset we mounted the fisheye camera previously described on a vehicle. We mounted the camera in front for safety reasons, but in a typical rear-view installation pose: 107cm height and 25◦ downward tilt angle. The dataset contains 15 filming sessions, each taken in a different day with different scenarios. Each session contains multiple clips with duration ranging from several seconds to several minutes. In total, the dataset contains 250 clips with a total duration of 76 minutes and over 200K annotated pedestrian bounding boxes. Each clip was recorded as a raw 30f ps with simultaneous recording of the vehicle CAN information. There are two types of sessions, containing either staged or “in-the-wild” pedestrians. The staged scenarios include mainly pedestrians walking in front of the camera at different positions and directions in a controlled manner, spanning the different use cases of a rear alert or breaking automotive feature. In the remaining sessions the vehicle drove either in public roads or in parking lots and captured incidental pedestrians. The different locations include: indoor parking lots, outdoor paved/sand parking lots, city roads and private driveways. We filmed both day and night scenarios, with different weather and lighting conditions. We also filmed one session in our lab with staged people in various poses. Figure 7 illustrates the variety of scenarios captured. The entire dataset was annotated semi-manually by marking a bounding box around each pedestrian in each video frame, using the annotation tool provided by [7]. This tool allows skipping several frames, completing the intervals automatically and allowing fast corrections. We chose 70 clips as the test portion of the dataset for evaluation purposes, while the training part was kept for future system improvements. Night and lab sessions were excluded from the test set. B. Evaluation criteria For quantifying the detection accuracy of the system we developed a video-based evaluation criterion previously published in [1]. Evaluation methods in previous computer vision and automotive pedestrian detection work [10], [7] were devised to test detection performance in single images or independent movie frames. The ground truth annotations in this methodology are provided as bounding boxes in the images. For two detection bounding boxes A, B, their Area of Overlap Ratio (AOR) is defined by: ∩ |A B| ∪ AOR(A, B) = |A B| A detection bounding box is intersected with all the ground truth boxes in the image and if their AOR(A, B) > t then the detection is considered to be a true positive. Otherwise, the

(a)

(b)

(c) 1 .80 .64 .50 .40

miss rate

.30 .20

.10

.05

99% VJ 97% Shapelet 93% LatSvm−V1 93% PoseInv 93% HogLbp 89% HikSvm 87% HOG 87% FtrMine 86% LatSvm−V2 84% MultiFtr 82% MultiFtr+CSS 82% Pls 80% MultiFtr+Motion 79% AccFeatSynth 78% FPDW 78% FeatSynth 78% AccFeatSynth+Geometry 77% ChnFtrs −3

10

−2

10

−1

10

0

10

1

10

false positives per image

Fig. 6. a: Frame-based evaluation on the rear-view pedestrians dataset. In parenthesis: log-average miss rate between 10−2 and 100 fppi (where such data exists). b: Video-based evaluation on the rear-view pedestrians dataset. In parenthesis miss rate at 1 false alarm per second. c: Frame-based evaluation on the Caltech dataset medium scale pedestrians. The “AccFeatSynth+Geometry” corresponds to the image pyramid block + AFS algorithm block from the presented system, and the “AccFeatSynth” corresponds to the AFS block only. See [7] for more details on the dataset and the evaluated methods.

Fig. 7. Examples of system outputs. Top-left image shows the full view of the rear-view camera, with blue-dotted line marking the operation area of the detection system. The rest of the images were cropped to working zone. Marked by green rectangles are the true positive detections above the 1 FAPS threshold and in red the false positives above the same threshold.

detection is considered a false positive. Ground truth objects not matched to any detection are counted as misses. The miss detection statistic is the ratio of number of misses to the number of ground truth objects, and the false positives per image (fppi) rate is the number of false positives divided by the total number of frames. We first use this criteria, which we denote by frame based evaluation, to obtain statistics on the frame independent detection quality. In frame based evaluation, the number of misses is calculated by summing them over all frames. This definition is simple, but it has the disadvantage that both of the following scenarios would have the same missed detection rate: Scenario A: Targets 1 and 2 are both repeatedly detected and missed with a cycle period of half a second. Scenario B: Target 1 is detected in all frames and Target 2 is missed in all frames. Clearly Scenario A is preferable to B from an application point of view, but both would result in the same score. Therefore, we define a grace period, Time of Miss Gap (TMG). Misses lasting for TMG seconds or less, which occur after a successful detection, are not counted. For false positives, the per-frame false alarm rate is meaningless from a user point of view. Even when injected the same movie, systems are different from each other with respect to the number of frames they process. We therefore shift to a false alarm per-second measure. In addition, we define the minimum duration Time of False Gap (TFG) as the time duration counted as a single false alarm. Once a false alarm occurs, and additional false alarm is counted for every TFG seconds that the false target track persists. There is no grace period for the initial false target report: a false detection is counted when it first appears. The evaluation below uses the following values: TFG= 0.5 seconds, TMG= 0.5 seconds, and t = 0.2. These values were chosen subjectively and in the future we recommend choosing them according to user experience studies. The result is reported as the miss ratio versus the number of false alarms per second (FAPS), and called the video based evaluation. The evaluation also considers the goal range and FOV of the detection system. In our case we set it to 7 meters and 130◦ . We translate it roughly to a minimal height of the bounding box around the pedestrian (64 pixels), and to left/right bounds on the area of detection in the image (corresponding to the dotted-blue marked area in figure 7). In order to avoid false alarms which are correct or incorrect detections outside these bounds, we mark the ground-truth objects and detections outside the bounds as “don’t care” and disregard them in the evaluation. If a bounding box within bounds is associated with a “don’t care” ground-truth it will be not counted as a negative or positive detection. The same goes for distant grouped people, which were difficult to annotate as separate individuals, and were marked as “people”. These objects are added to a “don’t care” list of objects.

C. Evaluation results We evaluate our system using both frame and video based evaluation on the test part of the rear-view pedestrian dataset described previously. As a baseline method we use the exact same system except for the AFS (Block C in figure 2) replaced by the HoG [5] template-based detection. The HoG is trained on the INRIA pedestrians dataset [5], as is the AFS. We refer to the two rear view detection systems in short as rv-AFS and rv-HoG. In figure 7 we show several examples of the detection results of the our system (rvAFS) including typical false alarms. In all these examples the detection threshold is set to obtain 1 FAPS in the videobased evaluation, which is the working point we use in real time operation of the system. As can be observed the system can perform in poor lighting conditions, heavy occlusions and large image distortions. Figure 6(a) summarizes the results using the frame-based evaluation. We plot the miss rate as a function of the fppi on a log-log scale. The number in parentheses beside each method name is the log-average of the miss rates between 10−2 and 100 fppi. Figure 6(b) shows the video based evaluation result. We plot the FAPS versus the miss detection rate. Here, the number in parentheses is the miss rate at 1 FAPS. At this working point our system (rv-AFS) miss rate is 5% compared to the rv-HoG with 7.4%. Results from both evaluation methods show that using part-based detection in this experimental setting enhances the overall system performance. Figure 6(c) shows the frame-based evaluation of the AFS with the geometry constrains blocks only on the Caltech pedestrians dataset [7], for medium size pedestrians, using the standard t = 0.5 overlap criterion. Our method is ranked second best among the compared methods. This result was previously published in [17] where additional Caltech dataset results can be found. The on-vehicle rv-AFS system runs in real-time on an i7 laptop processing between 4 and 12 frames per second depending on the scene complexity. D. Conclusions We presented a system for detecting pedestrians in the rear-view camera tested on a dataset we collected for training and evaluating such systems. The system is unique in using a part-based detection method, allowing it higher robustness for detecting close-by, partially occluded and non-upright pedestrians. In the future we aim at further increasing robustness and eliminating false alarms. We believe that incorporating a more sophisticated tracking algorithm and using motion-based features to complement appearance ones can improve the overall performance. In addition we aim at improving our appearance based detector, to specifically address the difficult challenge of children detection. R EFERENCES [1] Yaniv Alon and Aharon Bar-Hillel. Off-vehicle evaluation of camerabased pedestrian detection. In Intelligent Vehicles Symposium, pages 352–358. IEEE, 2012. [2] Aharon Bar-Hillel, Dan Levi, Eyal Krupka, and Chen Goldberg. Partbased feature synthesis for human detection. In European Conference on Computer Vision, volume 6314, pages 127–142. Springer-Verlag, 2010.

[3] A. Broggi, M. Bertozzi, A. Fascioli, and M. Sechi. Shape-based pedestrian detection. In in Procs. IEEE Intelligent Vehicles Symposium 2000, pages 215–220, 2000. [4] Hyunggi Cho, Paul E. Rybski, Aharon Bar-Hillel, and Wende Zhang. Real-time pedestrian detection with deformable part models. In Intelligent Vehicles Symposium, pages 1035–1042. IEEE, 2012. [5] Navneet Dalal and Bill Triggs. Histograms of oriented gradients for human detection. In IEEE Conference on Computer Vision and Pattern Recognition, pages 886–893, 2005. [6] Thomas Dean, Mark Ruzon, Mark Segal, Jonathon Shlens, Sudheendra Vijayanarasimhan, and Jay Yagnik. Fast, accurate detection of 100,000 object classes on a single machine. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2013. [7] P. Doll´ar, C. Wojek, B. Schiele, and P. Perona. Pedestrian detection: An evaluation of the state of the art. IEEE Transactions on Pattern Analysis and Machine Intelligence, 99, 2011. [8] Charles Dubout and Francois Fleuret. Exact acceleration of linear object detectors. In Proceedings of the European Conference on Computer Vision, 2012. [9] Markus Enzweiler and Dariu M. Gavrila. Monocular pedestrian detection: Survey and experiments. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(12):2179–2195, 2009. [10] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The PASCAL Visual Object Classes Challenge 2009 (VOC2009) Results. http://www.pascalnetwork.org/challenges/VOC/voc2009/workshop/index.html. [11] Pedro Felzenszwalb, Ross Girshick, David McAllester, and Deva Ramanan. Object detection with discriminatively trained part-based models. In IEEE Transactions on Pattern Analysis and Machine Intelligence, 2009. [12] Pedro F. Felzenszwalb, Ross B. Girshick, and David A. McAllester. Cascade object detection with deformable part models. In IEEE Conference on Computer Vision and Pattern Recognition, pages 2241– 2248. IEEE, 2010. [13] Pedro F. Felzenszwalb, Ross B. Girshick, David A. McAllester, and Deva Ramanan. Object detection with discriminatively trained part-

[14] [15] [16]

[17]

[18] [19] [20] [21]

[22]

[23]

[24]

based models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 32(9):1627–1645, 2010. R. Fergus, P. Perona, and A. Zisserman. Object class recognition by unsupervised scale invariant learning. In CVPR, 2003. Dariu M. Gavrila and Stefan Munder. Multi-cue pedestrian detection and tracking from a moving vehicle. International Journal of Computer Vision, 73(1):41–59, 2007. David Geronimo, Antonio M Lopez, Angel Domingo Sappa, and Thorsten Graf. Survey of pedestrian detection for advanced driver assistance systems. IEEE Transactions on Pattern Analysis and Machine Intelligence, 32(7):1239–1258, 2010. Dan Levi, Shai Silberstein, and Aharon Bar-Hillel. Fast multiplepart based object detection using kd-ferns. In IEEE Conference on Computer Vision and Pattern Recognition, pages 947–954. IEEE, 2013. David G. Lowe. Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 60:91–110, 2004. Elizabeth N. Mazzae and W. Riley Garrott. Experimental evaluation of the performance of available backover prevention technologies. National Highway Traffic Safety Administration, 2006. Constantine Papageorgiou and Tomaso Poggio. A trainable system for object detection. International Journal of Computer Vision, 38(1):15– 33, 2000. Antonio Prioletti, Andreas Mogelmose, Paolo Grisleri, Mohan M. Trivedi, Alberto Broggi, and Thomas B. Moeslund. Part-based pedestrian detection and feature-based tracking for driver assistance: Real-time, robust algorithms, and evaluation. IEEE Transactions on Intelligent Transportation Systems, 14(3):1346–1359, 2013. Amnon Shashua, Yoram Gdalyahu, and Gaby Hayun. Pedestrian detection for driving assistance systems: Single-frame classification and system level performance. In intelligent vehicles symposium, pages 1–6, 2004. D. Tsishkou and S. Bougnoux. Experimental evaluation of multi-cue monocular pedestrian detection system using built-in rear view camera. In Telecommunications, 2007. ITST ’07. 7th International Conference on ITS, pages 1–6, 2007. www.mobileye.com.

Vision-Based Pedestrian Detection for Rear-View ...

object detection tasks and improve benchmark results [10],. [7]. As illustrated in .... system running on the laptop is composed of several algo- rithmic blocks as ...

4MB Sizes 1 Downloads 289 Views

Recommend Documents

Detection of Sudden Pedestrian Crossings for Driving ...
IJRIT International Journal of Research in Information Technology, Volume 1, Issue 11, ... VNR Vignana Jyothi Institute of Technology, Hyderabad, India.

Detection of Sudden Pedestrian Crossings for Driving ... - IJRIT
First, USB camera video images are collected by the embedded Linux system, processed, ... Keywords: Video Capture, S3C2440A, Embedded Linux, DC Motor.

Multispectral Pedestrian Detection: Benchmark Dataset ...
Multispectral Pedestrian Detection: Benchmark Dataset and Baseline. Soonmin Hwang, Jaesik Park, Namil Kim, Yukyung Choi, In So Kweon. Korea Advanced Institute of Science and Technology (KAIST), Republic of Korea. Figure 1: Examples of proposed multis

Detection of Sudden Pedestrian Crossings for Driving ... - IJRIT
a graphical user interface and also used for developing non-GUI programs such as command-line tools and consoles ... access, XML parsing; thread management, network support, and a unified cross- ... projects and jump-start research by providing them

Real-time Pedestrian Detection with Deformable Part Models
of a pedestrian model to detect small pixel-sized pedestrians normally missed by ..... processing the image data that consists of first creating a densely sampled ...

Real-Time Pedestrian Detection With Deep ... - Research at Google
More specifically, we employ the soft-cascade from Benenson et al. [4] ..... ments. IEEE Trans. on Pattern Analysis and Machine Intelligence (T-PAMI), 2009.

Real-Time Pedestrian Detection With Deep Network ...
We present a new real-time approach to object detection that exploits the efficiency of cascade .... Figure 2: The architecture of the baseline deep network for pedestrian detection. ..... Histograms of oriented gradients for human detection.

Fast Pedestrian Detection Using a Cascade of Boosted ...
on pedestrian detection using state-of-the-art locally extracted fea- tures (e.g. ... meaningful features if there is a large variation in object's ap- pearance .... The final strong classifier can be .... simple nonpedestrian patterns in the early s

Pedestrian Detection with a Large-Field-Of-View ... - Semantic Scholar
miss rate on the Caltech Pedestrian Detection Benchmark. ... deep learning methods have become the top performing ..... not to, in the interest of speed.

Cheap Wireless Uhf Transmitter For Tv ⁄ Car Rearview Camera ...
Cheap Wireless Uhf Transmitter For Tv ⁄ Car Rearview ... Cctv Transmission Free Shipping & Wholesale Price.pdf. Cheap Wireless Uhf Transmitter For Tv ⁄ Car ...

Rearview-Mirror-A-Memoir.pdf
Free Books, no matter. whether Rearview Mirror (A Memoir) PDF eBooks or in other format, are accessible in a heap on the net. Rearview Mirror (A Memoir).

Rearview-Mirror-A-Memoir.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Rearview-Mirror-A-Memoir.pdf. Rearview-Mirror-A-Memoir.pdf. Open. Extract. Open with. Sign In. Main menu.

A Predictive Collision Avoidance Model for Pedestrian ... - Springer Link
Abstract. We present a new local method for collision avoidance that is based on collision prediction. In our model, each pedestrian predicts pos- sible future collisions with other pedestrians and then makes an efficient move to avoid them. Experime

A Predictive Collision Avoidance Model for Pedestrian ...
A Predictive Collision Avoidance Model for. Pedestrian Simulation. Ioannis Karamouzas, Peter Heil, Pascal van Beek, and Mark H. Overmars. Center for ...

Pedestrian Advisory Council -
and walkable city for people of all abilities;. WHEREAS, the City of Austin Sidewalk Master Plan / ADA Transition Plan promises at least $40 million per year.

pedestrian short story.pdf
stood silent where two main highways crossed the. town. During the day it was a thunderous surge of. cars, the gas stations .... small Midwestern town. He has also written poetry,. screenplays, and stage plays. Page 2 of 2. pedestrian short story.pdf

The Pedestrian - Ray Bradbury.pdf
screenplays, and stage plays. Page 2 of 2. The Pedestrian - Ray Bradbury.pdf. The Pedestrian - Ray Bradbury.pdf. Open. Extract. Open with. Sign In. Main menu.

Verrazano Pedestrian/Bicycle Access: Planning / Design Feasibility
Pedestrian/Bigde Aaess: Pamming/Design Faasibility. Rudolph W. Giuliani, Mayor. City ºf New Yºrk. Joseph B. Rose, Director. Department ºf City Planning.

Working Group Pedestrian Safety Materials.pdf
District could accommodate pedestrians in these areas by taking the parking lane to create. more room for pedestrians and lower the rates in surrounding parking garages. Additionally, the District could provide free Circulator service to support tran

Cheap Fulcol For Iphone 6 5S Samsung Rearview Mirror Selfie Stick ...
Cheap Fulcol For Iphone 6 5S Samsung Rearview Mirror ... Palo Selfie Stick Free Shipping & Wholesale Price.pdf. Cheap Fulcol For Iphone 6 5S Samsung ...

Statewide Bicycle Pedestrian Plan -
Bicycle and Pedestrian Statewide Plan, and we'd like your input. This Plan will be used to help determine future investments in bicycle and pedestrian projects ...