MI3: Multi-Intensity Infrared Illumination Video Database Chia-Hsin Chan#1, Hua-Tsung Chen#2, Wen-Chih Teng#3, Chin-Wei Liu#4, Jen-Hui Chuang#5 # Computer Science Department, National Chiao Tung University 1001 University Road, Hsinchu, Taiwan {1terry0201.cs98g, 3tenvunchi}@gmail.com {2huatsung, 5jchuang}@cs.nctu.edu.tw [email protected] Abstract—Vision-based video surveillance systems have gained increasing popularity. However, their functionality is substantially limited under nighttime conditions due to the poor visibility caused by improper illumination. Equipped on night vision cameras, ordinary infrared (IR) illuminators of fixedintensity usually lead to the imaging problem of overexposure (or underexposure) when the object is too close to (or too far from) the camera. To overcome this limitation, we use a novel multiintensity IR illuminator to extend the effective range of distance of camera surveillance, and establish in this paper the MI3 (MultiIntensity Infrared Illumination) database based on such an illuminator. The database contains intensity varying video sequences of several indoor and outdoor scenes. Ground truths including people counting and foreground labelling are provided for different research usages. Performances of related algorithms are tested for demonstration and evaluation. Index Terms—database, infrared, nighttime surveillance, foreground identification, human detection, people counting

I. INTRODUCTION Nighttime surveillance cameras are equipped with infrared (IR) illuminators for night vision enhancement. Benezeth et al. [1] use a far-IR camera to take images under low-light conditions and apply Gaussian mixture model for human object extraction. Dai et al. [2] use joint shape and appearance cues of to detect and track pedestrians from IR imagery. Li et al. [3] present an active near-IR imaging system capable of producing face images of good condition regardless of ambient visible lights. Two statistical learning algorithms based on Linear Discriminant Analysis and AdaBoost are used to build face recognition classifiers. Though existing works of human/object detection and event analysis show good results for daytime cases, the limitations of poor visibility due to improper illumination are often encountered for nighttime surveillance. In order to enhance the nighttime image quality, Cai et al. [4] utilize an object extraction technique and replace the low-quality static parts of the nighttime images with the high-quality counterparts acquired in the daytime. Nevertheless, unnatural mixtures may occur due to the errors in object extraction. Yamasaki et al. [5] produce natural-appearing enhanced images via decomposing video images into two components (luminance and reflectance) and then modifying the luminance component by referring to the daytime background.

978-1-4673-7314-2/15/$31.00 ©2015 IEEE

(a) (b) Fig. 1. Example of fixed-intensity infrared video for a person at (a) far and (b) near distances. The face is hard to recognize under conditions of (a) underexposure and (b) overexposure.

Although those previous works enhanced the image quality of nighttime surveillance videos to some extent, the IR illuminator of fixed intensity confronts the limitation that only the object(s) at certain distance can be clearly observed. An example of an IR video is shown in Fig. 1, wherein a person walks from far to near. One can see that the face is hard to recognize due to the underexposure (or overexposure) problem, when the person is too far from (or close to) the illuminator. Concerning the foregoing motivation and limitation of existing works, a novel multi-intensity infrared (MIIR) illuminator is proposed in [6] to extend the coverage area of a surveillance camera. Based on such an illuminator, we establish in this paper the MI3 (Multi-Intensity Infrared Illumination) database, which contains intensity varying video sequences of several indoor and outdoor scenes. Ground truths including (1) people counting and (2) foreground labelling are provided for different research usages. II. MULTI-INTENSITY IR ILLUMINATOR The main feature of the MIIR illuminator is the periodic change of illumination intensity. Such feature enables the surveillance camera to capture more information, enhancing nighttime video surveillance and overcoming the underexposure/overexposure problem. Example image frames of a multi-intensity infrared video are shown in Fig. 2, wherein foreground image(s) of applausive quality may be obtained from certain illumination level(s). Fig. 3 gives another example of multiple persons standing at different distances from the camera. One can see that features of human faces near to the camera remain clear under a low illumination level in Fig. 3(a). On the other hand, the person far from the camera can be clearly observed with a high illumination level, as shown in Fig. 3(b). There are several preliminary studies in which the MIIR

IEEE VCIP 2015

(a)

(b)

(c)

(d)

Fig. 2. Images captured at the same time instance under low to high illumination levels by using an MIIR illuminator. (e) (f) Fig. 4. Video clips and its daytime scenes in the MI3 database, including (a) Bus, (b) Doorway, (c) Pathway1, (d) Pathway2, (e) Room, and (f) Staircase.

(a) (b) Fig. 3. Example images for objects located at different ranges captured by the camera with an MIIR illuminator. (a) Features of the human face close to the camera can be seen clearly under a low illumination condition. (b) Objects far away can be clearly observed under a high illumination level.

illuminator is used for improving the video surveillance [7-10]. In [7], Lu et al. develop and compare two methods for foreground object detection. The first method clusters video frames according to their intensity levels and applies background modelling separately for each cluster. The second method constructs two background models to record periodic changes of the min/max pixel intensities of a scene to model the backgrounds. In [8], the MIIR illuminator is used to detect license plates in nighttime scenes. First, potential locations of license plates are estimated using the gradient and edge features, and verified by considering the stroke width of the license ID. Then the results of license plate detection obtained from images captured with different illumination levels are integrated into a synthesized high dynamic range (HDR) image for better visualization. In [9], a video summary method is proposed for the MIIR video. First the Gaussian Mixture Modelling (GMM) is applied to isolate the foreground and background for each frame, then the best foreground with the most plausible quality is selected among different intensities within the same time period. Finally, the selected foreground is synthesized with a pre-selected clear background. Therefore the video size is shrunk and high quality contents are preserved for comfortable viewing. Moreover, the MIIR illumination can also improve the surveillance in the Intensive Care Unit (ICU) [10]. Encouraging results from the above research works reveal that the use of such MIIR illuminator may be a promising research direction for nighttime security surveillance. However, there is no database established systematically to facilitate a quantitative analysis and evaluation for nighttime surveillance videos. Therefore, we establish a database of MIIR videos in this paper.

III. MI3: MULTI-INTENSITY INFRARED ILLUMINATION DATABASE In the proposed MI3 database, video sequences captured with the MIIR illuminator for various scenes are provided with ground truths for different applications. Details of the MI3 database are described as follows. A. Video Acquisition and Contents For MIIR video acquisition, we use an ordinary surveillance camera (fixed aperture and shutter speed) equipped with an MIIR illuminator of six illumination levels. The camera is set to record the raw video at 30 frames per second (fps). Since six illumination levels alternate periodically, the frame rate for each illumination level is 30/6 = 5 fps. Each video clip is captured with the first 100 frames contains no foreground objects. The captured videos consist of six scenes, including Bus, Doorway, Pathway1, Pathway2, Room, and Staircase. Each scene contains several test video clips. Typical examples of these scenes in MI3 are presented in Fig. 4, wherein the left images show the original scenes with high ambient light captured by a digital camera in daytime while the right images are the video frames captured as MIIR videos at nighttime. Each scene is introduced as follows. Bus More generally than ever, surveillance cameras are equipped inside buses, trains, and other means of public transportation. Video clips containing people get on/off the bus and take seats with different distance to the camera are provided for this scene. Doorway Doorways are usually the chokepoints where people must pass through to enter a certain area. Anyone entering an area without permission should be found instantly. Video clips containing single/multiple persons go in/out through the door (which means directly facing/back onto the camera) are provided for this scene. Pathway1 Installation of surveillance cameras at outdoor areas is highly desirable to protect people from safety hazard or criminal activities. In this scene there are ambient lights that

may lead to misjudgement of foregrounds. Video clips containing a single person walking through the pathway and multiple persons playing hide and seek are provided. Pathway2 This is another pathway without ambient lights and it may become very dark in the night. Video clips containing several activities including single/multiple person walking through the pathway, walking with flashlight in hand, riding a bicycle, leaving and picking up a bag are provided for this scene. Room Cameras installed in rooms can enable a security surveillance system to quickly identify the person(s) entering the room and issue an alarm if necessary. Different illumination levels provided by the MIIR illuminator will help the examination of persons located at different areas of a room. Video clips containing single/multiple person straying in the room, entering then leaving the room are provided for this scene. Staircase Staircases are the places where security incidents frequently occur. Besides suspicious event/person detection, cameras located at stairways can also help detect safety problems such as falling downs, so that the security personnel can be informed immediately and take prompt actions in a timely fashion. Video clips containing single/multiple person go up/down the staircase (which means directly facing/back onto the camera) are provided for this scene. B. Ground Truths and Evaluation Criteria In the proposed MI3 database, two kinds of ground truths with related evaluation are provided. 1) People Counting: the number of persons in each video clip is manually identified frame by frame, and for each illumination level. The average error of person number for the entire video clip is used as evaluation criterion, as defined by ଵ

‫ ܧ‬ൌ σே ෝప െ ‫݌‬௜ ȁ ௜ୀଵ ȁ‫݌‬ ே

(1)

where ‫݌‬ෝప and ‫݌‬௜ represent the estimated person number and ground truth of the ith frame, respectively, and N is the number of total frames. 2) Foreground Labelling: Given several video frames with only backgrounds as reference, human viewers pixel-wisely mark foreground regions for each testing video frame based on subjective judgement. Since such task is time-consuming and labour-intensive, not all channels of the whole frames are marked. Instead, we take turns marking one channel for each frame, from brightest to darkest (6, 5, 4, 3, 2, 1, 6 …), and all pixels in frames of the darkest illumination level are totally marked as backgrounds. For evaluation, we employ precision and recall, defined as ‫ ݊݋݅ݏ݅ܿ݁ݎ݌‬ൌ 

்௉ ்௉ାி௉

(2)

‫ ݈݈ܽܿ݁ݎ‬ൌ 

்௉ ்௉ାிே

(3)

where true positives (TP), false positives (FP), and false negatives (FN) respectively denote the numbers of correctly judged foreground pixels, incorrect background pixels (misjudged as foregrounds), and incorrect foreground pixels (misjudged as backgrounds). IV. SIMULATION RESULTS To demonstrate the usage of the proposed MI3 database, we apply existing methods of different usages, use the ground truths for evaluations, and try to compare the performance between single/multiple channels. Note that since the illumination level of channel 1 is too low for human observers to observe anything, in this paper channel 1 will not be considered for testing. However, channel 1 is still useful when the foreground object is very close to the camera, for example when a criminal is about to tamper with the camera. For people counting, the Viola-Jones object detection algorithm [11] is employed to detect human upper bodies, each represents a person. Since some persons may be miss-detected in some channels due to improper illumination but can be detected in other channel(s), we select the maximum of detected person numbers among all channels as the detection result for a frame. And since the raw detection number fluctuates through consecutive frames due to miss or wrong detections, the result of each frame is refined by a mean filter of length 15, and then rounded to the nearest integer. Experimental results for several sequences1 are shown in Table I. It is readily observable that the performance of the above method, which jointly considers all channels, is better than applying detection on a single channel. For example, as shown in Fig. 5, all people are correctly detected in channel 2, but only two persons are detected in channel 6 since the overexposure caused by an “over-illumination” condition makes the identification difficult. As for foreground labelling, the modified GMM background modelling approach in [9] is tested and partial results are shown in Table II. GMM achieves good performance of over 90% precision and recall in most cases since it can adapt to background variation and illumination changes. From the observation of the performance among different channels, it is found that brighter channels tend to yield higher recall rates but lower precision rates, and vice versa. That is because when a higher illumination level is used, reflections from surroundings will cause rapid intensity changes, leading to wrong identifications. The precision rates for Pathway2_2 are much lower than other sequences since sometimes the person directs the flashlight toward the camera. Therefore, not only flashlight but also the surroundings will be wrongly identified as foregrounds, resulting in FP identifications, as shown in Fig. 6. V. CONCLUSIONS In this paper, we establish the MI3 (Multi-Intensity Infrared

1

For complete test results, please refer to the dataset website http:ġ//sites.google.com/site/miirsurveillance/.

TABLE II PERFORMANCE OF FOREGROUND LABELLING USING THE MODIFIED GMM IN [9] Sequence

Pathway2_1

(a) (b) Fig. 5. Example showing the identification of persons for (a) channel 2 and (b) channel 6 at frame #484 in Pathway2_3. A yellow rectangle identifies the position of a detected person upper body.

Pathway2_2

Pathway2_3

(a) (b) Fig. 6. (a) Original image and (b) foreground labelling result for a frame in Pathway2_2. White, blue, and red colors stand for TP, FN and FP pixels, respectively. TABLEġI RESULTS OF PEOPLE COUNTING USING THE UPPER BODY DETECTION Sequence

Person

Bus_1

1

Bus_2

Pathway2_1

Pathway2_2

Pathway2_3

3

1

1

5

Channel Ch2 Ch3 Ch4 Ch5 Ch6 Joint Ch2 Ch3 Ch4 Ch5 Ch6 Joint Ch2 Ch3 Ch4 Ch5 Ch6 Joint Ch2 Ch3 Ch4 Ch5 Ch6 Joint Ch2 Ch3 Ch4 Ch5 Ch6 Joint

Avg. Error 0.217 0.113 0.104 0.104 0.132 0.085 1.384 1.240 1.208 1.028 1.144 0.747 0.636 0.575 0.731 0.639 0.381 0.231 0.659 0.665 0.604 0.489 0.225 0.170 3.740 3.621 3.662 3.892 3.683 3.062

Channel Ch2 Ch3 Ch4 Ch5 Ch6 Avg. Ch2 Ch3 Ch4 Ch5 Ch6 Avg. Ch2 Ch3 Ch4 Ch5 Ch6 Avg.

Precision 99.90% 98.99% 94.75% 90.52% 87.43% 93.72% 59.19% 52.64% 38.43% 41.60% 75.21% 50.07% 99.61% 99.16% 97.63% 94.90% 92.21% 96.48%

Recall 91.21% 92.98% 94.37% 94.94% 96.30% 94.06% 96.77% 97.64% 98/46% 97.91% 97.83% 97.74% 89.25% 90.35% 91.62% 93.68% 95.73% 92.19%

ACKNOWLEDGMENT This research is supported in part by MOST-103-2221-E009-134-MY2, MOST-104-2218-E-009-008, MOST-1042622-E-009-009-CC2, and MOST-104-3115-E-009-001 of the Ministry of Science and Technology, Taiwan, R.O.C. REFERENCES [1]

[2]

[3]

[4]

[5]

[6] [7]

[8]

[9]

Illumination) database which contains nighttime surveillance video clips of various scenes obtained with time varying illumination intensity. Some ground truths are provided for [10] different research usages. Experimental results show that MIIR videos can enhance the functionalities of nighttime surveillance. [11] We believe that the proposed MI3 database can contribute to further research and development of innovative techniques in the field of nighttime vision-based video surveillance.

Y. Benezeth, B. Emile, H. Laurent, and C. Rosenberger, “A real time human detection system based on far infrared vision,” Image and Signal Processing, vol. 5099, pp. 76-84, 2008. C. Dai, Y. Zheng, and X. Li, “Pedestrian detection and tracking in infrared imagery using shape and appearance,” Computer Vision and Image Understanding, vol. 2-3, pp. 288-299, 2007. S.-Z. Li, R.-F. Chu, S.-C. Liao, and L. Zhang, “Illumination invariant face recognition using near-infrared images,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 29, issue 4, pp. 627-639, 2007. Y. Cai, K. Huang, T. Tan, and Y. Wang, “Context enhancement of nighttime surveillance by image fusion,” Proc. of 18th IEEE Int. Conf. on Pattern Recognition, vol. 1, pp. 980-983, 2006. A. Yamasaki, H. Takauji, S. Kaneko, T. Kanade, and H. Ohki, “Denighting: enhancement of nighttime images for a surveillance camera,” Proc. of 19th IEEE Int. Conf. on Pattern Recognition, vol. 1, pp. 1-4, 2008. W.-C. Teng, “A new design of IR Illuminator for nighttime surveillance,” MS Thesis, National Chiao Tung Univ., 2010. P.-J. Lu, J.-H. Chuang, and H.-H. Lin, “Intelligent nighttime video surveillance using multi-intensity infrared illuminator,” Proc. of the World Congress on Engineering and Computer Science, vol. 1, 2011. Y.-T. Chen, J.-H. Chuang, W.-C. Teng, H.-H. Lin, and H.-T Chen “Robust license plate detection in nighttime scenes using multiple intensity IR-illuminator,” Proc. of IEEE Int. Symp. on Industrial Electronics, pp. 93-898, 2012. J.-H. Chuang, W.-J. Tsai, C.-H. Chan, W.-C. Teng, and I.-C. Lu, “A novel video summarization method for multi-intensity illuminated infrared videos,” Proc. of IEEE Int. Conf. on Multimedia and Expo, pp. 1-6, 2013. W.-C. Teng and J.-H. Chuang, “Intelligent ICU video display for multiintensity IR illumination,” Proc. of IEEE Int. Conf. on Consumer Electronics - Taiwan 2014, pp. 183-184, 2014. P. Viola and M. Jones, “Rapid object detection using a boosted cascade of simple features,” Proc. of IEEE Int. Conf. on Computer Vision and Pattern Recognition, vol. 1, pp.511-518, 2001.

MI3: Multi-Intensity Infrared Illumination Video Database

Room. Cameras installed in rooms can enable a security surveillance system to quickly identify the person(s) entering the room and issue an alarm if necessary.

355KB Sizes 0 Downloads 94 Views

Recommend Documents

pdf-0753\chapter-03-near-infrared-mid-infrared-and-raman ...
Connect more apps... Try one of the apps below to open or edit this item. pdf-0753\chapter-03-near-infrared-mid-infrared-and-raman-spectroscopy-by-y-pico.pdf.

Perceptual Global Illumination Cancellation in ... - Computer Science
For the examples in this paper, we transform the appear- ance of an existing ... iterative design applications. 2. Related Work ..... On a desktop machine with an.

Perceptual Global Illumination Cancellation in ... - Computer Science
based radiosity framework [GTGB84]. We define Kp to be ... iterative design applications. 2. Related Work ..... On a desktop machine with an. NVIDIA GeForce ...

Illumination, RA 7920 and IRR.pdf
Illumination, RA 7920 and IRR.pdf. Illumination, RA 7920 and IRR.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying Illumination, RA 7920 and ...

ReadPDF Advanced Global Illumination, Second ...
Book Synopsis. This book provides a fundamental understanding of global illumination algorithms. It discusses a broad class of algorithms for realistic image ...

Achieving Color Constancy and Illumination ...
to mathematical transformation that occurs due to changing condition between the ... The input data will involve sequences images under varying lighting .... The figure 2, is a block diagram representation of the Multiscale retinex algorithm.

manual-infrared-thermometer-pce-889b.pdf
Whoops! There was a problem loading this page. Retrying... Whoops! There was a problem loading this page. Retrying... manual-infrared-thermometer-pce-889b.pdf. manual-infrared-thermometer-pce-889b.pdf. Open. Extract. Open with. Sign In. Main menu. Di

Achieving Color Constancy and Illumination ...
My Implementation of Retinex and Available Software: A “True View Imaging Company”, presents a software package which includes the retinex algorithm ...

Evaluating Combinational Illumination Estimation ... - IEEE Xplore
Abstract—Illumination estimation is an important component of color constancy and automatic white balancing. A number of methods of combining illumination ...

MULTI-VIDEO SUMMARIZATION BASED ON VIDEO-MMR
we propose a criterion to select the best combination of parameters for Video-MMR. ... Marginal Relevance can be used to construct multi-document summaries ... is meaningful to compare Video-MMR to human choice. In a video set, 6 videos ...

Influence of the illumination on weak antilocalization in ...
tion based on a self-consistent solution of the Schrödinger and Poisson equations,22 including the charge balance equa- tion and the effect of exchange correlation on the Coulomb interaction23 should be performed based on the assumptions of the origi

LCD device including an illumination device having a polarized light ...
Feb 24, 2000 - crystal display element employed in the notebook type personal computer. ...... iZing sheet 9 on the light-incident side of the liquid crystal.

Video steganography
Jul 11, 2003 - ing, resampling, conversion from digital to analog and back again, etc. ...... 855-884. Pitas et al., “Applying Signatures on Digital Images,” IEEE.

LCD device including an illumination device having a polarized light ...
Jan 30, 1998 - In a so-called “notebook type personal computer” which has come in ... crystal display element employed in the notebook type personal computer. A representative example is shown in. FIG. 5 in case of a super twisted nematic liquid

manual-infrared-thermometer-hi-temp-2400.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item.

In-cabin occupant tracking using a low-cost infrared system
In-cabin occupant tracking using a low-cost infrared system. Abstract – Vehicles in future will be safer and more intelligent, able to make appropriate and ...

Infrared spectroscopic studies of the urethane/ether ...
fax: 34-943-212236. ..... [10] M.M. Coleman, Y. Xu, P.C. Painter, Macromolecules 27. (1994) 127. [11] M.M. Coleman, J.F. Graf, P.C. Painter, Specific Interactions.

CI Series Infrared Sorting Machine Brochure.pdf
CI Series Infrared Sorting Machine Brochure.pdf. CI Series Infrared Sorting Machine Brochure.pdf. Open. Extract. Open with. Sign In. Main menu.