Tracking multiple moving objects in moving camera images

Number of pages: 106 File Format: word File Code: 31337
Year: 2013 University Degree: Master's degree Category: Electrical Engineering
  • Part of the Content
  • Contents & Resources
  • Summary of Tracking multiple moving objects in moving camera images

    Master thesis in the field of electrical engineering (communication-system)

    Abstract

     

    Tracking multiple moving objects in moving camera images

     

     

     

    Target tracking is one of the most important issues in machine vision science. The problem of tracking multiple targets in images obtained from a moving camera has been evaluated in this thesis. To solve this problem, two general methods are presented in this thesis. The first method is mainly based on block matching algorithms. In this method, by using motion compensation and categorizing different motions in the image, the target and background areas will be determined. Then the tracking operation has been done using the Kalman filter. According to the obtained results, this method has the ability to track targets with high accuracy in most situations. The second proposed method is basically based on the well-known point matching algorithm called SIFT. In this method, by making a match between the key points of the target area and the corresponding points in the next frame, an attempt is made to find the target area in successive frames. The problem of tracking the target areas in the frame sequence is done by Kalman filter. Due to the fact that the target is checked locally in this method, this method has great ability in different tracking situations such as fading. According to the graphs and tables, it can be concluded that this method has a much stronger performance than most existing methods in terms of accuracy in many cases of tracking. 

    1-1- Introduction

    Today, scientific advances have deeply affected human life. At the same time, with the introduction of technology into the personal life of society, the presence of devices and tools that play the role of the interface between man and machine is felt more and more day by day. An example of these devices are cameras[1]. The widespread use of these tools in today's societies, especially in more industrialized countries, is undeniable. Therefore, today, improving the quality and facilities of cameras is considered as an important factor in increasing their efficiency. One of the most important branches of science that investigates these issues is called machine vision[2].

    One of the main goals of machine vision is to make cameras intelligent in order to use them in surveillance systems[3], commercial, military and other applications. For this purpose, extensive studies have been conducted in order to create new methods of intelligentization and also to improve existing methods. Most of these studies are focused on the detection [4] and tracking [5] of targets [6]. The general purpose of conducting such studies is to reduce the volume of calculations and increase the accuracy in the stages of detection and tracking. In general, target detection means detecting a region of the image that can be considered as a candidate [7] for the target region. For example: specifying the areas of the image that are related to the license plate, or also revealing the areas of the image that can be considered as areas related to the human face. Also, the purpose of tracking is to specify the desired area in the set of consecutive frames. In this way, the general path of the target will be determined in a sequence of time during successive frames. In the following discussion, we will briefly introduce various tracking systems and their components as well as their operation. 1-1-1 Structure of tracking systems Different tracking systems are divided into different categories based on their applications. Cameras and targets are the main components of such systems. Therefore, just as these components play a decisive role in the type of tracking systems, they are also very important in determining the type of methods used in these systems. The systems have significant differences based on the number, type, and other conditions of the cameras and objectives. Likewise, these differences can be seen in the tracking methods used in them. In the following, we will discuss the factors that are effective in creating these changes.

    1-1-1-1-          Camera

    As the main component of the tracking system, the camera has the task of creating a sequence of frames over time.The type of cameras used, the number and the way they are placed will play a big role in determining the appearance of the frames. This effect is sometimes to such an extent that it creates methods with a different foundation.

    For example, the tracking methods in systems with visible cameras [8] are completely different from the methods used in systems with infrared cameras [9]. This problem comes from the fact that infrared cameras have a kind of prior information [10] about the targets. This means that in the images obtained from these cameras, the targets have stronger color intensity [11] than their surroundings. As a result, they have more detection capabilities. Although some of the presented algorithms can be applied to the images obtained from both visible and infrared cameras, the efficiency of these algorithms in these images is significantly different.

    In addition, the number of cameras used is also one of the most important factors in determining the method used in tracking. The difference in the viewing angle of the cameras creates different images from different angles of a certain scene. In this situation, it is necessary to find the corresponding points in the frames obtained from all cameras and also to calibrate [12] the cameras. It can be seen that these methods are generally different from tracking methods based on a single camera.

    In addition to the above, camera motion[13] should also be considered in some cases. This means that sometimes, in addition to the targets, the camera also has movement. In these cases, the components in the consecutive frames move relative to each other. This is despite the fact that some of these movements are due to the movement of the camera and some are also due to the presence of movement in objects. Therefore, the ultimate goal is to distinguish between the movements that are caused by the camera and the movements that are real. The need to perform this action is one of the main things that is not considered in tracking targets in fixed camera systems. The type of targets, their number as well as their appearance change in the sequence of frames are the determining factors in choosing the appropriate tracking method for the system. According to these differences in the structure of targets, different methods have been created.

    In some cases of tracking, the target has special appearance characteristics. For example, sometimes the tracking of the human face is considered. The human face has a certain dominant[14] consisting of eyes, nose and lips in a pink oval shape. In detecting these cases, the algorithm only looks for areas of the image that have the ability to match the desired dominant for the human face. In this way, as it can be seen, the presented method is specific to tracking in this case and is different from other tracking methods in general.  

    The number of targets in the image is also one of the important and determining factors in the tracking algorithms. With the increase in the number of targets, new and major problems arise in the fields of detection and tracking. One of the most important of these problems is the fading of targets [15] and the placement of targets in a similar position [16]. Algorithms presented by focusing on these problems have provided ways to solve them.

    Change of appearance [17] of the target is one of the major problems related to its appearance conditions. This means that sometimes goals do not have a rigid structure[18]. Therefore, they look different in successive frames. For example, the appearance of a pedestrian undergoes various appearance changes in successive frames. These changes are not seen in the appearance of solid objects such as cars. Therefore, the proposed methods should be such that they have the ability to adapt to these changes in the appearance conditions of the targets. 1-1-2 How tracking systems work Although it was mentioned in the previous articles that changes in the structure of tracking systems can lead to major changes in the tracking methods used, these methods have many commonalities in many aspects and basic principles. The main difference between these methods is in how to implement the general steps. In this thesis, we generally divide the algorithms used in target tracking systems into two main categories based on the use of predictability[19].  In the following, we will introduce each of these two categories and state their advantages and disadvantages.

  • Contents & References of Tracking multiple moving objects in moving camera images

    List:

     

     

    Page

    Title

    Chapter 1 Introduction- 1

    Introduction- 2 Structure of tracking systems- 3

    Camera- 3

    Purpose- 5

    How tracking systems work- 6

    Algorithms without predictive properties - 6

    Algorithms with predictive properties - 7

    Definition of the problem and the problems ahead - 8

    How to solve the problem - 10

    Chapter heads - 11

     

    Chapter 2    Review of the conducted research - 14

    Introduction - 15

    Specific methods for fixed camera - 15

    Background subtraction method - 15

    Methods applicable to moving camera - 17

    Mean Shift method - 17

    CAM Shift method - 20

    Visual flow method - 21

    Page

    Title

    Chapter 3 Algorithms proposed for detection - 24

    Introduction - 25

    The first proposed algorithm - 26 Motion compensation by block matching algorithms - 26

    The concept of block matching algorithm - 27

    Algorithms for searching the corresponding block - 29

    Getting the moving image area - 33

    Image segmentation by K-Means algorithm - 34

    Flow diagram of the first proposed algorithm - 37

    Second proposed algorithm - 39

    Building the scale space - 41

    Using LoG approximation - 44

    Finding key points in the image - 46

    Removing ineffective key points - 47

    Harris corner detector - 47

    Removing points with low resolution using Taylor expansion - 51

    Directing selected key points - 53

    Creating SIFT features - 54

    Chapter 4 Tracking by Kalman filter - 56

    Introduction - 57

    Kalman filter - 57

    Target movement type - 61

    Practical use of Kalman filter - 62

    Page

    Title

    Chapter 5    Simulation and comparison - 66

    Introduction - 67

    Sequence of case frames Use - 68

    First frame sequence - 69

    Second frame sequence - 71

    Third frame sequence - 73

    Fourth frame sequence - 75

    Fifth frame sequence - 78

    Chapter 6 Results and suggestions - 82

    Introduction - 83

    Conclusion - 83

    Suggestions - 84

    List of references - 86

    Source:

    List of references

     

    Wang, P. (2010, December). Moving Object Segmentation Algorithm Based on Edge Detection. In Computational Intelligence and Software Engineering (CiSE), 2010 International Conference on (pp. 1-4). IEEE.

    Saravanakumar, S., Vadivel, A., & Saneem Ahmed, C. G. (2010, December). Multiple human object tracking using background subtraction and shadow removal techniques. In Signal and Image Processing (ICSIP), 2010 International Conference on (pp. 79-84). IEEE.

    Zhao, C., & Reid, I. (2010). Mean-Shift Visual Tracking with NP-Windows Density Estimates. In proc. British Machine Vision Conference (BMVC).

    Chen, J., An, G., Zhang, S., & Wu, Z. (2010, March). A mean shift algorithm based on modified Parzen window for small target tracking. In Acoustics Speech and Signal Processing (ICASSP), 2010 IEEE International Conference on (pp. 1166-1169). IEEE.

    Zhao, J., Qiao, W., & Men, G. Z. (2009, July). An approach based on mean shift and Kalman filter for target tracking under occlusion. In Machine Learning and Cybernetics, 2009 International Conference on (Vol. 4, pp. 2058-2062). IEEE.

    Chu, H., Ye, S., Guo, Q., & Liu, X. (2007, August). Object tracking algorithm based on camshift algorithm combining with difference in frame. In Automation and Logistics, 2007 IEEE International Conference on (pp. 51-55). IEEE.

    Wang, A., Li, J., & Lu, Z. Improved Camshift with adaptive searching window. International Journal of Soft Computing, 2012

    Li, D., Li, H., & Wei, X. (2010, October). Kalman filtering-based modified Cam-Shift vehicle tracking algorithm for highway traffic conditions. In Computer Application and System Modeling (ICCASM),In Computer Application and System Modeling (ICCASM), 2010 International Conference on (Vol. 14, pp. V14-271). IEEE.

    Horn, B. K., & Schunck, B. G. (1981). Determining optical flow. Artificial intelligence, 17(1), 185-203.

    Lucas, B. D., & Kanade, T. (1981, August). An iterative image registration technique with an application to stereo vision. In IJCAI (Vol. 81, pp. 674-679). Dos Santos Pinheiro, M. A. (2010). Multi-Object Tracking in Video Sequences (Doctoral dissertation, Universidade do Porto).

    Kesrarat, D., & Patanavijit, V. Tutorial of Motion Estimation Based on Horn-Schunk Optical Flow Algorithm in MATLAB®.

    Jung, Y. K., & Ho, Y. S. (2003, June). Active camera tracking system using affine motion compensation. In Visual Communications and Image Processing 2003 (pp. 1973-1966). International Society for Optics and Photonics.

    Chan, M. H., Yu, Y. B., & Constantinides, A. G. (1990, August). Variable size block matching motion compensation with applications to video coding. In Communications, Speech and Vision, IEE Proceedings I (Vol. 137, No. 4, pp. 205-212). IET.

    Atrianfar, H., Atrianfar, M., & Atrianfar, H. (2008). Temporal Resolution Enhancement of Video Sequences Using Transform Domain. In IPCV (pp. 185-190).

    Ratnottar, J., Joshi, R., & Shrivastav, M. (2012, May). Review towards the Fast Block Matching Algorithms for Video Motion Estimation. In Communication Systems and Network Technologies (CSNT), 2012 International Conference on (pp. 153-156). IEEE.

    Lu, J., & Liou, M. L. (1997). A simple and efficient search algorithm for block-matching motion estimation. Circuits and Systems for Video Technology, IEEE Transactions on, 7(2), 429-433.

    Luo, Y., & Celenk, M. (2008, June). A hybrid block-matching approach to motion estimation with adaptive search area. In Systems, Signals and Image Processing, 2008. IWSSIP 2008. 15th International Conference on (pp. 85-88). IEEE.

    Gonzalez, R. C., Woods, R. E., & Eddins, S. L. (2009). Digital image processing using MATLAB (Vol. 2). Knoxville: Gatesmark Publishing.

    Adams, R., & Bischof, L. (1994). Seeded region growing. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 16(6), 641-647.

    Lowe, D. G. (2004). Distinctive image features from scale-invariant keypoints. International journal of computer vision, 60(2), 91-110.

    Vinukonda, P. (2011). A Study of the Scale-Invariant Feature Transform on a Parallel Pipeline (Doctoral dissertation, Louisiana State University).

    Babaryka, A. (2012). Recognition from collections of local features (Doctoral dissertation, KTH).

    Seo, N., & Schug, D. A. (2007). Image matching using scale invariant feature transform (SIFT). Final project report. University of Maryland, College Park, Maryland.

    Lee, H., Heo, P. G., Suk, J. Y., Yeou, B. Y., & Park, H. (2009). Scale-invariant object tracking method using strong corners in the scale domain. Optical Engineering, 48(1), 017204-017204.

    Matas, J., Chum, O., Urban, M., & Pajdla, T. (2004). Robust wide-baseline stereo from maximally stable extreme regions. Image and vision computing,22(10), 761-767.

    Harris, C., & Stephens, M. (1988, August). A combined corner and edge detector. In Alvey vision conference (Vol. 15, p. 50).

    Matthews, J. (2002). An introduction to edge detection: The sobel edge detector. Generation5: January, 27. Brown, M., & Lowe, D. G. (2002, September). Invariant Features from Interest Point Groups. In BMVC (No. s 1).

    Brown, M., Szeliski, R., & Winder, S. (2005, June). Multi-image matching using multi-scale oriented patches. In Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE Computer Society Conference on (Vol. 1, pp. 510-517). IEEE.

    Cui, Y., Hasler, N., Thorm?hlen, T., & Seidel, H. P. (2009). Scale invariant feature transform with irregular orientation histogram binning. In Image Analysis and Recognition (pp.

Tracking multiple moving objects in moving camera images