Video tracking: Difference between revisions

Content deleted Content added
Reverted 1 edit by Stephenli2000 (talk): Rv linkspam (TW)
Citation bot (talk | contribs)
Add: publisher. | Use this bot. Report bugs. | Suggested by Headbomb | Linked from Wikipedia:WikiProject_Academic_Journals/Journals_cited_by_Wikipedia/Sandbox2 | #UCB_webform_linked 1971/2044
 
(13 intermediate revisions by 10 users not shown)
Line 1:
{{distinguish|Camera tracking}}
 
'''Video tracking''' is the process of locating a [[Motion (physics)|moving]] object (or multiple objects) over time using a camera. It has a variety of uses, some of which are: human-computer interaction, security and surveillance, video communication and [[video compression|compression]], [[augmented reality]], traffic control, medical imaging<ref>{{cite journal|author1=Peter Mountney, Danail Stoyanov |author2=Guang-Zhong Yang |lastauthorampname-list-style=yesamp |title=Three-Dimensional Tissue Deformation Recovery and Tracking: Introducing techniques based on laparoscopic or endoscopic images." IEEE Signal Processing Magazine. 2010 July. Volume: 27|issue= 4|pages=14–24|doi=10.1109/MSP.2010.936728|year=2010|journal=IEEE Signal Processing Magazine|volume=27 |hdl=10044/1/53740 |s2cid=14009451 |url=http://spiral.imperial.ac.uk/bitstream/10044/1/53740/2/Three-Dimensional%20Tissue%20Deformation%20Recovery%20and%20Tracking_AuthorsVersion.pdf|hdl-access=free}}</ref> and [[video editing]].<ref>{{cite book |author=Lyudmila Mihaylova, Paul Brasnett, Nishan Canagarajan and David Bull |title=Object Tracking by Particle Filtering Techniques in Video Sequences; In: Advances and Challenges in Multisensor Data and Information |citeseerx=10.1.1.60.8510 |year=2007 |series=NATO Security Through Science Series, 8 |publisher=IOS Press |place= Netherlands |pages=260–268 |isbn= 978-1-58603-727-7 |url=http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.60.8510&rep=rep1&type=pdf}}</ref><ref>{{cite book |chapter-url=http://www.hitl.washington.edu/research/artoolkit/Papers/IWAR99.kato.pdf |doi=10.1109/IWAR.1999.803809 |isbn=0-7695-0359-4 |chapter=Marker tracking and HMD calibration for a video-based augmented reality conferencing system |title=Proceedings 2nd IEEE and ACM International Workshop on Augmented Reality (IWAR'99) |pages=85–94 |year=1999 |last1=Kato |first1=H. |last2=Billinghurst |first2=M. |s2cid=8192877 }}</ref> Video tracking can be a time-consuming process due to the amount of data that is contained in video. Adding further to the complexity is the possible need to use [[object recognition]] techniques for tracking, a challenging problem in its own right.
 
==Objective==
[[File:High-speed catching system.webm|thumb|An example of [[visual servoing]] for the robot hand to catch a ball by object tracking with visual feedback that is processed by a high-speed image processing system.<ref>{{cite web|title=High-speed Catching System (exhibited in National Museum of Emerging Science and Innovation since 2005)|url=http://www.k2.t.u-tokyo.ac.jp/fusion/MiraikanCatching/index-e.html|website=Ishikawa Watanabe Laboratory, University of Tokyo|accessdateaccess-date=12 February 2015}}</ref><ref>{{cite web|title=Basic Concept and Technical Terms|url=http://www.k2.t.u-tokyo.ac.jp/tech_terms/index-e.html#VisualFeedback|website=Ishikawa Watanabe Laboratory, University of Tokyo|accessdateaccess-date=12 February 2015}}</ref>]]
 
The objective of video tracking is to associate target objects in consecutive video frames. The association can be especially difficult when the objects are moving fast relative to the [[frame rate]]. Another situation that increases the complexity of the problem is when the tracked object changes orientation over time. For these situations video tracking systems usually employ a motion model which describes how the image of the target might change for different possible motions of the object.
Line 15:
 
==Algorithms==
[[File:Samples_of_object_co-segmentation.jpg|thumb|[[Object co-segmentation|Co-segmentation]] of objects in video frames]]
To perform video tracking an algorithm analyzes sequential [[video frame]]s and outputs the movement of targets between the frames. There are a variety of algorithms, each having strengths and weaknesses. Considering the intended use is important when choosing which algorithm to use. There are two major components of a visual tracking system: target representation and localization, as well as filtering and data association.
 
''Target representation and localization'' is mostly a bottom-up process. These methods give a variety of tools for identifying the moving object. Locating and tracking the target object successfully is dependent on the algorithm. For example, using blob tracking is useful for identifying human movement because a person's profile changes dynamically.<ref>{{cite journal |author1=S. Kang |author2=J. Paik |author3=A. Koschan |author4=B. Abidi |author5=M. A. Abidi |lasteditor-authorfirst1=Kenneth W |editor-ampfirst2=yesFabrice |editor-last1=Tobin, Jr |editor-last2=Meriaudeau |name-list-style=amp |title=Real-time video tracking using PTZ cameras |journal=Proc. SPIE |volume= 5132 |pages=103–111 |year=2003 |citeseerx=10.1.1.101.4242 |doi=10.1117/12.514945 |series=Sixth International Conference on Quality Control by Artificial Vision |bibcode=2003SPIE.5132..103K |s2cid=12298526 }}</ref> Typically the computational complexity for these algorithms is low. The following are some common ''target representation and localization'' algorithms:
 
* '''Kernel-based tracking''' ([[mean-shift]] tracking<ref>Comaniciu, D.; Ramesh, V.; Meer, P., "[http://imagine.enpc.fr/~de-la-gm/cours/UPEM/projects/Real-Time%20Tracking%20of%20Non-Rigid%20Objects%20using%20Mean%20Shift.pdf Real-time tracking of non-rigid objects using mean shift]," Computer Vision and Pattern Recognition, 2000. Proceedings. IEEE Conference on, vol.2, no., pp. 142, 149 vol.2, 2000</ref>): an iterative localization procedure based on the maximization of a [[similarity measure]] ([[Bhattacharyya coefficient]]).
* '''Contour tracking''': detection of object boundary (e.g. active contours or [[Condensation algorithm]]). Contour tracking methods iteratively evolve an initial contour initialized from the previous frame to its new position in the current frame. This approach to contour tracking directly evolves the contour by minimizing the contour energy using gradient descent.
 
''Filtering and data association'' is mostly a top-down process, which involves incorporating prior information about the scene or object, dealing with object dynamics, and evaluation of different hypotheses. These methods allow the tracking of complex objects along with more complex object interaction like tracking objects moving behind obstructions.<ref>{{cite journal |author=Black, James, Tim Ellis, and Paul Rosin |title=A Novel Method for Video Tracking Performance Evaluation |citeseerx=10.1.1.10.3365|journal=Joint IEEE Int. Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance |year=2003 |pages= 125–132}}</ref> Additionally the complexity is increased if the video tracker (also named TV tracker or target tracker) is not mounted on rigid foundation (on-shore) but on a moving ship (off-shore), where typically an inertial measurement system is used to pre-stabilize the video tracker to reduce the required dynamics and bandwidth of the camera system.<ref>[httphttps://www.targettrackingimar-navigation.de/index.phpde/enprodukte-uebersicht/applications/trackerproduct-systemsoverview-forby-product/item/iipsc-tr-target-trackingtracker-iipsc.htmlrange-control Gyro Stabilized Target Tracker for Off-shore Installation]</ref>
The computational complexity for these algorithms is usually much higher. The following are some common filtering algorithms:
* [[Kalman filter]]: an optimal recursive Bayesian filter for linear functions subjected to Gaussian noise. It is an algorithm that uses a series of measurements observed over time, containing noise (random variations) and other inaccuracies, and produces estimates of unknown variables that tend to be more precise than those based on a single measurement alone.<ref>{{cite journal |author1=M. Arulampalam |author2=S. Maskell |author3=N. Gordon |author4=T. Clapp |lastname-authorlist-ampstyle=yesamp |title=A Tutorial on Particle Filters for Online Nonlinear/Non-Gaussian Bayesian Tracking |journal=IEEE Transactions on Signal Processing |volume= 50 |issue= 2 |year=2002 |doi=10.1109/78.978374 | citeseerx=10.1.1.117.1144 |pages=174 |bibcode=2002ITSP...50..174A |s2cid=55577025 }}</ref>
* [[Particle filter]]: useful for sampling the underlying state-space distribution of nonlinear and non-Gaussian processes.<ref name="Video Tracking">
{{cite book
Line 30 ⟶ 31:
| volume = 1
| year= 2010
|publisher=Addison-Wesley Professional | url= https://books.google.com/books?id=56LNfE2QGtYC&pg=PA50&dq=rhythms&pg=PA50
| quote = Video Tracking provides a comprehensive treatment of the fundamental aspects of algorithm and application development for the task of estimating, over time.
|isbn=9780132702348 }}
Line 39 ⟶ 40:
| volume = 1
| year= 2010
| url= https://books.google.com/books?id=Ws0JThymM-EC
| quote = Background subtraction is the process by which we segment moving regions in image sequences.
| isbn = 9780549524892
Line 59 ⟶ 60:
==External links==
*[https://www.youtube.com/watch?v=2y5oVHNfbf8&list=FL8KomHNN46yntJPaerp06uA&feature=mh_lolz – Interesting historical example (1980)] of [[Cromemco Cyclops]] Camera used to track a ball going through a maze.
 
{{Computer vision}}
 
[[Category:Motion in computer vision]]
Line 64 ⟶ 67:
[[Category:Tracking]]
[[Category:Articles containing video clips]]
 
[[fr:Tracking]]