Be the first to review this item and earn 25 Rakuten Super Points™
This book deals with the creation of the algorithm backbone that can enable a computer to perceive humans in some monitored space by processing the same signals that humans use in order to perform the same task e.g. audio and video. To do so, computers use sensors and algorithms to detect and tract multiple interacting humans, their faces and hands or their voices. This application domain is challenging, because audio and visual signals are cluttered both by background and by foreground audiovisual objects. After establishing particle filtering as the framework for tracking then audio, visual and audiovisual tracking is detailed. Each modality is analyzed starting with sensor configurations, detection for tracker initialization and the trackers themselves. Technique to fuse the modalities are then considered. Instead of offering a monolithic approach to the tracking problem, the book also focuses on implementation issues by providing MATLAB code for every presented component. This way the reader connects every presented concept with a corresponding piece of code that follows immediately after the theory. Finally, the applications of the tracking systems in different application domains are considered.