Please use this identifier to cite or link to this item:
http://dspace.dtu.ac.in:8080/jspui/handle/repository/15590
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | JAIN, KANIKA | - |
dc.date.accessioned | 2017-02-09T10:19:10Z | - |
dc.date.available | 2017-02-09T10:19:10Z | - |
dc.date.issued | 2015-07 | - |
dc.identifier.uri | http://dspace.dtu.ac.in:8080/jspui/handle/repository/15590 | - |
dc.description.abstract | Object tracking has always been a very interesting and important research area in the field of Computer Vision and Artificial Intelligence. Tracking an object of interest is an application that can benefit from multiple sensing modalities. If the object of interest emits sound then information from both audio and video sensors can be fused together to remove effects of clutter and background noise. Therefore, the use of same visual and audio interface modalities that humans take for granted can make indoor spaces more intelligent. Audio and visual modalities complement each other when background noise impairs a single modality. This work presents a new approach for modeling and processing data from audio and visual sensors for tracking multiple objects simultaneously. This approach is based on graphical model for visual data and Time Delay of Arrival (TDOA) analysis for sound cue. Then both the cues are modeled by a data likelihood function. Finally, Particle Filtering for multiple target tracking and Dezert-Smarandache theory (DSmT) for fusing the information provided by audio-visual cues are combined. For modeling the visual cue, initially dominant motion is detected from the video frames. Then some dominant motion points are selected for depicting movements of target object. When some occlusion occurs, these motion points estimate the object position from a graphical model. As for modeling the sound cue, the Time Delay of Arrival (TDOA) which occurs between the two audio signals received by the two different microphones kept a fixed distance apart from each other provides an indication of the position of the sound source(s) relative to the microphone pair. This provides an estimate of the horizontal position of object in the image. | en_US |
dc.language.iso | en | en_US |
dc.relation.ispartofseries | TD NO.1895; | - |
dc.subject | TRACKING MULTIPLE OBJECTS | en_US |
dc.subject | MULTIPLE SENSING MODALITIES | en_US |
dc.subject | DEZERT-SMARANDACHE THEORY | en_US |
dc.subject | PARTICLE FILTERING | en_US |
dc.title | MULTI MODEL TRACKING OF MULTIPLE OBJECTS BASED ON SPEECH | en_US |
dc.type | Thesis | en_US |
Appears in Collections: | M.E./M.Tech. Electronics & Communication Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
KanikaJain_thesis_2k13_spd_07 _actual.pdf | 1.42 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.