Show simple item record

dc.contributor.authorHu, Gang
dc.date.accessioned2012-07-25T12:46:17Z
dc.date.available2012-07-25T12:46:17Z
dc.date.issued2012-07-25
dc.identifier.urihttp://hdl.handle.net/10222/15095
dc.descriptionPhD Thesisen_US
dc.description.abstractCurrent developments of hardware devices have allowed the computer vision technologies to analyze complex human activities in real time. High quality computer algorithms for human activity interpretation are required by many emerging applications, such as patient behavior analysis, surveillance, gesture control video games, and other human computer interface systems. Despite great efforts that have been made in the past decades, it is still a challenging task to provide a generic gesture recognition solution that can facilitate the developments of different gesture-based applications. Human vision is able to perceive scenes continuously, recognize objects and grasp motion semantics effortlessly. Neuroscientists and psychologists have tried to understand and explain how exactly the visual system works. Some theories/hypotheses on visual perception such as the visual attention and the Gestalt Laws of perceptual organization (PO) have been established and shed some light on understanding fundamental mechanisms of human visual perception. In this dissertation, inspired by those visual attention models, we attempt to model and integrate important visual perception discoveries into a generic gesture recognition framework, which is the fundamental component of full-tier human activity understanding tasks. Our approach handles challenging tasks by: (1) organizing the complex visual information into a hierarchical structure including low-level feature, object (human body), and 4D spatiotemporal layers; 2) extracting bottom-up shape-based visual salience entities at each layer according to PO grouping laws; 3) building shape-based hierarchical salience maps in favor of high-level tasks for visual feature selection by manipulating attention conditions of the top-down knowledge about gestures and body structures; and 4) modeling gesture representations by a set of perceptual gesture salience entities (PGSEs) that provide qualitative gesture descriptions in 4D space for recognition tasks. Unlike other existing approaches, our gesture representation method encodes both extrinsic and intrinsic properties and reflects the way humans perceive the visual world so as to reduce the semantic gaps. Experimental results show our approach outperforms the others and has great potential in real-time applications.en_US
dc.language.isoenen_US
dc.subjectVisual attentionen_US
dc.subjectPerceptual organizationen_US
dc.subjectSalience mapen_US
dc.subject3D Cameraen_US
dc.subjectPerceptual gesture salience entityen_US
dc.subjectGeneric edge tokenen_US
dc.subjectCurve partition pointen_US
dc.subjectGesture recognitionen_US
dc.titleA Generic Gesture Recognition Approach based on Visual Perceptionen_US
dc.date.defence2012-06-22
dc.contributor.departmentFaculty of Computer Scienceen_US
dc.contributor.degreeDoctor of Philosophyen_US
dc.contributor.external-examinerDr. Minglun Gongen_US
dc.contributor.graduate-coordinatorDr. Qigang Gaoen_US
dc.contributor.thesis-readerDr. Evangelos E. Miliosen_US
dc.contributor.thesis-readerDr. Norman Scrimgeren_US
dc.contributor.thesis-readerDr. Stephen Brooksen_US
dc.contributor.thesis-supervisorDr. Qigang Gaoen_US
dc.contributor.ethics-approvalNot Applicableen_US
dc.contributor.manuscriptsNot Applicableen_US
dc.contributor.copyright-releaseNot Applicableen_US
 Find Full text

Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record