Open Access System for Information Sharing

Login Library

 

Thesis
Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads
Full metadata record
Files in This Item:
There are no files associated with this item.
DC FieldValueLanguage
dc.contributor.author곽수하en_US
dc.date.accessioned2014-12-01T11:49:16Z-
dc.date.available2014-12-01T11:49:16Z-
dc.date.issued2014en_US
dc.identifier.otherOAK-2014-01699en_US
dc.identifier.urihttp://postech.dcollection.net/jsp/common/DcLoOrgPer.jsp?sItemId=000001677486en_US
dc.identifier.urihttps://oasis.postech.ac.kr/handle/2014.oak/2201-
dc.descriptionDoctoren_US
dc.description.abstractIn these days, a vast amount of videos is recorded and archived every minute and demand on automatic video understanding increases consequently. As a step toward video understanding, in this thesis we focus on composite event detection from video. A composite event is composed of multiple primitive actions, which are arranged in a specific temporal-logical context to attain full meaning. Composite event detection aims to discover high-level interpretation of video through the context as well as to detect primitive actions accurately. The contextual structure of a composite event is called scenario, and we assume that scenario is described manually by domain experts.We first propose a new scenario description method. A scenario description method is important for video event detection since its expressive power determines the range of events to be detected. A set of temporal-logical predicates is defined to represent relationships between primitive actions more fluently. The proposed description method is in a form of regular grammar, which is based on the temporal-logical predicates instead of simple ordering of the original grammar. Consequently, the description method has more expressive power and is easy to describe complex composite events at the same time. More flexible scenarios are required to represent complicated composite events in real videos, but enlarge the search space prohibitively as well. We move to an inference algorithm to detect composite event efficiently and exactly even with the huge search space. To this end, we propose constraint flow, which is a combinatorial state transition machine and equivalent with scenario. Our inference algorithm is based on dynamic programming with the constraint flow. We show that the search space containing the globally optimal solution can be reduced significantly by constraint flow and the compact search space allows an on-line and efficient inference algorithm.Most event detection frameworks including the above assume that every agent in video participate in an event with known role. However, such assumption is invalid in real videos, where it is unknown which agent participates in an event with which role. We finally propose an efficient method to identify participants and their roles jointly. We observe that the role of an agent can be estimated by analyzing actions of the agent. Also, the agent-wise role analysis is much more efficient than event detection. Given the results of the agent-wise role analysis, the joint identification problem is solved efficiently by a two-step optimization. By applying event detector only to the identified participants, composite event detection in real videos could be done more efficiently and accurately than a naive approach that detects events from all possible agent combinations.en_US
dc.languageengen_US
dc.publisher포항공과대학교en_US
dc.rightsBY_NC_NDen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/2.0/kren_US
dc.titleGrammar-Based Event Detection from Videoen_US
dc.typeThesisen_US
dc.contributor.college일반대학원 컴퓨터공학과en_US
dc.date.degree2014- 2en_US
dc.type.docTypeThesis-

qr_code

  • mendeley

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Views & Downloads

Browse