Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Learning to detect, associate, and recognize human actions and surrounding scenes in untrimmed videos

Full metadata record
DC Field Value Language
dc.contributor.authorPark, J.-
dc.contributor.authorLee, J.-
dc.contributor.authorJeon, S.-
dc.contributor.authorKim, S.-
dc.contributor.authorKim, S.-
dc.contributor.authorSohn, K.-
dc.date.accessioned2023-04-21T01:40:24Z-
dc.date.available2023-04-21T01:40:24Z-
dc.date.issued2018-10-
dc.identifier.issn0000-0000-
dc.identifier.urihttps://yscholarhub.yonsei.ac.kr/handle/2021.sw.yonsei/6639-
dc.description.abstractWhile recognizing human actions and surrounding scenes addresses different aspects of video understanding, they have strong correlations that can be used to complement the singular information of each other. In this paper, we propose an approach for joint action and scene recognition that is formulated in an end-to-end learning framework based on temporal attention techniques and the fusion of them. By applying temporal attention modules to the generic feature network, action and scene features are extracted efficiently, and then they are composed to a single feature vector through the proposed fusion module. Our experiments on the CoVieW18 dataset show that our model is able to detect temporal attention with only weak supervision, and remarkably improves multi-task action and scene classification accuracies. © 2018 Association for Computing Machinery.-
dc.format.extent6-
dc.language영어-
dc.language.isoENG-
dc.publisherAssociation for Computing Machinery, Inc-
dc.titleLearning to detect, associate, and recognize human actions and surrounding scenes in untrimmed videos-
dc.typeArticle-
dc.identifier.doi10.1145/3265987.3265989-
dc.identifier.scopusid2-s2.0-85058144673-
dc.identifier.bibliographicCitationCoVieW 2018 - Proceedings of the 1st Workshop and Challenge on Comprehensive Video Understanding in the Wild, co-located with MM 2018, pp 21 - 26-
dc.citation.titleCoVieW 2018 - Proceedings of the 1st Workshop and Challenge on Comprehensive Video Understanding in the Wild, co-located with MM 2018-
dc.citation.startPage21-
dc.citation.endPage26-
dc.type.docTypeConference Paper-
dc.description.isOpenAccessN-
dc.description.journalRegisteredClassother-
dc.subject.keywordPlusSemantics-
dc.subject.keywordPlusAction classifications-
dc.subject.keywordPlusLearning frameworks-
dc.subject.keywordPlusScene classification-
dc.subject.keywordPlusSemantic features-
dc.subject.keywordPlusSingular information-
dc.subject.keywordPlusStrong correlation-
dc.subject.keywordPlusVideo classification-
dc.subject.keywordPlusVideo understanding-
dc.subject.keywordPlusClassification (of information)-
dc.subject.keywordAuthorAction Classification-
dc.subject.keywordAuthorScene Classification-
dc.subject.keywordAuthorSemantic Feature Fusion-
dc.subject.keywordAuthorVideo Classification-
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Engineering > 공과대학 전기전자공학부 > 공과대학 전기전자공학과 > 1. Journal Articles

qrcode

Items in Scholar Hub are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Park, Jungin photo

Park, Jungin
공과대학 전기전자공학과
Read more

Altmetrics

Total Views & Downloads

BROWSE