Detailed Information

Cited 22 time in webofscience Cited 26 time in scopus
Metadata Downloads

Action Recognition in Videos Using Pre-Trained 2D Convolutional Neural Networks

Full metadata record
DC Field Value Language
dc.contributor.authorKim, Jun-Hwa-
dc.contributor.authorWon, Chee Sun-
dc.date.accessioned2023-04-28T00:41:18Z-
dc.date.available2023-04-28T00:41:18Z-
dc.date.issued2020-
dc.identifier.issn2169-3536-
dc.identifier.urihttps://scholarworks.dongguk.edu/handle/sw.dongguk/7167-
dc.description.abstractA pre-trained 2D CNN (Convolutional Neural Network) can be used for the spatial stream in the two-stream CNN structure for videos, treating the representative frame selected from the video as an input. However, the CNN for the temporal stream in the two-stream CNN needs training from scratch using the optical flow frames, which demands expensive computations. In this paper, we propose to adopt a pre-trained 2D CNN for the temporal stream to avoid the optical flow computations. Specifically, three RGB frames selected at three different times in the video sequence are converted into grayscale images and are assigned to three R(red), G(green), and B(blue) channels, respectively, to form a Stacked Grayscale 3-channel Image (SG3I). Then, the pre-trained 2D CNN is fine-tuned by SG3Is for the temporal stream CNN. Therefore, only pre-trained 2D CNNs are used for both spatial and temporal streams. To learn long-range temporal motions in videos, we can use multiple SG3Is by partitioning the video shot into sub-shots and a single SG3I is generated for each sub-shot. Experimental results show that our two-stream CNN with the proposed SG3Is is about 14.6 times faster than the first version of the two-stream CNN with the optical flow, and yet achieves a similar recognition accuracy for UCF-101 and a 5.7% better result for HMDB-51.-
dc.format.extent10-
dc.language영어-
dc.language.isoENG-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleAction Recognition in Videos Using Pre-Trained 2D Convolutional Neural Networks-
dc.typeArticle-
dc.publisher.location미국-
dc.identifier.doi10.1109/ACCESS.2020.2983427-
dc.identifier.scopusid2-s2.0-85083249776-
dc.identifier.wosid000527413100044-
dc.identifier.bibliographicCitationIEEE ACCESS, v.8, pp 60179 - 60188-
dc.citation.titleIEEE ACCESS-
dc.citation.volume8-
dc.citation.startPage60179-
dc.citation.endPage60188-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaTelecommunications-
dc.relation.journalWebOfScienceCategoryComputer Science, Information Systems-
dc.relation.journalWebOfScienceCategoryEngineering, Electrical & Electronic-
dc.relation.journalWebOfScienceCategoryTelecommunications-
dc.subject.keywordAuthorConvolutional neural network (CNN)-
dc.subject.keywordAuthoraction recognition-
dc.subject.keywordAuthorvideo analysis-
dc.subject.keywordAuthortwo-stream convolutional neural networks-
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Engineering > Department of Electronics and Electrical Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE