HOME > Detail View

Detail View

Online temporal action localization from untrimmed video stream by foreseeing upcoming frames

Online temporal action localization from untrimmed video stream by foreseeing upcoming frames

Material type
학위논문
Personal Author
윤다혜
Title Statement
Online temporal action localization from untrimmed video stream by foreseeing upcoming frames / 윤다혜
Publication, Distribution, etc
Seoul :   Graduate School, Korea University,   2018  
Physical Medium
vi, 36장 : 천연색삽화, 도표 ; 26 cm
기타형태 저록
Online Temporal Action Localization from Untrimmed Video Stream by Foreseeing Upcoming Frames   (DCOLL211009)000000079838  
학위논문주기
학위논문(석사)-- 고려대학교 대학원, 컴퓨터·전파통신공학과, 2018. 2
학과코드
0510   6D36   1072  
General Note
지도교수: 이성환  
Bibliography, Etc. Note
참고문헌: 장 30-36
이용가능한 다른형태자료
PDF 파일로도 이용가능;   Requires PDF file reader(application/pdf)  
비통제주제어
Online action detection,,
000 00000nam c2200205 c 4500
001 000045932645
005 20230526124030
007 ta
008 171228s2018 ulkad bmAC 000c eng
040 ▼a 211009 ▼c 211009 ▼d 211009
041 0 ▼a eng ▼b kor
085 0 ▼a 0510 ▼2 KDCP
090 ▼a 0510 ▼b 6D36 ▼c 1072
100 1 ▼a 윤다혜
245 1 0 ▼a Online temporal action localization from untrimmed video stream by foreseeing upcoming frames / ▼d 윤다혜
260 ▼a Seoul : ▼b Graduate School, Korea University, ▼c 2018
300 ▼a vi, 36장 : ▼b 천연색삽화, 도표 ; ▼c 26 cm
500 ▼a 지도교수: 이성환
502 0 ▼a 학위논문(석사)-- ▼b 고려대학교 대학원, ▼c 컴퓨터·전파통신공학과, ▼d 2018. 2
504 ▼a 참고문헌: 장 30-36
530 ▼a PDF 파일로도 이용가능; ▼c Requires PDF file reader(application/pdf)
653 ▼a Online action detection
776 0 ▼t Online Temporal Action Localization from Untrimmed Video Stream by Foreseeing Upcoming Frames ▼w (DCOLL211009)000000079838
900 1 0 ▼a 이성환, ▼g 李晟瑍, ▼d 1962-, ▼e 지도교수 ▼0 AUTH(211009)151678
945 ▼a KLPA

Electronic Information

No. Title Service
1
Online temporal action localization from untrimmed video stream by foreseeing upcoming frames (34회 열람)
View PDF Abstract Table of Contents

Holdings Information

No. Location Call Number Accession No. Availability Due Date Make a Reservation Service
No. 1 Location Science & Engineering Library/Stacks(Thesis)/ Call Number 0510 6D36 1072 Accession No. 123058297 Availability Available Due Date Make a Reservation Service B M
No. 2 Location Science & Engineering Library/Stacks(Thesis)/ Call Number 0510 6D36 1072 Accession No. 123058298 Availability Available Due Date Make a Reservation Service B M

Contents information

Abstract

Online temporal action localization from an untrimmed video stream is a difficult problem in computer vision. It is challenging because i) in an untrimmed video stream, more than one action instance may appear, including background scenes, and ii) in online settings, only past and current information is available. Therefore, temporal priors, such as the average action duration of training data, which have been exploited by previous action detections methods, are not suitable for this task because of the high intra-class variation in human actions.

We propose a novel online action detection framework that considers actions as a set of temporally ordered subclasses and leverages a future frame generation network to cope with the limited information issue associated with problem outlined above. Additionally, we augment our data by varying the lengths of videos to allow the proposed method to learn about the high intra-class variation in human actions. We evaluate our method using the THUMOS'14 dataset for an online action detection scenario and demonstrate that the performance is comparable to state-of-the-art methods that have been proposed for offline settings.

Table of Contents

1 Introduction 1
2 Related Work 5
3 Method 8
3.1 Detecting action candidate spots . . . . . . . . . . . . . . . . . 10
3.2 Learning visual traits regarding a temporal order . . . . . . . . 10
3.3 Generating future frames . . . . . . . . . . . . . . . . . . . . . . 14
3.4 Detecting actions by modeling temporal correlations . . . . . . . 14
3.5 Training . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.6 Data augmentation . . . . . . . . . . . . . . . . . . . . . . . . . 17
4 Experiments 19
4.1 Experimental Setup . . . . . . . . . . . . . . . . . . . . . . . . . 19
4.2 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . 21
5 Conclusion and Future work 29