A framework for video event classification by modeling temporal context of multimodal features using HMM

Hsuan Sheng Chen*, W. J. Tsai

*Corresponding author for this work

研究成果: Article同行評審

6 引文 斯高帕斯(Scopus)

摘要

Semantic high-level event recognition of videos is one of most interesting issues for multimedia searching and indexing. Since low-level features are semantically distinct from high-level events, a hierarchical video analysis framework is needed, i.e., using mid-level features to provide clear linkages between low-level audio-visual features and high-level semantics. Therefore, this paper presents a framework for video event classification using temporal context of mid-level interval-based multimodal features. In the framework, a co-occurrence symbol transformation method is proposed to explore full temporal relations among multiple modalities in probabilistic HMM event classification. The results of our experiments on baseball video event classification demonstrate the superiority of the proposed approach.

原文English
頁(從 - 到)285-295
頁數11
期刊Journal of Visual Communication and Image Representation
25
發行號2
DOIs
出版狀態Published - 1 二月 2014

指紋 深入研究「A framework for video event classification by modeling temporal context of multimodal features using HMM」主題。共同形成了獨特的指紋。

引用此