Extracting content from instructional videos by statistical modelling and classification |
| |
Authors: | Chekuri Choudary Tiecheng Liu |
| |
Affiliation: | (1) Department of Computer Science and Engineering, University of South Carolina, Columbia, SC 29208, USA |
| |
Abstract: | This paper presents a robust approach to extracting content from instructional videos for handwritten recognition, indexing
and retrieval, and other e-learning applications. For the instructional videos of chalkboard presentations, retrieving the
handwritten content (e.g., characters, drawings, figures) on boards is the first and prerequisite step towards further exploration
of instructional video content. However, content extraction in instructional videos is still challenging due to video noise,
non-uniformity of the color in board regions, light condition changes in a video session, camera movements, and unavoidable
occlusions by instructors. To solve this problem, we first segment video frames into multiple regions and estimate the parameters
of the board regions based on statistical analysis of the pixels in dominant regions. Then we accurately separate the board
regions from irrelevant regions using a probabilistic classifier. Finally, we combine top-hat morphological processing with
a gradient-based adaptive thresholding technique to retrieve content pixels from the board regions. Evaluation of the content
extraction results on four full-length instructional videos shows the high performance of the proposed method. The extraction
of content text facilitates the research on full exploitation of instructional videos, such as content enhancement, indexing,
and retrieval.
|
| |
Keywords: | Instructional video analysis Video text detection E-learning |
本文献已被 SpringerLink 等数据库收录! |
|