Next Article in Journal
Comparative Analysis of Energy and Exergy Performance of Hydrogen Production Methods
Previous Article in Journal
Generative Art with Swarm Landscapes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Video Summarization Based on Mutual Information and Entropy Sliding Window Method

1
School of Computer Science and Engineering, South China University of Technology, Guangzhou 510006, China
2
School of Software Engineering, South China University of Technology, Guangzhou 510006, China
*
Authors to whom correspondence should be addressed.
WenLin Li and DeYu Qi contributed equally to this work.
Entropy 2020, 22(11), 1285; https://doi.org/10.3390/e22111285
Submission received: 9 October 2020 / Revised: 6 November 2020 / Accepted: 10 November 2020 / Published: 12 November 2020

Abstract

:
This paper proposes a video summarization algorithm called the Mutual Information and Entropy based adaptive Sliding Window (MIESW) method, which is specifically for the static summary of gesture videos. Considering that gesture videos usually have uncertain transition postures and unclear movement boundaries or inexplicable frames, we propose a three-step method where the first step involves browsing a video, the second step applies the MIESW method to select candidate key frames, and the third step removes most redundant key frames. In detail, the first step is to convert the video into a sequence of frames and adjust the size of the frames. In the second step, a key frame extraction algorithm named MIESW is executed. The inter-frame mutual information value is used as a metric to adaptively adjust the size of the sliding window to group similar content of the video. Then, based on the entropy value of the frame and the average mutual information value of the frame group, the threshold method is applied to optimize the grouping, and the key frames are extracted. In the third step, speeded up robust features (SURF) analysis is performed to eliminate redundant frames in these candidate key frames. The calculation of Precision, Recall, and F m e a s u r e are optimized from the perspective of practicality and feasibility. Experiments demonstrate that key frames extracted using our method provide high-quality video summaries and basically cover the main content of the gesture video.

1. Introduction

Nowadays, camera-based human-computer interaction (HCI) is increasingly being applied to intelligent life. The technological advancements in video acquisition approaches led to independence in video storage capacities [1]. Extracting the frames that can express the main content of video sequences, called key frame extraction, is one of the important methods of video summarization [2]. Due to the diversity features of different videos, there is no universal key frame extraction algorithm, especially for specific content. As a more convenient way of communication, gesture video analysis is widely studied in the field of human-computer interaction. Effectively extracting key-frames from gesture video repositories is still a challenging task. According to the structure information and diversity features, there are four major categories of methods used in key frame extraction.
The first method is based on shot detection and considers key frame extraction as a shot segmentation problem. In [3], the key frames are extracted by using histogram features based on the shots change detecting. This method is suitable for video sequences with less shot changes and simple content with fewer redundant frames.
The second major category of key frame extraction is based on clustering, where frames are clustered into groups according to the principle of similarity. In [4], a modified k-means clustering method is used to extract key frames. In [5], video summarization is formulated as a sequential decision, and treated diversity reward measure as a k-medoids problem; therefore, cluster centers are selected as key frames. The clustering algorithm is an unsupervised learning process and it is sensitive to dirty data. It has a large amount of computation, which is more suitable for small data calculation.
The third type of method is based on motion features such as motion vector estimation [6] and the optical flow method [7]. However, the motion feature is usually local information, and for locally complicated videos, using a motion estimation method may extract many redundant frames. Optical flow estimates the displacement area between two frames, providing motion and structure information. Because the calculation is the displacement of sub-pixel, there will be higher accuracy for motion estimation and more suitable for complex movement; however, pixel-level processing may result in excessive computation.
The fourth method is based on feature extraction and analysis, which uses feature descriptors such as color characteristics and texture as the basis for analysis. By setting appropriate threshold parameters, key frames with significant variations are extracted through inter-frame feature comparison [8].
In general, video retrieval technology includes: scene analysis, video structure processing, shot segmentation, sequence feature extraction, video content summary, and so forth. Above all, key frame extraction is the premise of indexing and summarization for the video sequence. These algorithms either use certain features to distinguish key frames, resulting in inaccurate extraction when the features are insufficient or disturbed by noise, or based on pixel-level motion extraction, which may lead to excessive calculations. To address these problems, we propose a key frame extraction method for gesture video called the Mutual Information and Entropy based adaptive Sliding Window (MIESW). The first step is to preprocess the video into a series of frames and resize the image. The second step refers to a rough temporal segment—it divides pre-processed videos into short segments by employing an improved sliding window with mutual information and entropy value as metrics for the sake of extracting candidate key-frames. In the third step, we apply speeded up robust features (SURF) features to determine similarity between candidate key-frames, then remove frames with high similarity to eliminate redundancy of key-frames sequences. As a result, the most representative frames can basically cover the main content of the given video. Precision, recall, and F-measure criterion are optimized conceptually, aiming to more accurately evaluate the pros and cons of the algorithm.
The remaining chapters of the paper are arranged as follows: Section 2 covers related work, Section 3 introduces algorithms and experimental methods, Section 4 conducts the experiment and clarifies results, and Section 5 summarizes the content of the paper and points out future work.

2. Related Work

Indexing, matching, and summarization of video sequences are popular areas, and the most popular method is based on key frame extraction. In order to extract key frames, domestic and foreign experts have conducted extensive research in different fields.
Shot detection are methods of extracting key frames. Rachida et al. [9] proposed a technology based on shot clip detection and video summarization, and used it in the editing of movies, documentaries, sports, and other videos. This method uses adaptive mean shift algorithm and global orientation features to build a key frames extraction method termed mean shift-based keyframes for video summarization (MSKVS). The algorithm proposed global frame feature vector (GFFV) descriptor and used it to describe visual content. However, the mean shift algorithm may get stuck at a saddle point, and more analysis should be done to make MSKVS suitable for real-time scenarios. In the action shot summary method proposed by Meghdadi et al. [10], the summary of the action shot is completed by merging still images aligned with the action frame. The algorithm relies on the motion trajectory of the action to generate the shot summary. When there are multiple targets or occlusions in the video, this may cause segmentation errors of the action shots. In the field of visual positioning, key frames extraction approach based on shot detection is used to construct an offline image database to store visual maps with location information. The algorithm proposed by [11] presets the first key frames and uses fixed-size groupings as shots to select key frames instead of grouping according to the similarity of content, which may lead to a decrease in the accuracy of key frame extraction.
The cluster-based key frames extraction methods are also very popular. In [4], a method based on sparse coding and k-means clustering is used to extract key frames, and added some conditions to select the number of key frames from each shot. However, these conditions are relatively harsh, and the constraint conditions of the sparse coding vector quantization method are too strict, which leads to more complicated parameters for extracting key frames. Avila et al. [12] propose a method of static video summaries based on color feature and K-means clustering. In [5], video summarization is formulated as an unsupervised decision process, and treated diversity reward measure as a k-medoids problem. Since it is necessary to specify the number of cluster centers, that is, the number of key frames obtained, this may result in key frame missed detection or excessive detection due to incorrect parameter settings. There are also some improved algorithms for extracting key frames based on semantic clustering. For example, Yin et al. [13] constructed an innovative algorithm based on semantic connections among the elements. In order to construct SeTree, a normalized graph cut clustering algorithm by combining visual features, textual information and user preferences is proposed. For key frames extraction, apart from using clustering to detect shot boundaries, it also uses personalized saliency and spatio–temporal saliency to score the importance of segments. Nevertheless, this algorithm needs to encode and cluster image features, which is slightly complicated for occasions with high real-time requirements. Generally, the clustering algorithms used to extract key frames are unsupervised learning processes that require clustering centers and are sensitive to dirty data. The clustering methods may also have the disadvantage that the obtained key frames lose the time information of the original video. These method usually have large amounts of calculation and are more suitable for small data calculations.
The key frames methods based on motion feature extraction have also attracted people’s attention. Zhu et al. [14] maps the original high-dimensional data to the low-dimensional state space, clusters motion data, and uses the segmentation points as the extraction basis of key frames. However, this method adopts the idea of dimensionality reduction, combined with motion features, it will lose local features and may lead to incorrect results. Wolf [15] first proposed the method of using optical flow calculation to extract key frames. Inspired by Wolf, Gao et al. [16] proposes a method of summarizing video using optical flow tensor (OFT) combined with hidden Markov model (HMM). The combination of the two methods can accurately capture the motion information of the video, but this method is sensitive to the video content and is more suitable for precise video extraction of subtle differences. Optical flow estimation is a pixel-level calculation method, which estimates the displacement by calculating the gray change value in two frames. Using optical flow estimation algorithms to extract key frames require that the brightness and space of the video remain almost constant, which greatly limits the application of the algorithm.
Scholars have also done some research on constructing feature descriptors to extract key frames from videos. Extraction algorithms based on image features are usually all non-domain specific and general. In [17], color features and structural features are used to describe the feature of the image content. In [18], the local feature SURF is selected to extract local points as the frame feature descriptor, and then used to analyze the frame sequence with the sliding window and extract key frames. In each shot, the most typical frames are extracted as [19] calculates the characteristics of the energy and standard deviation of each sub-band after contour wave transformation and forms a feature vector representation to extract key frames. The method of extracting key frames based on a single feature cannot effectively express the video content to some extent, especially when the video shots is complicated.
The above research progress on key frames is mainly for specific areas such as medicine, animation, content extraction, indoor visual positioning, monitoring analysis, human motion data capture, etc. In these fields, key frame extraction technology provides convenience for quickly and accurately understanding relevant video content. Gesture recognition is a prevalent way of human-computer interaction. Research on dynamic gesture recognition can better optimize the real-time requirements of human-computer interaction. However, dynamic gesture videos generally contain hundreds of frames, which complicates the task of directly analyzing and processing video data. By using key frame extraction technology to summarize the main content of the video, the calculation of redundant information can be reduced and the real-time performance of human-computer interaction can be improved; therefore, it is of great significance to extract key frames from gesture video as an advanced gesture recognition algorithm.

3. Key Frames Extraction and Feature Fusion Principle

The key frame extraction method proposed in this paper is mainly divided into the following steps, namely, extracting image information, optimizing image information construction efficiency to improve storage occupation, and analyzing the relationship between video sequences to extract representative frames. The principles, feasibility and experimental demonstration of each step are separately described below.

3.1. Entropy and Mutual Information Theory

As the basis of information theory, since 1995, mutual information has been applied to the fields of image comparison and multi-modal medical image matching [20] and has gradually been applied to shots detection [21] and image segmentation. The literature [22] combines the mutual information algorithm and k-means method to segment medical and automobile images. For a given image, it assumes that pixel gray values are independent and appear randomly in the image. The image distribution is expressed as P = { p 0 , p 1 , p L 1 } , where p i represents the image and the gray value.
For the image frames X, their entropy H [ X ] can be expressed as (1):
H [ X ] = k p X ( k ) ln ( p X ( k ) )
where p X ( k ) denotes the probability density function of frames X, the probability function can be obtained by normalized histogram of the gray-scale pixel.
Mutual information between X and Y noted as I [ Y , X ] , which can be defined as the difference between entropy of X and conditional entropy of Y given the condition of X noted as H ( Y | X ) as follows:
I [ Y , X ] = H [ X ] H [ X | Y ]
In this paper, X , Y denote frame images, then the joint probability distribution is the normalized histogram of the two images. Mutual information is applied in video extraction as a basis for judging the similarity between frames. If the mutual information value is greater than the threshold, that means the content of the two frames has changed a lot. Entropy H [ X ] measures the degree of change in gesture movements. The H [ X , Y ] indicates the similarity of the two frames.
In Figure 1, Venn diagram represents the relationship between image entropy, joint distribution and mutual information. The histogram represents the probability distribution of grayscale images.
The key frame extraction algorithm based on entropy and mutual information involves the following steps:
  • Input video: Input the video to extract key frames.
  • Conversion part: Convert the video into a frame sequence, count the length of the sequence, and adjust the frame size to N × N .
  • First, an adaptive sliding window is used to process the frame sequence. The mutual information value is calculated for the frame group entering the window, the window size is adaptively adjusted according to the difference between the frames, and similar frames are classified into a group. Then according to the entropy value of the frame, the groups with similar content are further merged, and finally the frame closest to the average mutual information value between the groups is selected to enter the candidate sequence of the key frame.
  • Eliminate redundancy: Use SURF to extract the features of the frames in the candidate sequence, and eliminate the redundant frames with high similarity.
  • Output key frames: Output the final key frame extraction results. It should be noted that during the key frame extraction process, the frame sequence retains time sequential information.
Figure 2 elaborates the mechanism used in the proposed solution. The input and output modules are represented by blue squares, the red squares represent the pre-conversion module and the candidate sequence adjustment module, and the inside of the yellow square is the MIESW algorithm module.

3.2. The Proposed Mutual Information and Entropy Based Sliding Window Method (MIESW) for Key Frames Extraction

The key frame extraction method proposed in this paper is called MIESW, which is mainly divided into the following steps: calculate the inter-frame mutual information value to acquire spatial-temporal information, use an adaptive sliding window method to group similar content frames, and apply an entropy-based threshold method optimization grouping. As a result, a candidate key frame sequence is obtained. In order to obtain the key frame sequence, the feature extraction algorithm is used to analyze the similarity of the frames in the candidate frame sequence and eliminate redundant frames.

3.3. Improved Adaptive Sliding Window Method to Extract Key Frames

Traditional key frame extraction algorithms based on sliding window measure the similarities between adjacent frames. The purpose of processing data using a sliding window is to analyze the local features and overall characteristics of the video sequences. In [23], combined with ensemble learning, a sliding window-based support vector regression method is proposed to predict micrometeorological data. In [24], three consecutive sliding windows are considered to calculate G1-continuities errors as the criterion. This method takes the shots boundaries and transitions types into consideration, however, calculation may be huge. The fixed window size will also result in poor adaptability of the sliding window algorithm to different types of video sequences. This method considers the boundary and transition type of the camera, but the amount of calculation may be large.
Taking these factors into consideration, an improved the sliding window method is proposed, which can be seen in Figure 3. In the segmentation algorithm proposed in paper, we calculate the average similarity on mutual information between the frame we have traveled and frames within the current sliding window.
As shown in Figure 3, F 1 represents the first frame in the video sequence V, and F N indicates that the sequence has N frames. The t axis represents a time-line. In Figure 3 the green line represents the sliding window, the changing in its length represents the process of adaptively changing the size of the sliding window, and the red line represents the size of the window no longer increases. The frames in the fixed window are regarded as a group. The last group is represented by a blue line, because the length of the frame sequence entering the sliding window may be less than the initial value of the sliding window. In this case, the frame sequence is directly grouped and marked as g n . The grouped frame sequence is denoted as V = { g 1 , g 2 , , g n } . The key frames extraction framework named MIESW consists of two stages: segmentation and grouping, which are summarized as Algorithms 1 and 2. The framework of Algorithms 1 and 2 are shown in Figure 4.
We designed Algorithm 1 as follows: When using the MIESW method to extract key frames, we use the adaptive sliding window method to group the frame sequence with primary similarity according to Algorithm 1. First, initialize the sliding window, set the threshold (step 1), then determine whether the last frame that entered the sliding window is the final frame (step 2). Next, we calculate mutual information value of the frame group in the sliding window (step 3), and then add the frame next to the current window’s right boundary to judge the degree of similarity changes within the group (step 4). Thereafter, the size of the sliding window is adaptively adjusted according to the threshold (step 5), and the frame group are output after the loop ends. The detailed algorithm is as follows:
step 1.
Pre-defined the initial length of the sliding window to w. Set the threshold T H 1 . Start the algorithm from the first frame.
step 2.
Make sure the last frame in the sliding window is not the final frame F N . If true, denote the current group as g n , algorithm is over, output the result. Otherwise, initial the current window size s = w , go to step 3.
step 3.
The window slides on the sequence, and frames entering the sliding window are denoted as G q = { F p , F p + 1 , , F q } , satisfying: q = p + s . Calculate the mutual information value M of each consecutive two frames in G q , recorded as M I q = { M p , M p + 1 , , M q } , and then calculate the mean value as M I q ¯ = i = p q M I p q / s .
step 4.
Add the frame next to the current window’s right boundary F q + 1 into the group and obtain M I q + 1 ¯ . Calculate the absolute value of the difference between M I q ¯ and M I q + 1 ¯ as D q . D q = | M I q ¯ M I q + 1 ¯ | If D q > T H 1 , it means that the newly added frame F s + 1 reduces the overall correlation of the original group G q , so denote the group as g q . Set p = q + 1 , go to step 2. Otherwise go to step 5.
step 5.
D p < T H 1 , it indicates that the newly added frame has a correlation with the original group. If F q F N , then s = s + 1 , go to step 4. If F q = F N , go to step 2.
After the first grouping, the video sequence is approximately divided into some groups. Moreover, to merge similar adjacent groups, we perform Algorithm 2. Algorithm 2 is used to extract candidate key frames. Use the frame groups output by Algorithm 1 as input (step 1), then calculate the entropy indicator and threshold (step 2). After high similarity groups are merged according to the entropy indicator and the threshold (step 3), candidate key frames are extracted from the new groups based on mutual information characteristics (step 4). The detailed algorithm is as follows:
step 1.
Represented groups in the sequence as V = { g 1 , g 2 , , g n } as input.
step 2.
Calculate the standard deviation of the entropy values of each group as e V = { e 1 , e 2 , , e n } , and the average standard deviation is obtained as the threshold T H 2 = 1 n e / n .
step 3.
In the second grouping, adjacent segment which is smaller than the threshold T H 2 will be merged. For e i e V , if e i < T H 2 , merge g i to g i 1 ; if not, keep the group unchanged.
step 4.
After this process, the final groups is denoted as V = { g 1 , g 2 , . , g m } , in each group, the frame closets to the average mutual information value is selected as the key frame: f k e y = { f 1 , f 2 , , f m } .
Figure 5 shows a sequence of frames from the gesture sequence `Clic’, which contains 99 frames. The detailed process of the MIESW algorithm is shown in the Figure 6, the green line indicates the number of frames included in the frame segment which is the frame group length. The red line indicates that the newly added frame does not meet the group condition. For the first frame segment g 1 , its length is s. For the case where the difference D 1 is smaller than the threshold, as shown by the F s + 1 frame in the Figure 6. The segment g 2 shown in the Figure 6 has a length of m. It should be noted that if the length of the last group is smaller than the original window size s, the last segment is directly set to g n as shown by the blue line.

3.4. Remove Redundant Frames

To remove redundant frames from candidate key frames, we analyze sequences similarity through image feature extraction. In image processing, feature extraction has become a popular method in image classification [25], image retrieval [26], target detection [27], and image segmentation [28]. In recent years, most of the research has focused on the area of deep learning, which extracts eigenvector or characteristic vector by using different convolution kernels.
However, traditional methods with mathematical operators such as Harris Corner Detector, SURF, scale-invariant feature transform (SIFT), etc. still have the advantages in key-points orientation description, scale analysis, and illumination variation. Proposed by David, the SIFT [29] detector can find feature points on different scale spaces. As a robust local descriptor, SURF is usually used in key-frame extraction tasks. We evaluate SURF detectors as a metric to counting the number of features in each key-frame. The ultimate matching of SURF feature is to measure the similarity of the two key-frames and serve as the basis for removing redundant frames. SURF is inspired by SIFT descriptor, which is several times faster than SIFT at the cost of fewer feature points to be detected. Based on Haar wavelet transform, SURF performs Hessian algorithm to detect key points.
SURF uses a Hessian Matrix to extract feature points. The Hessian Matrix is a square matrix composed of the second-order partial derivative, which describes the local curvature of the multivariate function. For an image I ( x , y ) , we denote the Hessian Matrix as:
H ( x , y , σ ) = L x x ( x , y , σ ) L x y ( x , y , σ ) L x y ( x , y , σ ) L y y ( x , y , σ )
with,
L ( x , y , σ ) = G ( σ ) I ( x , y )
G ( σ ) = 1 2 π σ 2 e x 2 + y 2 2 σ 2
SURF compares each pixel point processed by the Hessian matrix with its image domain (image of the same size) and all neighboring points in the adjacent scale space. When it is greater than (or less than) all neighboring points, the point is feature point. In the SURF algorithm, it counts the Harr wavelet features in the circular neighborhood of the feature point with the fan-shaped area as a unit, and takes the fan-shaped direction of the largest feature value as the main direction of the feature point. SURF has excellent scale-invariant features which are advantageous for specific target-gesture feature extraction and image comparison. SURF vector matching is used for the candidate frame sequences to eliminate redundant frames. Taking the experimental sequence as an example, the candidate key-frames numbers are 14, 33, 72, and 98. We evaluated the SURF detectors ability to process matching images—the results are shown in Figure 7:
Analyzed from the Figure 7 and Table 1, the number of point pairs on SURF matching in the third figure is the maximum compared to the first and the second figure. In other words, the frame 72 and the frame 98 have a high degree of similarity. We calculate the similarity between candidate key-frames by analyzing SURF features and matching degree.
SURF similarity describes the resemblance of candidate frames in visual features. If the indicator is as high as 50%, it demonstrates that there are redundant frames in the candidate frames, and redundant frames will be eliminated. Based on this method, key-frames are determined. So the extraction results of this video is shown as Figure 8.

4. Experiment

4.1. Improved Precision, Recall, and F m e a s u r e Criteria

Three quantitative assessment metrics based on the visual comparison which have been used to evaluate the quality of keyframes extraction methods, i.e., Precision, Recall, and F m e a s u r e , that are defined as follows:
P = N m N A K
R = N m N U K
F m e a s u r e = 2 × P × R P + R = 2 × N m N A K + N U K
where P denotes Precision, R denotes Recall, N A K is the number of automatic keyframes and N U K is the number of user-set keyframes, N m as the number of matching keyframes indexes between N A K and N U K .
However, the quantitative metrics rely too much on the key frames set by the user, which may cause the extracted representative key frames to be incorrectly judged due to different set frame numbers. Another method [30] takes this situation into consideration and adds the parameter of similar frames to the improved metrics to reduce the misjudgment, as shown in (9)–(11):
P = N U K N U K + N S
R = N U K N U K + N M S
F m e a s u r e = 2 × P × R P + R
where N S denotes the number of keyframes found similar to the user-set keyframes, N M S is the number of missed detection of the correct result, which is user-set keyframes.
Considering the continuity of frame sequences, it is important to reduce the subjective in this method. Instead of obtaining precision and recall by counting automatic key frames and user-set key frames as shown in (6)–(8), or calculating similar user-set key frames and missed correct key frames as shown in (9)–(11), an optimized assessment metrics has been approved. Since the gestures appearing in the video are coherent, the key frame set by the user cannot be accurate to a specific value, so setting it within a reasonable range is a feasible measure. The improved evaluation metrics are as (12)–(14):
P = N s N g
R = N s N u
F m e a s u r e = 2 × P × R P + R = 2 × N s N g + N u
where N u is a collection of key frame areas set by the user, N g is the key frame extracted by the algorithm. N s is the number of matched frames, which refers to the number of frames that fall correctly in the user-set key frame area set.
Only one key frame is allowed to fall in each area. If two or more key frames extracted by the algorithm fall into the same area, the correct matching number can only count one frame; if there is no algorithm extracted frame falling within a user-set key frame area, the area is determined as the missed detection area and the correct matching number is not changed in this case.

4.2. Experimental Results

Experiments were conducted on many videos, and some examples of experimental results are shown below. Experiments were performed on the hand posture video sequences ’Clic’ and ’StopGraspOk’ of Dr.Sebastien Marcel-IDIAP2001. The ’Clic’ sequences contain three gestures: raising the index finger, clenching fists, and then raising the index finger.
For the comparison, four different baselines were selected: (1) entropy-based (with color histograms) [31], (2) color-based [17], (3) sliding window based (with gist and sift features) [18], and (4) our method without deduplication optimization.
In Table 2, these sequences denoted as {C1, C2} and {S1, S2}, respectively. Frames are color images saved in portable any map (PNM) format with the size of 62 × 58 × 3 . Therefore, in general, the cubic spline interpolation is used to adjust the image size to 64 × 64 × 3 . The proposed(w/o optimization) method and our method are performed on the experimental data using the threshold of 0.3. The other three algorithms are implemented according to the settings of corresponding original papers.
Experimental results are evaluated using the modified Precision (P), Recall (R), and F m e a s u r e (F) measurement. Table 2 shows the results. The entropy-based [31] method calculated the entropy value of the frame to perform key frames selection, and the color-based [17] method extract color characteristics to obtain temporal segment on the video. The sliding window based [18] extracts features of frames by using gist and sift features and then use the sliding window to extract key frames. These three methods share different cogitation on temporal segments, comparing the approval method of others in the identical dataset is a fair and convincing way.
In order to make the results more intuitive and easier for objective comparison, Figure 9 shows the PRF comparison results. The PRF criterion reflect quantitative results of key frame selection. It can be seen that our method has achieved better results. There is improvement on Precision, Recall, and F m e a s u r e in our algorithm compared with these algorithms.
Some results of key frames extraction are shown in Figure 10. It can be seen with the comparison results that our proposed method conducted better key frames extraction for the dataset. We infer that it is because our algorithm fully considers the inter-frame information and the characteristics of frames, then performs adaptive sliding-window based adaptive grouping to extract candidate key frames, and finally eliminates redundant frames. Entropy-based method depends only on the maximum and minimum of the frames’ entropy, it may missing some key frames that are not at extreme points. Since the background color of our dataset is relatively close to the foreground color, so the color-based method does not perform well. Due to the shot almost unchanged, the sliding-window based method may have trouble dividing frames in groups. Although the proposed method without optimization has selected key frames that can represent video content, there are also many redundant frames. However, we can remove some redundant frames by performing the SURF similarity method. Using the key frame extraction algorithm proposed in this paper can increase feature extraction efficiency and eliminate redundant information.

5. Conclusions

In this paper, we have devised and performed a video summarization approach called MIESW for gesture videos. The innovations of this method are as follows: considering the correlation between frames, an improved key frame extraction algorithm based on adaptive sliding window is proposed. When sliding on the video sequences, the window size is adaptively adjusted to accommodate the characteristics of the sequences expressed by mutual information values. After the second frame grouping measured by entropy, it is ensured that similar content is correctly classified and representative frames are extracted as key frames. Extract the SURF of candidate key frames, and then analyze the similarity between them to delete redundant frames and obtain key frames. The calculation of Precision, Recall, and F m e a s u r e are optimized from the perspective of practicality and feasibility which improves the definition of user-set key frames to key frame intervals, taking into account the continuity of video content. Experimental results demonstrate that the proposed method performs well on the dataset. However, with the continuous expansion of the application range of gestures, the key frames extraction for videos in more complex backgrounds may be a limitation of our algorithm. In future, we plan to optimize the algorithm in a complex background, and use deep learning framework for automatic feature extraction. The key frame extraction algorithm proposed in this paper is not limited by video content and is applied to gesture extraction. However, the video content is varied and the video lengths are different. The increase in video services has also challenged video summarization and key frame extraction algorithms. In the subsequent improvements, we will further study the promotion of algorithm to the extraction of multiple videos.

Author Contributions

Data curation, J.G. and J.Y.; formal analysis, C.Z.; funding acquisition, D.Q.; writing—original draft, W.L. and D.Q. All authors contributed to the writing and revisions. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (No. 61070015), Natural Science Fund Team Project of Guangdong Province (No. 10351806001000000), Frontier and key technology innovation projects of Guangdong Province (No. 2014B010110004) and Industry-University-Research Collaborative Innovation Project of Guangzhou City (No. 201604016074). https://search.crossref.org/funding, any errors may affect your future funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Money, A.G.; Agius, H. Video summarisation: A conceptual framework and survey of the state of the art. J. Vis. Commun. Image Represent. 2008, 19, 121–143. [Google Scholar] [CrossRef] [Green Version]
  2. Hu, W.; Xie, N.; Li, L.; Zeng, X.; Maybank, S. A Survey on Visual Content-Based Video Indexing and Retrieval. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2011, 41, 797–819. [Google Scholar] [CrossRef]
  3. Ejaz, N.; Tariq, T.B.; Baik, S.W. Adaptive key frame extraction for video summarization using an aggregation mechanism. J. Vis. Commun. Image Represent. 2012, 23, 1031–1040. [Google Scholar] [CrossRef]
  4. Amiri, A.; Fathy, M. Hierarchical Keyframe-based Video Summarization Using QR-Decomposition and Modified-Means Clustering. EURASIP J. Adv. Signal Process. 2010, 2010, 892124. [Google Scholar] [CrossRef] [Green Version]
  5. Zhou, K.; Qiao, Y.; Xiang, T. Deep reinforcement learning for unsupervised video summarization with diversity-representativeness reward. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  6. Hannane, R.; Elboushaki, A.; Afdel, K.; Naghabhushan, P.; Javed, M. An efficient method for video shot boundary detection and keyframe extraction using SIFT-point distribution histogram. Int. J. Multimed. Inf. Retr. 2016, 5, 89–104. [Google Scholar] [CrossRef]
  7. Zhu, Y.; Li, K.; Jiang, J. Video super-resolution based on automatic key-frame selection and feature-guided variational optical flow. Signal Process. Image Commun. 2014, 29, 875–886. [Google Scholar] [CrossRef]
  8. Smeaton, A.F.; Over, P.; Doherty, A.R. Video shot boundary detection: Seven years of TRECVid activity. Comput. Vis. Image Underst. 2010, 114, 411–418. [Google Scholar] [CrossRef] [Green Version]
  9. Hannane, R.; Elboushaki, A.; Afdel, K. MSKVS: Adaptive mean shift-based keyframe extraction for video summarization and a new objective verification approach. J. Vis. Commun. Image Represent. 2018, 55, 179–200. [Google Scholar] [CrossRef]
  10. Meghdadi, A.H.; Irani, P. Interactive Exploration of Surveillance Video through Action Shot Summarization and Trajectory Visualization. IEEE Trans. Vis. Comput. Graph. 2013, 19, 2119–2128. [Google Scholar] [CrossRef]
  11. Ma, L.; Yang, H.; Tan, X.; Feng, G. Image Keyframe-based Visual-Depth Map Establishing Method. J. Harbin Inst. Technol. 2018, 50, 23–31. [Google Scholar]
  12. De Avila, S.E.F.; Lopes, A.P.B.; da Luz, A., Jr.; de Albuquerque Araújo, A. VSUMM: A mechanism designed to produce static video summaries and a novel evaluation method. Pattern Recognit. Lett. 2011, 32, 56–68. [Google Scholar] [CrossRef]
  13. Yin, Y.; Thapliya, R.; Zimmermann, R. Encoded Semantic Tree for Automatic User Profiling Applied to Personalized Video Summarization. IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 181–192. [Google Scholar] [CrossRef]
  14. Zhu, D.; Wang, Z. Extraction of keyframe from motion capture data based on motion sequence segmentation. J. Comput.-Aided Des. Comput. Graph. 2008, 20, 787–792. [Google Scholar]
  15. Wolf, W. Key frame selection by motion analysis. In Proceedings of the 1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings, Atlanta, GA, USA, 9 May 1996; Volume 2, pp. 1228–1231. [Google Scholar] [CrossRef]
  16. Gao, X.; Li, X.; Feng, J.; Tao, D. Shot-based video retrieval with optical flow tensor and HMMs. Pattern Recognit. Lett. 2009, 30, 140–147. [Google Scholar] [CrossRef]
  17. Shi, Y.; Yang, H.; Gong, M.; Liu, X.; Xia, Y. A fast and robust key frame extraction method for video copyright protection. J. Electr. Comput. Eng. 2017, 2017, 1231794. [Google Scholar] [CrossRef]
  18. Yu, L.; Cao, J.; Chen, M.; Cui, X. Key frame extraction scheme based on sliding window and features. Peer-Netw. Appl. 2018, 11, 1141–1152. [Google Scholar] [CrossRef]
  19. Rao, P.C.; Das, M.M. Keyframe Extraction Method Using Contourlet Transform. In Proceedings of the 2012 International Conference on Electronics, Communications and Control; IEEE Computer Society: Washington, DC, USA, 2012; pp. 437–440. [Google Scholar]
  20. Maes, F.; Collignon, A.; Vandermeulen, D.; Marchal, G.; Suetens, P. Multimodality image registration by maximization of mutual information. IEEE Trans. Med Imaging 1997, 16, 187–198. [Google Scholar] [CrossRef] [Green Version]
  21. Ngo, C.W.; Ma, Y.F.; Zhang, H.J. Video summarization and scene detection by graph modeling. IEEE Trans. Circuits Syst. Video Technol. 2005, 15, 296–305. [Google Scholar]
  22. Rigau, J.; Feixas, M.; Sbert, M.; Bardera, A.; Boada, I. Medical image segmentation based on mutual information maximization. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2004; pp. 135–142. [Google Scholar]
  23. Kaneda, Y.; Mineno, H. Sliding window-based support vector regression for predicting micrometeorological data. Expert Syst. Appl. 2016, 59, 217–225. [Google Scholar] [CrossRef] [Green Version]
  24. Huang, C.; Wang, H. Novel Key-frames Selection Framework for Comprehensive Video Summarization. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 577–589. [Google Scholar] [CrossRef]
  25. Huang, G.; Liu, Z.; Laurens, V.D.M.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  26. Kan, S.; Cen, Y.; He, Z.; Zhang, Z.; Zhang, L.; Wang, Y. Supervised Deep Feature Embedding with Hand Crafted Feature. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2019, 28, 5809–5823. [Google Scholar] [CrossRef]
  27. Wang, Y.J.; Ding, M.; Kan, S.; Zhang, S.; Lu, C. Deep Proposal and Detection Networks for Road Damage Detection and Classification. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018. [Google Scholar]
  28. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  29. Lowe, D.G. Distinctive Image Feature from Scale-Invariant Key points. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  30. Chasanis, V.T.; Ioannidis, A.I.; Likas, A.C. Efficient key-frame extraction based on unimodality of frame sequences. In Proceedings of the IEEE 2014 12th international conference on signal processing (ICSP), HangZhou, China, 19–23 October 2014; pp. 1133–1138. [Google Scholar]
  31. Tang, H.; Xiao, W.; Liu, H.; Sebe, N. Fast and Robust Dynamic Hand Gesture Recognition via Key Frames Extraction and Feature Fusion. Neurocomputing 2018, 331, 424–433. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Entropy H [ F ] measures the degree of change in gesture movements. The correlation entropy H [ F t , F ( t + 1 ) ] indicates the similarity of the two frames. The mathematical relationship between information entropy and mutual information is illustrated in (a), F t denotes a frame at time t, and H [ F t ] represents the information entropy of F t , I [ F t , F t + 1 ] represents the mutual information value of two consecutive frames at time t and t + 1 . (b) Shows two gray-scale gesture images, and (c) shows the gray-scale distribution histograms, which can be used to calculate the entropy of the image. In (d), the horizontal axis represents the gray gradation value, and the vertical axis shows the number of pixels. The joint histogram counts the frequencies at which different gray-value combinations appear at corresponding positions in the two images. The shapes of the two histograms are similar, indicating that their probability distribution of the pixel gray values in approximate.
Figure 1. Entropy H [ F ] measures the degree of change in gesture movements. The correlation entropy H [ F t , F ( t + 1 ) ] indicates the similarity of the two frames. The mathematical relationship between information entropy and mutual information is illustrated in (a), F t denotes a frame at time t, and H [ F t ] represents the information entropy of F t , I [ F t , F t + 1 ] represents the mutual information value of two consecutive frames at time t and t + 1 . (b) Shows two gray-scale gesture images, and (c) shows the gray-scale distribution histograms, which can be used to calculate the entropy of the image. In (d), the horizontal axis represents the gray gradation value, and the vertical axis shows the number of pixels. The joint histogram counts the frequencies at which different gray-value combinations appear at corresponding positions in the two images. The shapes of the two histograms are similar, indicating that their probability distribution of the pixel gray values in approximate.
Entropy 22 01285 g001
Figure 2. Flow chart of the Mutual Information and Entropy based adaptive Sliding Window (MIESW) algorithm for extracting key frames.
Figure 2. Flow chart of the Mutual Information and Entropy based adaptive Sliding Window (MIESW) algorithm for extracting key frames.
Entropy 22 01285 g002
Figure 3. The improved sliding window schematic.
Figure 3. The improved sliding window schematic.
Entropy 22 01285 g003
Figure 4. The framework of Algorithms 1 and 2.
Figure 4. The framework of Algorithms 1 and 2.
Entropy 22 01285 g004
Figure 5. Frames in the video of a gesture for the word `Clic’.
Figure 5. Frames in the video of a gesture for the word `Clic’.
Entropy 22 01285 g005
Figure 6. Schematic diagram of extracting key frames. A hand gesture sequence sample from Marcel-IDIAP2001, which contains 99 frames. The horizontal axis coordinate represents the mutual information value obtained by successive frame pairs in the video sequence. Data1 is the inter-frame mutual information value range from 2 to 3.5. Blocks of different colors indicate the grouping results after using Algorithms 2 and 3. The horizontal lines in color blocks indicate the mean value of each group. The key frames obtained by our method are in red boxes, which are the 17, 33, 72, and 98 frames.
Figure 6. Schematic diagram of extracting key frames. A hand gesture sequence sample from Marcel-IDIAP2001, which contains 99 frames. The horizontal axis coordinate represents the mutual information value obtained by successive frame pairs in the video sequence. Data1 is the inter-frame mutual information value range from 2 to 3.5. Blocks of different colors indicate the grouping results after using Algorithms 2 and 3. The horizontal lines in color blocks indicate the mean value of each group. The key frames obtained by our method are in red boxes, which are the 17, 33, 72, and 98 frames.
Entropy 22 01285 g006
Figure 7. SURF (Speeded Up Robust Features) analysis.
Figure 7. SURF (Speeded Up Robust Features) analysis.
Entropy 22 01285 g007
Figure 8. The key frames extraction result of the test video.
Figure 8. The key frames extraction result of the test video.
Entropy 22 01285 g008
Figure 9. Precision, Recall, and F m e a s u r e of different techniques on test videos. The purple bar, yellow bar, and blue line represent Precision, Recall, and F-measure, respectively. The horizontal axis in the figure represents four test videos, and the vertical axis represents the PRF evaluation metrics of the key frame extraction result. It can be seen that using the same test video, our proposed algorithm can obtain higher precision, recall, and F m e a s u r e . For detailed experimental results, see Table 2.
Figure 9. Precision, Recall, and F m e a s u r e of different techniques on test videos. The purple bar, yellow bar, and blue line represent Precision, Recall, and F-measure, respectively. The horizontal axis in the figure represents four test videos, and the vertical axis represents the PRF evaluation metrics of the key frame extraction result. It can be seen that using the same test video, our proposed algorithm can obtain higher precision, recall, and F m e a s u r e . For detailed experimental results, see Table 2.
Entropy 22 01285 g009
Figure 10. The key frames extractions applied on video S2 that contains 141 frames and consists of four gestures. Its gesture starts with the fist, then opens fingers, next clenches the fist, and finally opens the palm. For the 1st to 3rd comparison algorithms, especially the third gesture, has been leaked, which would be bad in key-frames extraction. However, in the fourth algorithm we proposed, although the third gesture was extracted, unfortunately, there are many redundant frames. The results of our proposed method shows a series of gesture changes can be extracted more obviously.
Figure 10. The key frames extractions applied on video S2 that contains 141 frames and consists of four gestures. Its gesture starts with the fist, then opens fingers, next clenches the fist, and finally opens the palm. For the 1st to 3rd comparison algorithms, especially the third gesture, has been leaked, which would be bad in key-frames extraction. However, in the fourth algorithm we proposed, although the third gesture was extracted, unfortunately, there are many redundant frames. The results of our proposed method shows a series of gesture changes can be extracted more obviously.
Entropy 22 01285 g010
Table 1. Similarity based on SURF descriptor.
Table 1. Similarity based on SURF descriptor.
Compared ImagesThe Best Matching ValueThe Worst Matching ValueSURF Similarity
(14, 33)0.04140.57460.3469
(33, 72)0.03130.55110.3864
(72, 98)0.01870.63720.5882
Table 2. Comparison of Precision (P), Recall (R), and F m e a s u r e (F) of different techniques on the test videos. The best results are highlighted in bold.
Table 2. Comparison of Precision (P), Recall (R), and F m e a s u r e (F) of different techniques on the test videos. The best results are highlighted in bold.
MethodVideo Name N u N g N s PRF
entropy-basedc15630.50000.60000.5445
c25210.50000.20000.2857
s17430.75000.42860.5455
s27940.44440.57140.5000
color-basedc15630.50000.60000.5445
c251040.40000.80000.5333
s17950.55560.71430.6250
s271140.36360.57140.4444
sliding-window-basedc15630.50000.60000.5445
c25730.42860.60000.5000
s17850.62500.71430.6667
s27940.44440.57140.5000
proposed (w/o optimization)c15530.60000.60000.6000
c25530.60000.60000.6000
s17840.50000.57140.5333
s27630.50000.42860.4615
our methodc15540.80000.80000.8000
c25740.57140.80000.6667
s17650.83330.71430.7692
s27650.83330.71430.7692
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, W.; Qi, D.; Zhang, C.; Guo, J.; Yao, J. Video Summarization Based on Mutual Information and Entropy Sliding Window Method. Entropy 2020, 22, 1285. https://doi.org/10.3390/e22111285

AMA Style

Li W, Qi D, Zhang C, Guo J, Yao J. Video Summarization Based on Mutual Information and Entropy Sliding Window Method. Entropy. 2020; 22(11):1285. https://doi.org/10.3390/e22111285

Chicago/Turabian Style

Li, WenLin, DeYu Qi, ChangJian Zhang, Jing Guo, and JiaJun Yao. 2020. "Video Summarization Based on Mutual Information and Entropy Sliding Window Method" Entropy 22, no. 11: 1285. https://doi.org/10.3390/e22111285

APA Style

Li, W., Qi, D., Zhang, C., Guo, J., & Yao, J. (2020). Video Summarization Based on Mutual Information and Entropy Sliding Window Method. Entropy, 22(11), 1285. https://doi.org/10.3390/e22111285

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop