Next Article in Journal
Exploring Signals on L5/E5a/B2a for Dual-Frequency GNSS Precise Point Positioning
Previous Article in Journal
Experimental Verification of the Concept of Using LOFAR Radio-Telescopes as Receivers in Passive Radiolocation Systems
Previous Article in Special Issue
DDoS Flood and Destination Service Changing Sensor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Important Scenes in Baseball Videos via a Time-Lag-Aware Multimodal Variational Autoencoder †

1
Graduate School of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-ku, Sapporo 060-0814, Hokkaido, Japan
2
Office of Institutional Research, Hokkaido University, N-8, W-5, Kita-ku, Sapporo 060-0808, Hokkaido, Japan
3
Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-ku, Sapporo 060-0814, Hokkaido, Japan
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in: Hirasawa, K.; Maeda, K.; Ogawa, T.; Haseyama, M. Important Scene Detection Based on Anomaly Detection using Long Short-Term Memory for Baseball Highlight Generation. In the Proceedings of the IEEE International Conference on Consumer Electronics—Taiwan (IEEE 2020 ICCE-TW), Taoyuan, Taiwan, 28–30 September 2020.
Sensors 2021, 21(6), 2045; https://doi.org/10.3390/s21062045
Submission received: 28 January 2021 / Revised: 4 March 2021 / Accepted: 10 March 2021 / Published: 14 March 2021

Abstract

:
A new method for the detection of important scenes in baseball videos via a time-lag-aware multimodal variational autoencoder (Tl-MVAE) is presented in this paper. Tl-MVAE estimates latent features calculated from tweet, video, and audio features extracted from tweets and videos. Then, important scenes are detected by estimating the probability of the scene being important from estimated latent features. It should be noted that there exist time-lags between tweets posted by users and videos. To consider the time-lags between tweet features and other features calculated from corresponding multiple previous events, the feature transformation based on feature correlation considering such time-lags is newly introduced to the encoder in MVAE in the proposed method. This is the biggest contribution of the Tl-MVAE. Experimental results obtained from actual baseball videos and their corresponding tweets show the effectiveness of the proposed method.

1. Introduction

The development of various network technologies and devices has resulted in the availability of many video distribution services. A tremendous number of videos has become viewable via these services. Since sports videos are popular videos among them, a large number of sports games is provided by sports video distribution services. Although users can easily obtain their desired videos, it is difficult for viewers to watch all of the games provided by these services [1]. Among various sports, a baseball game takes a longer time than other sports such as basketball and soccer. Therefore, techniques for viewers to understand the content of a long game in a short time are necessary [1].
The generation of highlights is one of the important techniques for enabling a long game to be watched in a short time. Highlights consist of important scenes such as scoring and home runs. Therefore, since important scene detection methods are necessary for the generation of highlights, various detection methods have been proposed [2,3,4,5,6,7,8]. The methods in [2,3,4] using videos detect important scenes by applying a hidden Markov model and a maximum entropy model to player movements and cheers from the audiences obtained from the target videos. Therefore, conventional video-based methods are effective for the representation of visual information of players and audiences. On the other hand, along with the development of microblogging services, methods using Twitter (https://twitter.com/ (accessed on 12 March 2021) for the detection of important scenes have been proposed [5,6,7]. By using Twitter, which is one of the microblogging services, the reactions of viewers during games can be obtained. People watching baseball games often post tweets, which are short text messages, related to games. Since posted tweets often include information on the reactions of viewers and the contents of the games, conventional Twitter-based methods have an advantage for the extraction of opinions or feelings of viewers. Therefore, by taking advantage of each modality, it is expected that a method using both Twitter and video analysis would enable high-quality detection of important scenes.
In order to use both tweets and videos, we need to consider multiple different modalities. Several methods focusing on the relationships across multiple different modalities have been proposed [9,10,11]. The method in [9] realizes efficient cross-modal video-text retrieval using multi-modal features such as visual characteristics, audio, and text. Then, the method in [10] learns multimodal embeddings between modalities such as text, video, and audio via deep canonical correlation analysis. The method in [11] has a multimodal variational autoencoder (MVAE) [12] including a fake news detection network using tweets and visual information, and the performance has been improved. Thus, MVAE is effective for our task, which is the detection of important scenes using tweets and videos. Generally, the MVAE can detect correlations across modalities by training shared representations across them. Since the MVAE can flexibly express the relationship between heterogeneous modalities, the construction of a highly accurate method for our tasks is expected by adopting the MVAE. However, to efficiently use tweets and videos, we need to consider the following problem. Since tweets posted on Twitter are influenced by multiple previous events in the videos, they are closely related to each other. Thus, there are time-lags between posted tweets and multiple corresponding events. Since these time-lags cannot be considered in previous MVAE-based approaches, the development of the MVAE-based detection method considering time-lags between posted tweets and multiple corresponding events would contribute to an improvement in performance.
In this paper, we propose a new method for the detection of important scenes in baseball videos via the MVAE considering time-lags between tweets and corresponding multiple previous events. The newly proposed method is the time-lag-aware MVAE (Tl-MVAE). To realize the detection using tweets and videos, tweet, video, and audio features extracted from tweets and videos are used in the Tl-MVAE. The Tl-MVAE consists of an encoder, a decoder, and an important scene detector. In the encoding architecture, multimodal features are transformed into latent features based on feature correlations considering time-lags. The decoder reconstructs the original transformed features from the latent features. Moreover, the important scene detector detects important scenes by using the latent features. Since feature transformation based on feature correlations considering such time-lags is introduced to the Tl-MVAE, the proposed method can derive latent features that are efficient for the consideration of the relationships between tweets and videos. From this novelty, the proposed method can realize accurate detection based on the Tl-MVAE using multimodal features extracted from tweets and videos.
It should be noted that this paper is an extended version of [13]. Specifically, we newly introduce the novel Tl-MVAE to the detection of important scenes.

2. Detection of Important Scenes via the TL-MVAE

The details of the new method for the detection of important scenes in baseball videos via the Tl-MVAE are shown in this section. As shown in Figure 1, the Tl-MVAE consists of an encoder (Section 2.1), a decoder (Section 2.2), and an important scene detector (Section 2.3). Each of them is explained in detail below. The final loss and the optimization of the parameters of the encoder, decoder, and important scene detector are also shown in Section 2.4.

2.1. Encoder

The encoder is explained in this subsection. The inputs of the encoder are tweets posted during baseball games and videos, and the outputs are the latent features of the features extracted from tweets and videos. Specifically, in feature extraction architectures, by using the i-th baseball video ( i = 1 , 2 , , I ; I being the number of videos) and corresponding tweets, features x i , j m R d x ( j = 1 , 2 , , J i ; J i being the number of tweets of the i-th video) are calculated. Note that m { t , v , a } is the modality in the feature extraction architecture. Consequently, feature matrices X i m = [ x i , 1 m , , x i , j m , , x i , J i m ] R d x × J i are obtained. The feature extraction architectures are explained in detail in Section 3.1.
The proposed network transforms extracted features with the consideration of time-lags between tweets and videos. Specifically, by transforming feature matrices X i m via multi-layered neural networks, Y i m = [ y i , 1 m , , y i , j m , , y i , J i m ] R d y × J i considering time-lags are output from the last layer of the neural network. Since posted tweets are related to corresponding multiple previous events rather than just one event, there exists a correlation between tweet features at a target time and corresponding video and audio features of multiple previous events. For example, the immediate previous event strongly influences the tweet, and the influence of past events tends to be gradually weakened. Thus, we assume that tweets are affected by events from the present to the past according to a Poisson distribution, which means that the number of events occurring in a fixed interval of time can be regarded as the degree of influence by baseball events as shown in Figure 2. From this assumption, we construct multi-layered neural networks with consideration of the time-lags between different features based on the influence defined from the Poisson distribution. In order to calculate Y i m considering these time-lags with the optimization of the parameters of the multi-layered neural networks, we maximize the average inter-set correlation [14] defined as:
ρ = 1 d y ( M 1 ) d = 1 d y ψ d R B ψ d ψ d R W ψ d ,
where ψ d R d y ( d = 1 , 2 , . . . , d y ) means the optimal weight common to all modalities. Furthermore, M ( = 3 ) means the number of modalities. Then, R B is the between-set covariance matrix considering the time-lags, and R W is the within-set covariance matrix. They are respectively defined as:
R B = i = 1 I m 1 { t , v , a } m 2 { t , v , a } , m 2 m 1 C ¯ i m 1 , m 2 ,
R W = i = 1 I m { t , v , a } C ̲ i m , m .
Note that we omit the same scaling value ( J i 1 ) 1 M 1 in the above equations. Furthermore,
C ¯ i m 1 , m 2 = l = 0 L 1 e λ λ l l ! Y ˜ i , 0 m 1 Y ˜ i , l m 2 l = 0 L 1 e λ λ l l ! ( m 1 { t } , m 2 { v , a } ) l = 0 L 1 e λ λ l l ! Y ˜ i , l m 1 Y ˜ i , 0 m 2 l = 0 L 1 e λ λ l l ! ( m 1 { v , a } , m 2 { t } ) Y ˜ i , 0 m 1 Y ˜ i , 0 m 2 ( otherwise ) ,
C ̲ i m , m = Y ˜ i , 0 m Y ˜ i , 0 m ,
where L determines the number of previous events affecting the target tweet. In addition, λ is a parameter of the Poisson distribution that controls the peak of the distribution. Note that λ is the same as the mean and the variance of the distribution. Then, feature matrices Y ˜ i , l m = [ y i , L l m , , y i , J i l m ] ( l = 0 , , L 1 ) are mean-normalized.
Finally, we calculate latent features from the transformed features y i , j m . Specifically, we pass y i , j m through multiple fully connected layers and obtain low-dimensional features u i , j m . The low-dimensional tweet features u i , j t , video features u i , j v , and audio features u i , j a are concatenated and passed through a fully connected layer to form the shared representation s i , j . Then, the mean μ i , j and the variance σ i , j of the distribution of the shared representation s i , j are obtained. By using random variables ϵ sampled from the Gaussian distribution, the latent features z i , j are defined as:
z i , j = μ i , j + σ i , j ϵ .
By denoting the network that outputs the latent features z i , j from the transformed features y i , j m as G enc ( y i , j m , θ enc ) , the latent features z i , j are denoted as:
z i , j = G enc ( y i , j m , θ enc ) ,
where θ enc means all of the parameters to be trained in the network that outputs the latent features z i , j from the transformed features y i , j m .
Introducing the consideration of time-lags between tweets and corresponding multiple previous events into the encoder is the biggest contribution. The encoder of a general MVAE cannot consider time-lags between different modalities. However, the Tl-MVAE can flexibly consider the characteristics of tweets and videos by the consideration of time-lags based on Equation (4).

2.2. Decoder

In this subsection, we explain the decoder that reconstructs the original data from the latent features. The decoder is an inverted network of the network that outputs the latent features from the transformed features. Specifically, the input of the decoder is the latent features z i , j obtained in Section 2.1, and it outputs reconstructed features y ^ i , j m . By denoting the decoder as G dec ( z i , j , θ dec ) , the reconstructed features are denoted as:
y ^ i , j m = G dec ( z i , j , θ dec ) ,
where θ dec means all of the parameters to be trained in the decoder.

2.3. Important Scene Detector

In this subsection, we explain the important scene detector consisting of multiple fully connected layers. The important scene detector takes the latent features obtained in Section 2.1 as the input and aims to classify the scene as important or normal. Specifically, latent features z i , j are input, and the probability p ^ i , j of the scene corresponding to z i , j being important is output. By denoting the important scene detector as G isd ( z i , j , θ isd ) , the probability p ^ i , j is defined as:
p ^ i , j = G isd ( z i , j , θ isd ) ,
where θ isd means all of the parameters to be trained in the important scene detector. When p ^ i , j > τ , the scene is determined to be an important scene. Note that we predetermine a threshold value τ .

2.4. Final Loss

In this subsection, we explain the final loss of the Tl-MVAE. We jointly train the encoder, decoder, and important scene detector. Thus, the final loss can be defined as:
L final ( θ enc , θ dec , θ isd ) = ( m { t , v , a } ξ rec m L rec m ) + ξ kl L kl + ξ isd L isd ,
where L rec m is the reconstruction loss between the outputs of the decoder and features considering time-lags obtained in the encoder. The reconstruction loss L rec m is defined as:
L rec m = 1 I i = 1 I { 1 J i d y j = 1 J i d = 1 d y ( y ^ i , j , d m y i , j , d m ) 2 } .
Note that y ^ i , j , d m and y i , j , d m are the d-th values of y ^ i , j m and y i , j m , respectively. Then, the minimization of the KL divergence L kl means optimizing the probability distribution parameters ( μ and σ ) to closely resemble those of the target distribution (Gaussian distribution). The KL divergence L kl is defined as:
L kl = 1 I i = 1 I { 1 2 J i j = 1 J i d = 1 d z ( μ i , j , d 2 + σ i , j , d 2 log ( σ i , j , d ) 1 ) } ,
where μ i , j , d and σ i , j , d are the d-th values of μ i , j and σ i , j , respectively. Moreover, L isd , which is the detection loss of important scenes, is defined as:
L isd = 1 I i = 1 I { 1 J i j = 1 J i ( l i , j log ( l ^ i , j ) + ( 1 l i , j ) log ( 1 l ^ i , j ) ) } ,
where l ^ i , j is the label calculated on the basis of the probability p ^ i , j and l i , j is the ground truth label. Then, ξ rec m , ξ kl , and ξ isd are parameters to balance the individual terms of the loss function. By minimizing the final loss, we can calculate the optimal parameters.
Our biggest contribution is the development of a method for the detection of important scenes in baseball videos via the Tl-MVAE, which can consider time-lags between tweets and their corresponding multiple previous events. The Tl-MVAE assumes that tweet features are related to the previous video and audio features, and feature transformation considering these characteristics is introduced into the encoder of the Tl-MVAE. Then, by using the latent features calculated from the features transformed on the basis of feature correlations considering time-lags, the accurate detection of important scenes can be realized. This is the novel idea of the Tl-MVAE.

3. Experimental Results

3.1. Experimental Setting

As datasets, twelve baseball videos (30 fps) and tweets posted during those games were used. The baseball videos were broadcast from 13 June to 27 September 2019 by Pacific League TV according to our previous experiments [13,15]. The details of these videos are shown in Table 1. By using the query “#lovefighters”, which is an official hashtag of the professional baseball team, we collected tweets for this experiment. We used seven games for the training of the Tl-MVAE and the other five games for the test. The previous works [3] and [4] used 6 and 10 games for the experiments. Moreover, since the proposed method adopts not only videos, but also tweets, more information is available in the proposed method than in the previous works. Thus, the number of games used in this experiment is considered to be sufficient. Furthermore, since tweets related to each experiment were collected on their own in previous works [7,8], we also followed their schemes in the same manner and verified the performance of the Tl-MVAE in this experiment. In the feature extraction architecture, as the tweet feature extraction networks described in Section 2.1, we adopted Tweet2Vec [16], which is one of the common language models. Tweet2Vec contains a bi-directional gated recurrent unit [17], a linear layer, and a softmax layer. Tweet2Vec used in the Tl-MVAE was trained by using tweets collected on the basis of 27 hashtags related to professional baseball teams. In order to extract the video features, we used 3D ResNet [18], which was pre-trained on the Kinetics dataset [19], as the video feature extraction network described in Section 2.1. 3D ResNet includes 17 convolutional blocks, a global average pooling layer, a fully connected layer, and a softmax layer. In addition, VGG16 [20], which was pre-trained with the ImageNet dataset [21], was used as the audio feature extraction network described in Section 2.1. VGG16 consists of 13 convolutional layers, five max layers, three fully connected layers, and a softmax layer. Generally, VGG16 is used for feature extraction from images. However, since spectrogram-based feature calculation based on a pre-trained CNN model leads to effective representation of audio data [22,23], we extracted audio features from spectrograms of audio based on the pretrained VGG16. Furthermore, d x , d y , L, λ ξ rec m , ξ kl , ξ isd , and τ were set to 500, 256, 12, 3, 1, 1, 1, and 0.5, respectively. d x , d y , L, λ , and τ were set to the values at which the proposed method had the highest specificity. In addition, ξ rec m , ξ kl , and ξ isd were set to 1 according to [11].
In order to confirm the validity of introducing the tweet, video, and audio features, the following comparative methods (CMs1-7) were adopted. In addition, we adopted CMs8 and 9 to confirm the validity of the consideration of time-lags between tweets and corresponding multiple previous events. Furthermore, CMs10 and 11 were used to confirm the validity of using the MVAE as the model for the detection of important scenes. We used CM12 for the comparison with our previous method.
CMs1-6: Methods using features shown in Table 2. Note that CMs4 and 5 consider time-lags.
CM7: A method based on the simple integration of detection models constructed for each modality. This approach determines important scenes by majority voting of the detection results based on CMs1-3.
CM8: A method not considering time-lags based on the MVAE [11].
CM9: A method based on [8]. This approach considers time-lags between tweets and only one corresponding previous event.
CM10: A method [15] that flexibly expresses the relationship between tweets and videos based on time-lag-aware deep multiset canonical correlation analysis (Tl-dMCCA). This approach considers time-lags between tweets and corresponding multiple previous events.
CM11: A method using a long short-term memory (LSTM) [24], which can maintain a long-term memory, which is effective for training time series features calculated from tweets and videos. In the same manner as the Tl-MVAE, the inputs are tweets and videos, and the outputs are the probabilities that the scenes are important. This approach can consider time-lags.
CM12: A method using LSTM based on our previous method [13]. The inputs of LSTM are features transformed based on dMCCA, and the outputs are the probabilities that the scenes are important. This approach cannot consider time-lags.
In this experiment, each at-bat (an at-bat is a period from when the batter enters the batter box until he/she is out or reaches the bases) was regarded as an important scene when at least 80% of its length is detected as important. Since the specificity of the proposed method is the highest when this percentage of length is 80%, we adopted this percentage. As a quantitative evaluation index, we adopted specificity when sensitivity was almost 1.0 (i.e., maximizing sensitivity). Specificity when sensitivity is almost 1.0 means how much over-detection of normal scenes can be reduced when important scenes are detected. The detection of all important scenes is needed to maximize sensitivity. Thus, the over-detection of normal scenes needs to be close to 0.0 to obtain high specificity when sensitivity is almost 1.0. Since that is a difficult task, the specificity when sensitivity is almost 1.0 does not become a very large value in this experiment. Note that we defined pre-edited highlights broadcast by Pacific League TV as important scenes, i.e., ground truth. Other scenes are defined as normal scenes.

3.2. Performance Evaluation

The specificity of the proposed method (PM) and CMs1-12 are shown in Table 3. Since the PM has higher specificity than the specificity of CMs1-6, we can confirm the effectiveness of introducing tweet, video, and audio features. Furthermore, since the specificity of the PM is higher than that of CM7, simply utilizing these features does not necessarily improve detection performance. Thus, we can see that the consideration of the relationships between tweets and videos is effective for the detection of important scenes. Since the specificity of CM9 is higher than that of CM8, the effectiveness of the consideration of time-lags can be confirmed. When the t-test was performed on the probabilities calculated by CM8 and CM9, the p-value was 0.046, so there is a significant difference in this result. Furthermore, the PM has higher specificity than that of CM9. However, since the results of the PM and CM9 are very close, we compared these two methods by using the F-measure, which is the harmonic mean of recall and precision. As a result, the F-measure of the PM and CM9 were 0.416 and 0.404. Thus, we can confirm that the consideration of time-lags between tweets and corresponding multiple previous events is effective. Since the PM outperforms CMs10 and 11, feature transformation considering time-lags is effective for the detection of important scenes. In addition, the PM has higher specificity than that of CM12, which is our previous method. From the results of the PM and CMs1-12, accurate detection of important scenes of baseball videos via the Tl-MVAE considering time-lags between tweets and corresponding multiple previous events is realized.
Next, two examples of important scenes detected by our method (PM) and their corresponding tweets are shown in Figure 3. The tweets correspond to the scenes surrounded by a rectangle of the same color. The important scene surrounded by the red rectangle, a scene in which the batter takes a chance, was detected by the PM. Interestingly, the contents of tweets for the scene surrounded by the red rectangle indicate the enthusiasm and expectations of the viewers for this scene. In addition, the important scene surrounded by the blue rectangle is a scene in which the batter hits a home run. The contents of tweets for the scene surrounded by the blue rectangle include the delight and excitement of the viewers watching this important scene. These tweets were posted after the corresponding important scenes. From these results, since there obviously exist time-lags between tweets and their corresponding events, we can confirm that the PM accurately detects these important scenes from this qualitative evaluation.
Finally, we show the results of the Tl-MVAE obtained when parameters λ and L of the Poisson distribution are changed in Table 4. The results of each parameter λ express how much the peak of the distribution should be slid. Thus, we can confirm when tweets are posted after corresponding events have occurred. Furthermore, the results obtained when L is changed express how much previous events affect the tweets. From the specificity of each parameter, we can confirm that the highest specificity is achieved when λ is three and L is 12. In addition, the tweet for the test is posted about every 24 s on average. Therefore, these results suggest that the time-lag between the tweet and its most related event is about 72 s, and events up to 288 s in the past affect the tweets. We therefore consider that the calculation based on the parameters λ and L of the distribution is effective for revealing the time-lags.

4. Conclusions

In this paper, a new method for the detection of important scenes in baseball videos via the Tl-MVAE is presented. By introducing the consideration of time-lags, the Tl-MVAE can correctly consider the relationships between tweets and videos. Experimental results confirm that the Tl-MVAE realizes the accurate detection of important scenes of baseball videos. Since the immediately previous event strongly influences the tweet and the influence of past events tends to be gradually weakened, we adopt a Poisson distribution for the consideration of time-lags. In future work, we will use this distribution and other distributions in order to verify the optimal distribution for the consideration of time-lags.

Author Contributions

Conceptualization, K.H., K.M., T.O., and M.H.; methodology, K.H., K.M., T.O., and M.H.; software, K.H.; validation, K.H., K.M., T.O., and M.H.; data curation, K.H.; writing—original draft preparation, K.H.; writing—review and editing, K.M., T.O, and M.H; visualization, K.H.; funding acquisition, T.O. and M.H. All authors read and agreed to the published version of the manuscript.

Funding

This work was partly supported by the MIC/SCOPE #181601001.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shukla, P.; Sadana, H.; Bansal, A.; Verma, D.; Elmadjian, C.; Raman, B.; Turk, M. Automatic cricket highlight generation using event-driven and excitement-based features. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1800–1808. [Google Scholar]
  2. Cheng, C.C.; Hsu, C.T. Fusion of audio and motion information on HMM-based highlight extraction for baseball games. IEEE Trans. Multimed. 2006, 8, 585–599. [Google Scholar] [CrossRef]
  3. Chang, P.; Han, M.; Gong, Y. Extract highlights from baseball game video with hidden Markov models. In Proceedings of the International Conference on Image Processing, Rochester, NY, USA, 22–25 September 2002; Volume 1, pp. 609–612. [Google Scholar]
  4. Gong, Y.; Han, M.; Hua, W.; Xu, W. Maximum entropy model-based baseball highlight detection and classification. Comput. Vis. Image Underst. 2004, 96, 181–199. [Google Scholar] [CrossRef]
  5. Nakazawa, M.; Erdmann, M.; Hoashi, K.; Ono, C. Social indexing of TV programs: Detection and labeling of significant TV scenes by Twitter analysis. In Proceedings of the International Conference on Advanced Information Networking and Applications Workshops, Fukuoka, Japan, 26–29 March 2012; pp. 141–146. [Google Scholar]
  6. Doman, K.; Tomita, T.; Ide, I.; Deguchi, D.; Murase, H. Event detection based on twitter enthusiasm degree for generating a sports highlight video. In Proceedings of the ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 949–952. [Google Scholar]
  7. Hsieh, L.C.; Lee, C.W.; Chiu, T.H.; Hsu, W. Live semantic sport highlight detection based on analyzing tweets of twitter. In Proceedings of the IEEE International Conference on Multimedia and Expo, Melbourne, Australia, 9–13 July 2012; pp. 949–954. [Google Scholar]
  8. Jai-Andaloussi, S.; Mohamed, A.; Madrane, N.; Sekkaki, A. Soccer video summarization using video content analysis and social media streams. In Proceedings of the IEEE/ACM International Symposium on Big Data Computing, Limassol, Cyprus, 7–9 July 2014; pp. 1–7. [Google Scholar]
  9. Mithun, N.C.; Li, J.; Metze, F.; Roy-Chowdhury, A.K. Learning joint embedding with multimodal cues for cross-modal video-text retrieval. In Proceedings of the ACM on International Conference on Multimedia Retrieval, Amsterdam, The Netherlands, 27–30 May 2018; pp. 19–27. [Google Scholar]
  10. Sun, Z.; Sarma, P.; Sethares, W.; Liang, Y. Learning relationships between text, audio, and video via deep canonical correlation for multimodal language analysis. In Proceedings of the AAAI Conf. Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 8992–8999. [Google Scholar]
  11. Khattar, D.; Goud, J.S.; Gupta, M.; Varma, V. MVAE: Multimodal variational autoencoder for fake news detection. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2915–2921. [Google Scholar]
  12. Suzuki, M.; Nakayama, K.; Matsuo, Y. Joint multimodal learning with deep generative models. arXiv 2016, arXiv:1611.01891. [Google Scholar]
  13. Hirasawa, K.; Maeda, K.; Ogawa, T.; Haseyama, M. Important scene detection based on anomaly detection using long short-term memory for baseball highlight generation. In Proceedings of the IEEE International Conference on Consumer Electronics, Taoyuan, Taiwan, 28–30 September 2020; pp. 1–2. [Google Scholar]
  14. Parra, L.C.; Haufe, S.; Dmochowski, J.P. Correlated components analysis-extracting reliable dimensions in multivariate data. arXiv 2018, arXiv:1801.08881. [Google Scholar]
  15. Hirasawa, K.; Maeda, K.; Ogawa, T.; Haseyama, M. Important scene detection of baseball videos via time-lag-aware deep multiset canonical correlation maximization. In Proceedings of the IEEE International Conference on Image Processing, Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 1236–1240. [Google Scholar]
  16. Dhingra, B.; Zhou, Z.; Fitzpatrick, D.; Muehl, M.; Cohen, W.W. Tweet2Vec: Character-based distributed representations for social media. arXiv 2016, arXiv:1605.03481. [Google Scholar]
  17. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  18. Hara, K.; Kataoka, H.; Satoh, Y. Can spatiotemporal 3D CNNs retrace the history of 2D CNNs and imagenet? In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6546–6555. [Google Scholar]
  19. Kay, W.; Carreira, J.; Simonyan, K.; Zhang, B.; Hillier, C.; Vijayanarasimhan, S.; Viola, F.; Green, T.; Back, T.; Natsev, P.; et al. The kinetics human action video dataset. arXiv 2017, arXiv:1705.06950. [Google Scholar]
  20. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  21. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–5 December 2012; pp. 1097–1105. [Google Scholar]
  22. Al Hadhrami, E.; Al Mufti, M.; Taha, B.; Werghi, N. Ground moving radar targets classification based on spectrogram images using convolutional neural networks. In Proceedings of the International Radar Symposium, Bonn, Germany, 20–22 June 2018; pp. 1–9. [Google Scholar]
  23. Amiriparian, S.; Gerczuk, M.; Ottl, S.; Cummins, N.; Freitag, M.; Pugachevskiy, S.; Baird, A.; Schuller, B.W. Snore sound classification using image-based deep spectrum features. In Proceedings of the Interspeech, Stockholm, Sweeden, 20–24 August 2017; pp. 3512–3516. [Google Scholar]
  24. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Outline of the time-lag-aware multimodal variational autoencoder (Tl-MVAE). The Tl-MVAE encodes tweet, video, and audio features into latent features based on feature correlations with the consideration of time-lags, as described in Section 2.1. Furthermore, the Tl-MVAE reconstructs the original features from the latent features as described in Section 2.2. Finally, the Tl-MVAE detects important scenes by using the latent features as described in Section 2.3.
Figure 1. Outline of the time-lag-aware multimodal variational autoencoder (Tl-MVAE). The Tl-MVAE encodes tweet, video, and audio features into latent features based on feature correlations with the consideration of time-lags, as described in Section 2.1. Furthermore, the Tl-MVAE reconstructs the original features from the latent features as described in Section 2.2. Finally, the Tl-MVAE detects important scenes by using the latent features as described in Section 2.3.
Sensors 21 02045 g001
Figure 2. The relationship between a tweet and its corresponding multiple previous events. The tweet corresponds to events from the present to the past. Furthermore, these events are weighted by using the degree of influence defined on the basis of the Poisson distribution.
Figure 2. The relationship between a tweet and its corresponding multiple previous events. The tweet corresponds to events from the present to the past. Furthermore, these events are weighted by using the degree of influence defined on the basis of the Poisson distribution.
Sensors 21 02045 g002
Figure 3. Two examples of important scenes detected by the Tl-MVAE and tweets corresponding to these scenes. The horizontal axis means time. The timing when important scenes occur and the timing when the corresponding tweets are posted are different.
Figure 3. Two examples of important scenes detected by the Tl-MVAE and tweets corresponding to these scenes. The horizontal axis means time. The timing when important scenes occur and the timing when the corresponding tweets are posted are different.
Sensors 21 02045 g003
Table 1. Details of the baseball video dataset used in our experiment.
Table 1. Details of the baseball video dataset used in our experiment.
GamesNumber of Important ScenesAverage Time Length of Important ScenesGame Time
1122 min 28 s3 h 23 min
2142 min 31 s2 h 59 min
3222 min 03 s3h 32 min
4181 min 58 s3 h 08 min
5111 min 40 s2 h 44 min
Table 2. Features used in Comparative Methods (CMs) 1-6.
Table 2. Features used in Comparative Methods (CMs) 1-6.
FeaturesCM1CM2CM3CM4CM5CM6
Tweet
Video
Audio
Table 3. Specificity of the detection of important scenes in the proposed method (PM) and CMs1-12.
Table 3. Specificity of the detection of important scenes in the proposed method (PM) and CMs1-12.
GamesPMCM1CM2CM3CM4CM5CM6CM7CM8 [11]CM9 [8]CM10 [15]CM11 [24]CM12 [13]
10.3770.3500.3110.3110.3500.3450.3500.3450.3700.3770.3600.3700.370
20.4540.3630.3630.3900.3900.4060.4060.3630.4360.4360.4060.4180.418
30.4720.4490.4620.4150.4490.4490.4530.4530.4530.4710.4620.4630.453
40.3200.3010.2970.2970.3010.2970.3010.2970.2970.3200.3080.3020.297
50.4230.3650.3460.3460.4090.4020.4090.3650.4230.4040.4210.4130.409
Average0.4090.3660.3560.3520.3800.3800.3840.3650.3960.4020.3910.3930.389
Table 4. Average specificity of the Tl-MVAE for all games when parameters the λ and L of the Poisson distribution are changed.
Table 4. Average specificity of the Tl-MVAE for all games when parameters the λ and L of the Poisson distribution are changed.
L = 10 L = 12 L = 14 L = 16 L = 18 L = 20 Average
λ = 1 0.3960.4000.3910.3870.3910.3750.390
λ = 3 0.4030.4090.3970.3920.3870.3810.395
λ = 5 0.3910.3970.3860.3810.3750.3740.384
λ = 7 0.3770.3860.3690.3640.3650.3630.371
λ = 9 0.3640.3750.3730.3650.3570.3470.364
Average0.3860.3930.3830.3780.3750.3680.381
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hirasawa, K.; Maeda, K.; Ogawa, T.; Haseyama, M. Detection of Important Scenes in Baseball Videos via a Time-Lag-Aware Multimodal Variational Autoencoder. Sensors 2021, 21, 2045. https://doi.org/10.3390/s21062045

AMA Style

Hirasawa K, Maeda K, Ogawa T, Haseyama M. Detection of Important Scenes in Baseball Videos via a Time-Lag-Aware Multimodal Variational Autoencoder. Sensors. 2021; 21(6):2045. https://doi.org/10.3390/s21062045

Chicago/Turabian Style

Hirasawa, Kaito, Keisuke Maeda, Takahiro Ogawa, and Miki Haseyama. 2021. "Detection of Important Scenes in Baseball Videos via a Time-Lag-Aware Multimodal Variational Autoencoder" Sensors 21, no. 6: 2045. https://doi.org/10.3390/s21062045

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop