Next Article in Journal
GMDNet: Grouped Encoder-Mixer-Decoder Architecture Based on the Role of Modalities for Brain Tumor MRI Image Segmentation
Previous Article in Journal
Enhanced Power Sharing Control of an Islanded DC Microgrid with Unmatched Line Impedances
Previous Article in Special Issue
Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Temporal Map-Based Boundary Refinement Network for Video Moment Localization

1
College of Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China
2
School of Artificial Intelligence, Anhui University, Hefei 230039, China
3
College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China
*
Authors to whom correspondence should be addressed.
Electronics 2025, 14(8), 1657; https://doi.org/10.3390/electronics14081657
Submission received: 29 March 2025 / Revised: 16 April 2025 / Accepted: 17 April 2025 / Published: 19 April 2025
(This article belongs to the Special Issue Big Model Techniques for Image Processing)

Abstract

Video moment localization has gradually become a hot research problem in video understanding. Despite much remarkable progress, there are still two limitations in the following aspects. Firstly, previous methods usually directly regarded the moment candidate with the highest confidence score as the final localization result; they overlooked the inevitable deviations between moment candidates and the ground truth. Secondly, past models have not considered the problem that moment candidates with various qualities have different impacts on model training. Therefore, this paper proposes a novel Temporal Map-based Boundary Refinement Network to solve the above problems. Specifically, besides the conventional confidence scores’ prediction network, we introduce a boundary refinement network based on the 2D temporal map, which can fine-tune the temporal boundaries of generated moment candidates to obtain more precise results. Additionally, to discriminately treat diverse moment candidates with different qualities and further boost the localization performance, we devise an innovative weighted refinement loss to guide the model to focus more on refining processes of those moment candidates closer to the ground truth. Finally, we evaluate our model on two publicly available datasets, and the results of extensive experiments show that our technique outperforms the state-of-the-art methods (e.g., for the ActivityNet Captions dataset, a relative improvement of 4.63% on R1@0.7).
Keywords: video moment localization; temporal map; boundary refinement network; weighted refinement loss video moment localization; temporal map; boundary refinement network; weighted refinement loss

Share and Cite

MDPI and ACS Style

Lyu, L.; Liu, D.; Zhang, C.; Zhang, Y.; Ruan, H.; Zhu, L. Temporal Map-Based Boundary Refinement Network for Video Moment Localization. Electronics 2025, 14, 1657. https://doi.org/10.3390/electronics14081657

AMA Style

Lyu L, Liu D, Zhang C, Zhang Y, Ruan H, Zhu L. Temporal Map-Based Boundary Refinement Network for Video Moment Localization. Electronics. 2025; 14(8):1657. https://doi.org/10.3390/electronics14081657

Chicago/Turabian Style

Lyu, Liang, Deyin Liu, Chengyuan Zhang, Yilin Zhang, Haoyu Ruan, and Lei Zhu. 2025. "Temporal Map-Based Boundary Refinement Network for Video Moment Localization" Electronics 14, no. 8: 1657. https://doi.org/10.3390/electronics14081657

APA Style

Lyu, L., Liu, D., Zhang, C., Zhang, Y., Ruan, H., & Zhu, L. (2025). Temporal Map-Based Boundary Refinement Network for Video Moment Localization. Electronics, 14(8), 1657. https://doi.org/10.3390/electronics14081657

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop