학술논문

GPTSee: Enhancing Moment Retrieval and Highlight Detection via Description-Based Similarity Features
Document Type
Periodical
Source
IEEE Signal Processing Letters IEEE Signal Process. Lett. Signal Processing Letters, IEEE. 31:521-525 2024
Subject
Signal Processing and Analysis
Computing and Processing
Communication, Networking and Broadcast Technologies
Semantics
Visualization
Feature extraction
Task analysis
Decoding
Natural languages
Computational modeling
Image description
semantic similarity
video moment retrieval
video highlight detection
Language
ISSN
1070-9908
1558-2361
Abstract
Moment retrieval (MR) and highlight detection (HD) aim to identify relevant moments and highlights in video from corresponding natural language query. Large language models (LLMs) have demonstrated proficiency in various computer vision tasks. However, existing methods for MR&HD have not yet been integrated with LLMs. In this letter, we propose a novel two-stage model that takes the output of LLMs as the input to the second-stage transformer encoder-decoder. First, MiniGPT-4 is employed to generate the detailed description of the video frame and rewrite the query statement, fed into the encoder as new features. Then, semantic similarity is computed between the generated description and the rewritten queries. Finally, continuous high-similarity video frames are converted into span anchors, serving as prior position information for the decoder. Experiments demonstrate that our approach achieves a state-of-the-art result, and by using only span anchors and similarity scores as outputs, positioning accuracy outperforms traditional methods, like Moment-DETR.