Highlight Detection
27 papers with code • 3 benchmarks • 2 datasets
Libraries
Use these libraries to find Highlight Detection models and implementationsMost implemented papers
Text-Aware Single Image Specular Highlight Removal
The core goal is to improve the accuracy of text detection and recognition by removing the highlight from text images.
Cross-category Video Highlight Detection via Set-based Learning
For tackling such practical problem, we propose a Dual-Learner-based Video Highlight Detection (DL-VHD) framework.
Single-Image Specular Highlight Removal via Real-World Dataset Construction
Specular reflections pose great challenges on various multimedia and computer vision tasks, e. g. , image segmentation, detection and matching.
UMT: Unified Multi-modal Transformers for Joint Video Moment Retrieval and Highlight Detection
Finding relevant moments and highlights in videos according to natural language queries is a natural and highly valuable common need in the current video content explosion era.
M2-Net: Multi-stages Specular Highlight Detection and Removal in Multi-scenes
The framework consists of three main components, highlight feature extractor module, highlight coarse removal module, and highlight refine removal module.
SpecSeg Network for Specular Highlight Detection and Segmentation in Real-World Images
Specular highlights detection and removal in images is a fundamental yet non-trivial problem of interest.
Query-Dependent Video Representation for Moment Retrieval and Highlight Detection
As we observe the insignificant role of a given query in transformer architectures, our encoding module starts with cross-attention layers to explicitly inject the context of text query into video representation.
Collaborative Noisy Label Cleaner: Learning Scene-aware Trailers for Multi-modal Highlight Detection in Movies
Based on existing efforts, this work has two observations: (1) For different annotators, labeling highlight has uncertainty, which leads to inaccurate and time-consuming annotations.
Joint Moment Retrieval and Highlight Detection Via Natural Language Queries
Video summarization has become an increasingly important task in the field of computer vision due to the vast amount of video content available on the internet.
UniVTG: Towards Unified Video-Language Temporal Grounding
Most methods in this direction develop taskspecific models that are trained with type-specific labels, such as moment retrieval (time interval) and highlight detection (worthiness curve), which limits their abilities to generalize to various VTG tasks and labels.