Unsupervised Video Summarization
16 papers with code • 2 benchmarks • 3 datasets
Unsupervised video summarization approaches overcome the need for ground-truth data (whose production requires time-demanding and laborious manual annotation procedures), based on learning mechanisms that require only an adequately large collection of original videos for their training. Specifically, the training is based on heuristic rules, like the sparsity, the representativeness, and the diversity of the utilized input features/characteristics.
Most implemented papers
Summarizing Videos using Concentrated Attention and Considering the Uniqueness and Diversity of the Video Frames
Instead of simply modeling the frames' dependencies based on global attention, our method integrates a concentrated attention mechanism that is able to focus on non-overlapping blocks in the main diagonal of the attention matrix, and to enrich the existing information by extracting and exploiting knowledge about the uniqueness and diversity of the associated frames of the video.
Contrastive Losses Are Natural Criteria for Unsupervised Video Summarization
Video summarization aims to select the most informative subset of frames in a video to facilitate efficient video browsing.
Adopting Self-Supervised Learning into Unsupervised Video Summarization through Restorative Score
We show that the reconstruction loss of the model for a video with masked frames correlates with the representativeness of the remaining frames in the video.
Adopting Self-Supervised Learning into Unsupervised Video Summarization through Restorative Score.
We show that the reconstruction loss of the model for a video with masked frames correlates with the representativeness of the remaining frames in the video.
Enhancing Video Summarization with Context Awareness
Despite the importance of video summarization, there is a lack of diverse and representative datasets, hindering comprehensive evaluation and benchmarking of algorithms.
Cluster-based Video Summarization with Temporal Context Awareness
In this paper, we present TAC-SUM, a novel and efficient training-free approach for video summarization that addresses the limitations of existing cluster-based models by incorporating temporal context.