Search Results for author: Guozhang Li

Found 3 papers, 2 papers with code

EtC: Temporal Boundary Expand then Clarify for Weakly Supervised Video Grounding with Multimodal Large Language Model

no code implementations5 Dec 2023 Guozhang Li, Xinpeng Ding, De Cheng, Jie Li, Nannan Wang, Xinbo Gao

To further clarify the noise of expanded boundaries, we combine mutual learning with a tailored proposal-level contrastive objective to use a learnable approach to harmonize a balance between incomplete yet clean (initial) and comprehensive yet noisy (expanded) boundaries for more precise ones.

Boundary Detection Language Modelling +2

Boosting Weakly-Supervised Temporal Action Localization with Text Information

1 code implementation CVPR 2023 Guozhang Li, De Cheng, Xinpeng Ding, Nannan Wang, Xiaoyu Wang, Xinbo Gao

For the discriminative objective, we propose a Text-Segment Mining (TSM) mechanism, which constructs a text description based on the action class label, and regards the text as the query to mine all class-related segments.

Sentence Weakly-supervised Temporal Action Localization +1

Weakly-Supervised Temporal Action Localization with Bidirectional Semantic Consistency Constraint

1 code implementation25 Apr 2023 Guozhang Li, De Cheng, Xinpeng Ding, Nannan Wang, Jie Li, Xinbo Gao

The proposed Bi-SCC firstly adopts a temporal context augmentation to generate an augmented video that breaks the correlation between positive actions and their co-scene actions in the inter-video; Then, a semantic consistency constraint (SCC) is used to enforce the predictions of the original video and augmented video to be consistent, hence suppressing the co-scene actions.

Weakly-supervised Temporal Action Localization Weakly Supervised Temporal Action Localization

Cannot find the paper you are looking for? You can Submit a new open access paper.