1 code implementation • 18 Mar 2024 • Mingyang Song, Mao Zheng, Xuan Luo
While recent research endeavors have concentrated on developing Large Language Models (LLMs) with robust long-context capabilities, due to the lack of appropriate evaluation strategies, relatively little is known about how well the long-context capability and performance of leading LLMs (e. g., GPT-4 Turbo and Kimi Chat).
no code implementations • 20 Feb 2023 • Weihong Zhong, Mao Zheng, Duyu Tang, Xuan Luo, Heng Gong, Xiaocheng Feng, Bing Qin
Although large-scale video-language pre-training models, which usually build a global alignment between the video and the text, have achieved remarkable progress on various downstream tasks, the idea of adopting fine-grained information during the pre-training stage is not well explored.
1 code implementation • CVPR 2022 • Shi Pu, Kaili Zhao, Mao Zheng
Further, we synthesize features of unseen classes by proposing a class generator that interpolates and extrapolates the features of seen classes.
Ranked #14 on Zero-Shot Action Recognition on UCF101
no code implementations • 26 Oct 2020 • Shi Pu, Yijiang He, Zheng Li, Mao Zheng
Existing video recommendation systems directly exploit features from different modalities (e. g., user personal data, user behavior data, video titles, video tags, and visual contents) to input deep neural networks, while expecting the networks to online mine user-preferred topics implicitly from these features.