no code implementations • 5 Jun 2023 • Han Xie, Da Zheng, Jun Ma, Houyu Zhang, Vassilis N. Ioannidis, Xiang Song, Qing Ping, Sheng Wang, Carl Yang, Yi Xu, Belinda Zeng, Trishul Chilimbi
Model pre-training on large text corpora has been demonstrated effective for various downstream applications in the NLP domain.
no code implementations • CVPR 2023 • Qian Jiang, Changyou Chen, Han Zhao, Liqun Chen, Qing Ping, Son Dinh Tran, Yi Xu, Belinda Zeng, Trishul Chilimbi
Hence we advocate that the key of better performance lies in meaningful latent modality structures instead of perfect modality alignment.
no code implementations • 22 Apr 2022 • Yubo Zhang, Feiyang Niu, Qing Ping, Govind Thattai
To solve video-and-language grounding tasks, the key is for the network to understand the connection between the two modalities.
no code implementations • 15 Feb 2022 • Cristian-Paul Bara, Qing Ping, Abhinav Mathur, Govind Thattai, Rohith MV, Gaurav S. Sukhatme
We introduce a novel privacy-preserving methodology for performing Visual Question Answering on the edge.
no code implementations • 14 Jan 2022 • Feng Gao, Qing Ping, Govind Thattai, Aishwarya Reganti, Ying Nian Wu, Prem Natarajan
Outside-knowledge visual question answering (OK-VQA) requires the agent to comprehend the image, make use of relevant knowledge from the entire web, and digest all the information to answer the question.
no code implementations • CVPR 2022 • Feng Gao, Qing Ping, Govind Thattai, Aishwarya Reganti, Ying Nian Wu, Prem Natarajan
Most previous works address the problem by first fusing the image and question in the multi-modal space, which is inflexible for further fusion with a vast amount of external knowledge.
Ranked #19 on Visual Question Answering (VQA) on OK-VQA
1 code implementation • CVPR 2021 • Tao Tu, Qing Ping, Govind Thattai, Gokhan Tur, Prem Natarajan
Most existing work for Guesser encode the dialog history as a whole and train the Guesser models from scratch on the GuessWhat?!
no code implementations • 2 Dec 2020 • Qing Ping, Feiyang Niu, Govind Thattai, Joel Chengottusseriyil, Qiaozi Gao, Aishwarya Reganti, Prashanth Rajagopal, Gokhan Tur, Dilek Hakkani-Tur, Prem Nataraja
Current conversational AI systems aim to understand a set of pre-designed requests and execute related actions, which limits them to evolve naturally and adapt based on human interactions.
no code implementations • 30 Jan 2020 • Jiangbo Yuan, Bing Wu, Wanying Ding, Qing Ping, Zhendong Yu
We conduct the learning in an adversarial learning process, which bears a close resemblance to the original GAN but again shifts the learning from image spaces to prior and latent code spaces.
no code implementations • 25 Dec 2019 • Qing Ping, Chaomei Chen
With the attention weights learned from the mutual-attention layer, final representations of a user and an item absorb information from both itself and its counterparts for making rating prediction.
4 code implementations • 16 Apr 2019 • Qing Ping, Bing Wu, Wanying Ding, Jiangbo Yuan
In this paper, we introduce attribute-aware fashion-editing, a novel task, to the fashion domain.
no code implementations • WS 2017 • Qing Ping, Chaomei Chen
With the prevalence of video sharing, there are increasing demands for automatic video digestion such as highlight detection.
1 code implementation • 7 Aug 2017 • Qing Ping, Chaomei Chen
With the prevalence of video sharing, there are increasing demands for automatic video digestion such as highlight detection.
no code implementations • 7 Aug 2017 • Qing Ping, Chaomei Chen
Such storylines help to preserve the semantic structure and logical thinking process of a scientific paper.