no code implementations • 22 Feb 2023 • Tingting Qiao, Gonzalo Moro Perez
We also interpret our models by visualizing the word importance given by the trained model, which indicates the model is capable to extract high-level semantic information of input documents.
2 code implementations • 14 May 2020 • Xinsheng Wang, Tingting Qiao, Jihua Zhu, Alan Hanjalic, Odette Scharenborg
An estimated half of the world's languages do not have a written form, making it impossible for these languages to benefit from any existing text-based technologies.
1 code implementation • NeurIPS 2019 • Tingting Qiao, Jing Zhang, Duanqing Xu, DaCheng Tao
Given a text description, we immediately imagine an overall visual impression using this prior and, based on this, we draw a picture by progressively adding more and more details.
2 code implementations • CVPR 2019 • Tingting Qiao, Jing Zhang, Duanqing Xu, DaCheng Tao
Generating an image from a given text description has two goals: visual realism and semantic consistency.
Ranked #8 on Text-to-Image Generation on CUB (Inception score metric)
1 code implementation • 2 Jan 2019 • Tingting Qiao, Weijing Zhang, Miao Zhang, Zixuan Ma, Duanqing Xu
By doing so, the ancient painting processing problems become natural image processing problems and models trained on natural images can be directly applied to the transferred paintings.
no code implementations • 19 Sep 2017 • Tingting Qiao, Jianfeng Dong, Duanqing Xu
Since there is a lack of human attention data, we first propose a Human Attention Network (HAN) to generate human-like attention maps, training on a recently released dataset called Human ATtention Dataset (VQA-HAT).