no code implementations • Findings (ACL) 2022 • Shuai Zhang, Wang Lijie, Xinyan Xiao, Hua Wu
Syntactic information has been proved to be useful for transformer-based pre-trained language models.
1 code implementation • EMNLP 2021 • Moye Chen, Wei Li, Jiachen Liu, Xinyan Xiao, Hua Wu, Haifeng Wang
Comparing with traditional methods, our method has two main advantages: (1) the relations between sentences are captured by modeling both the graph structure of the whole document set and the candidate sub-graphs; (2) directly outputs an integrate summary in the form of sub-graph which is more informative and coherent.
no code implementations • EMNLP 2020 • Yunjie Ji, Hao liu, Bolei He, Xinyan Xiao, Hua Wu, Yanhua Yu
To this end, we propose a novel Diversified Multiple Instance Learning Network (D-MILN), which is able to achieve aspect-level sentiment classification with only document-level weak supervision.
no code implementations • 6 Oct 2024 • Wenbo Li, Guohao Li, Zhibin Lan, Xue Xu, Wanru Zhuang, Jiachen Liu, Xinyan Xiao, Jinsong Su
Diffusion-based text-to-image models have demonstrated impressive achievements in diversity and aesthetics but struggle to generate images with legible visual texts.
1 code implementation • 24 Sep 2024 • Chuyang Zhao, Yuxing Song, Wenhao Wang, Haocheng Feng, Errui Ding, Yifan Sun, Xinyan Xiao, Jingdong Wang
Most existing multimodality methods use separate backbones for autoregression-based discrete text generation and diffusion-based continuous visual generation, or the same backbone by discretizing the visual data to use autoregression for both text and visual generation.
no code implementations • 23 Sep 2024 • Yuanhang Li, Qi Mao, Lan Chen, Zhen Fang, Lei Tian, Xinyan Xiao, Libiao Jin, Hua Wu
To enhance the motion-subject binding, we implement a syntax-guided contrastive constraint in the subsequent denoising phase, aimed at improving the correlations between the CA maps of verbs and their corresponding nouns. Both qualitative and quantitative evaluations demonstrate that the proposed framework significantly outperforms baseline approaches, producing higher-quality videos with improved subject-motion consistency.
no code implementations • International Journal of Computer Vision 2024 • Zhenyu Huang, Peng Hu, guocheng niu, Xinyan Xiao, Jiancheng Lv, Xi Peng
This paper studies a new learning paradigm for noisy labels, i. e., noisy correspondence (NC).
Cross-modal retrieval with noisy correspondence Text Retrieval +2
no code implementations • 24 Jan 2024 • Wei Li, Xue Xu, Jiachen Liu, Xinyan Xiao
This paper presents UNIMO-G, a simple multimodal conditional diffusion framework that operates on multimodal prompts with interleaved textual and visual inputs, which demonstrates a unified ability for both text-driven and subject-driven image generation.
no code implementations • 17 Jan 2024 • Ludan Ruan, Lei Tian, Chuanwei Huang, Xu Zhang, Xinyan Xiao
This cannot fully meet the needs of real-world application scenarios, as users are likely to input images and text conditions in a flexible manner, either individually or in combination.
no code implementations • 11 Jan 2024 • Hanzhang Wang, Haoran Wang, Jinze Yang, Zhongrui Yu, Zeke Xie, Lei Tian, Xinyan Xiao, Junjun Jiang, Xianming Liu, Mingming Sun
In the specific, our model is constructed based on Latent Diffusion Model (LDM) and elaborately designed to absorb content and style instance as conditions of LDM.
no code implementations • 2 Dec 2023 • Yuanbiao Gou, Haiyu Zhao, Boyun Li, Xinyan Xiao, Xi Peng
In contrast to close-set scenarios that restore images from a predefined set of degradations, open-set image restoration aims to handle the unknown degradations that were unforeseen during the pretraining phase, which is less-touched as far as we know.
no code implementations • 23 May 2023 • Hao Yang, Can Gao, Hao Líu, Xinyan Xiao, Yanyan Zhao, Bing Qin
The experimental results show that our model achieves state-of-the-art performance in various downstream tasks, and through ablation study can prove that effective cross-layer learning improves the model's ability of multimodal representation.
1 code implementation • 20 Dec 2022 • Wenhao Wu, Wei Li, Xinyan Xiao, Jiachen Liu, Sujian Li, Yajuan Lv
As a result, they perform poorly on the real generated text and are biased heavily by their single-source upstream tasks.
no code implementations • 1 Nov 2022 • Wenhao Wu, Wei Li, Jiachen Liu, Xinyan Xiao, Ziqiang Cao, Sujian Li, Hua Wu
We first measure a model's factual robustness by its success rate to defend against adversarial attacks when generating factual information.
no code implementations • 28 Oct 2022 • Wei Li, Xue Xu, Xinyan Xiao, Jiachen Liu, Hu Yang, Guohao Li, Zhanpeng Wang, Zhifan Feng, Qiaoqiao She, Yajuan Lyu, Hua Wu
Diffusion generative models have recently greatly improved the power of text-conditioned image generation.
no code implementations • 22 Oct 2022 • Wenhao Wu, Wei Li, Jiachen Liu, Xinyan Xiao, Sujian Li, Yajuan Lyu
Though model robustness has been extensively studied in language understanding, the robustness of Seq2Seq generation remains understudied.
no code implementations • 26 Aug 2022 • Saihao Huang, Lijie Wang, Zhenghua Li, Zeyang Liu, Chenhui Dou, Fukang Yan, Xinyan Xiao, Hua Wu, Min Zhang
As the first session-level Chinese dataset, CHASE contains two separate parts, i. e., 2, 003 sessions manually constructed from scratch (CHASE-C), and 3, 456 sessions translated from English SParC (CHASE-T).
no code implementations • 28 Jul 2022 • Yaozong Shen, Lijie Wang, Ying Chen, Xinyan Xiao, Jing Liu, Hua Wu
To fill in the gap, we propose a novel evaluation benchmark providing with both English and Chinese annotated data.
no code implementations • 23 May 2022 • Lijie Wang, Yaozong Shen, Shuyuan Peng, Shuai Zhang, Xinyan Xiao, Hao liu, Hongxuan Tang, Ying Chen, Hua Wu, Haifeng Wang
Based on this benchmark, we conduct experiments on three typical models with three saliency methods, and unveil their strengths and weakness in terms of interpretability.
no code implementations • 26 Apr 2022 • Kun Wu, Lijie Wang, Zhenghua Li, Xinyan Xiao
Grammar-based parsers have achieved high performance in the cross-domain text-to-SQL parsing task, but suffer from low decoding efficiency due to the much larger number of actions for grammar selection than that of tokens in SQL queries.
2 code implementations • ACL 2022 • Yaojie Lu, Qing Liu, Dai Dai, Xinyan Xiao, Hongyu Lin, Xianpei Han, Le Sun, Hua Wu
Information extraction suffers from its varying targets, heterogeneous structures, and demand-specific schemas.
Ranked #4 on Aspect-Based Sentiment Analysis (ABSA) on ASTE (using extra training data)
no code implementations • Findings (ACL) 2022 • Luyang Huang, guocheng niu, Jiachen Liu, Xinyan Xiao, Hua Wu
To bridge the gap between image understanding and generation, we further design a novel commitment loss.
no code implementations • ACL 2022 • Zhe Hu, Hou Pong Chan, Jiachen Liu, Xinyan Xiao, Hua Wu, Lifu Huang
Despite recent progress of pre-trained language models on generating fluent text, existing methods still suffer from incoherence problems in long-form text generation tasks that require proper content control and planning to form a coherent high-level logical flow.
1 code implementation • Findings (ACL) 2022 • Wei Li, Can Gao, guocheng niu, Xinyan Xiao, Hao liu, Jiachen Liu, Hua Wu, Haifeng Wang
In particular, we propose to conduct grounded learning on both images and texts via a sharing grounded space, which helps bridge unaligned images and texts, and align the visual and textual semantic spaces on different types of corpora.
no code implementations • 10 Mar 2022 • Wei Li, Wenhao Wu, Moye Chen, Jiachen Liu, Xinyan Xiao, Hua Wu
In this survey, we provide a systematic overview of the research progress on the faithfulness problem of NLG, including problem analysis, evaluation metrics and optimization methods.
1 code implementation • NeurIPS 2021 • Zhenyu Huang, guocheng niu, Xiao Liu, Wenbiao Ding, Xinyan Xiao, Hua Wu, Xi Peng
Based on this observation, we reveal and study a latent and challenging direction in cross-modal matching, named noisy correspondence, which could be regarded as a new paradigm of noisy labels.
Cross-modal retrieval with noisy correspondence Image-text matching +3
1 code implementation • 25 Oct 2021 • Moye Chen, Wei Li, Jiachen Liu, Xinyan Xiao, Hua Wu, Haifeng Wang
Comparing with traditional methods, our method has two main advantages: (1) the relations between sentences are captured by modeling both the graph structure of the whole document set and the candidate sub-graphs; (2) directly outputs an integrate summary in the form of sub-graph which is more informative and coherent.
no code implementations • 17 Sep 2021 • Hongxuan Tang, Hao liu, Xinyan Xiao, Hua Wu
Based on this, we propose a multimodal sentiment analysis dataset, named baiDu Video Sentiment dataset (DuVideoSenti), and introduce a new sentiment system which is designed to describe the sentimental style of a video on recommendation scenery.
no code implementations • 14 Sep 2021 • Zhe Hu, Zhiwei Cao, Hou Pong Chan, Jiachen Liu, Xinyan Xiao, Jinsong Su, Hua Wu
Controllable text generation is an appealing but challenging task, which allows users to specify particular attributes of the generated outputs.
no code implementations • EMNLP 2021 • Qing Liu, Hongyu Lin, Xinyan Xiao, Xianpei Han, Le Sun, Hua Wu
Conventional entity typing approaches are based on independent classification paradigms, which make them difficult to recognize inter-dependent, long-tailed and fine-grained entity types.
Ranked #8 on Entity Typing on Open Entity
no code implementations • 30 Aug 2021 • Lijie Wang, Hao liu, Shuyuan Peng, Hongxuan Tang, Xinyan Xiao, Ying Chen, Hua Wu, Haifeng Wang
Therefore, in order to systematically evaluate the factors for building trustworthy systems, we propose a novel and well-annotated sentiment analysis dataset to evaluate robustness and interpretability.
no code implementations • ACL 2021 • Wenhao Wu, Wei Li, Xinyan Xiao, Jiachen Liu, Ziqiang Cao, Sujian Li, Hua Wu, Haifeng Wang
Abstractive summarization for long-document or multi-document remains challenging for the Seq2Seq architecture, as Seq2Seq is not good at analyzing long-distance relations in text.
no code implementations • 18 May 2021 • Bofeng Wu, guocheng niu, Jun Yu, Xinyan Xiao, Jian Zhang, Hua Wu
This paper proposes an approach to Dense Video Captioning (DVC) without pairwise event-sentence annotation.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Jialong Tang, Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun, Xinyan Xiao, Hua Wu
One of the biggest bottlenecks in building accurate, high coverage neural open IE systems is the need for large labelled corpora.
1 code implementation • EMNLP 2021 • Kun Wu, Lijie Wang, Zhenghua Li, Ao Zhang, Xinyan Xiao, Hua Wu, Min Zhang, Haifeng Wang
For better distribution matching, we require that at least 80% of SQL patterns in the training data are covered by generated queries.
3 code implementations • ACL 2021 • Wei Li, Can Gao, guocheng niu, Xinyan Xiao, Hao liu, Jiachen Liu, Hua Wu, Haifeng Wang
Existed pre-training methods either focus on single-modal tasks or multi-modal tasks, and cannot effectively adapt to each other.
Ranked #4 on Image Captioning on MS COCO
2 code implementations • 2 Sep 2020 • Shuai Zhang, Lijie Wang, Ke Sun, Xinyan Xiao
DDParser is extended on the graph-based biaffine parser to accommodate to the characteristics of Chinese dataset.
2 code implementations • ACL 2020 • Wei Li, Xinyan Xiao, Jiachen Liu, Hua Wu, Haifeng Wang, Junping Du
Graphs that capture relations between textual units have great benefits for detecting salient information from multiple documents and generating overall coherent summaries.
7 code implementations • ACL 2020 • Hao Tian, Can Gao, Xinyan Xiao, Hao liu, Bolei He, Hua Wu, Haifeng Wang, Feng Wu
In particular, the prediction of aspect-sentiment pairs is converted into multi-label classification, aiming to capture the dependency between words in a pair.
Ranked #14 on Stock Market Prediction on Astock
1 code implementation • ACL 2020 • Chulun Zhou, Liang-Yu Chen, Jiachen Liu, Xinyan Xiao, Jinsong Su, Sheng Guo, Hua Wu
Unsupervised style transfer aims to change the style of an input sentence while preserving its original content without using parallel training data.
no code implementations • IJCNLP 2019 • Guocheng Niu, Hengru Xu, Bolei He, Xinyan Xiao, Hua Wu, Sheng Gao
For text classification, traditional local feature driven models learn long dependency by deeply stacking or hybrid modeling.
no code implementations • AAAI-2019 2019 • Dai Dai, Xinyan Xiao, Yajuan Lyu, Shan Dou, Qiaoqiao She, Haifeng Wang
Joint entity and relation extraction is to detect entity and relation using a single model.
Ranked #2 on Relation Extraction on NYT-single
no code implementations • ACL 2019 • Wei Jia, Dai Dai, Xinyan Xiao, Hua Wu
In this paper, we propose ARNOR, a novel Attention Regularization based NOise Reduction framework for distant supervision relation classification.
no code implementations • EMNLP 2018 • Wei Li, Xinyan Xiao, Yajuan Lyu, Yuanzhuo Wang
Information selection is the most important component in document summarization task.
Ranked #32 on Abstractive Text Summarization on CNN / Daily Mail
no code implementations • EMNLP 2018 • Wei Li, Xinyan Xiao, Yajuan Lyu, Yuanzhuo Wang
Recent neural sequence-to-sequence models have shown significant progress on short text summarization.
Ranked #43 on Abstractive Text Summarization on CNN / Daily Mail
no code implementations • ACL 2018 • Zhen Wang, Jiachen Liu, Xinyan Xiao, Yajuan Lyu, Tian Wu
While sophisticated neural-based techniques have been developed in reading comprehension, most approaches model the answer in an independent manner, ignoring its relations with other answer candidates.
3 code implementations • WS 2018 • Wei He, Kai Liu, Jing Liu, Yajuan Lyu, Shiqi Zhao, Xinyan Xiao, Yu-An Liu, Yizhong Wang, Hua Wu, Qiaoqiao She, Xuan Liu, Tian Wu, Haifeng Wang
Experiments show that human performance is well above current state-of-the-art baseline systems, leaving plenty of room for the community to make improvements.