no code implementations • 24 Jun 2024 • Deng Cai, Huayang Li, Tingchen Fu, Siheng Li, Weiwen Xu, Shuaiyi Li, Bowen Cao, Zhisong Zhang, Xinting Huang, Leyang Cui, Yan Wang, Lemao Liu, Taro Watanabe, Shuming Shi
Despite the general capabilities of pre-trained large language models (LLMs), they still need further adaptation to better serve practical applications.
1 code implementation • 12 Jun 2024 • Benjamin Hsu, Xiaoyu Liu, Huayang Li, Yoshinari Fujinuma, Maria Nadejde, Xing Niu, Yair Kittenplon, Ron Litman, Raghavendra Pappagari
Document translation poses a challenge for Neural Machine Translation (NMT) systems.
1 code implementation • 25 Mar 2024 • Huayang Li, Deng Cai, Zhi Qu, Qu Cui, Hidetaka Kamigaito, Lemao Liu, Taro Watanabe
In our work, we propose a new task formulation of dense retrieval, cross-lingual contextualized phrase retrieval, which aims to augment cross-lingual applications by addressing polysemy using context information.
1 code implementation • 16 Jan 2024 • Shuming Shi, Enbo Zhao, Deng Cai, Leyang Cui, Xinting Huang, Huayang Li
We present Inferflow, an efficient and highly configurable inference engine for large language models (LLMs).
no code implementations • 25 Nov 2023 • Zhanyu Wang, Longyue Wang, Zhen Zhao, Minghao Wu, Chenyang Lyu, Huayang Li, Deng Cai, Luping Zhou, Shuming Shi, Zhaopeng Tu
While the recent advances in Multimodal Large Language Models (MLLMs) constitute a significant leap forward in the field, these models are predominantly confined to the realm of input-side multimodal comprehension, lacking the capacity for multimodal content generation.
1 code implementation • 14 Sep 2023 • Huayang Li, Siheng Li, Deng Cai, Longyue Wang, Lemao Liu, Taro Watanabe, Yujiu Yang, Shuming Shi
We release our dataset, model, and demo to foster future research in the area of multimodal instruction following.
Ranked #158 on Visual Question Answering on MM-Vet
1 code implementation • 25 May 2023 • Yixuan Su, Tian Lan, Huayang Li, Jialu Xu, Yan Wang, Deng Cai
To do so, PandaGPT combines the multimodal encoders from ImageBind and the large language models from Vicuna.
1 code implementation • 22 May 2023 • Haoran Yang, Deng Cai, Huayang Li, Wei Bi, Wai Lam, Shuming Shi
We introduce a frustratingly simple, super efficient and surprisingly effective decoding method, which we call Frustratingly Simple Decoding (FSD), for neural text generation.
no code implementations • 27 Mar 2023 • Xuanfan Ni, Piji Li, Huayang Li
Text structuralization is one of the important fields of natural language processing (NLP) consists of information extraction (IE) and structure formalization.
1 code implementation • 26 Oct 2022 • Huayang Li, Deng Cai, Jin Xu, Taro Watanabe
The combination of $n$-gram and neural LMs not only allows the neural part to focus on the deeper understanding of language but also provides a flexible way to customize an LM by switching the underlying $n$-gram model without changing the neural model.
2 code implementations • 6 Jun 2022 • Jin Xu, Xiaojiang Liu, Jianhao Yan, Deng Cai, Huayang Li, Jian Li
While large-scale neural language models, such as GPT2 and BART, have achieved impressive results on various text generation tasks, they tend to get stuck in undesirable sentence-level loops with maximization-based decoding algorithms (\textit{e. g.}, greedy search).
no code implementations • Findings (ACL) 2022 • Jiannan Xiang, Huayang Li, Defu Lian, Guoping Huang, Taro Watanabe, Lemao Liu
To this end, we study the dynamic relationship between the encoded linguistic information and task performance from the viewpoint of Pareto Optimality.
no code implementations • Findings (ACL) 2022 • Jiannan Xiang, Huayang Li, Yahui Liu, Lemao Liu, Guoping Huang, Defu Lian, Shuming Shi
Current practices in metric evaluation focus on one single dataset, e. g., Newstest dataset in each year's WMT Metrics Shared Task.
no code implementations • 2 Feb 2022 • Huayang Li, Yixuan Su, Deng Cai, Yan Wang, Lemao Liu
Recently, retrieval-augmented text generation attracted increasing attention of the computational linguistics community.
no code implementations • ACL 2021 • Huayang Li, Lemao Liu, Guoping Huang, Shuming Shi
In this paper, we propose the task of general word-level autocompletion (GWLAN) from a real-world CAT scenario, and construct the first public benchmark to facilitate research in this topic.
no code implementations • ACL 2021 • Wei Bi, Huayang Li, Jiacheng Huang
Data augmentation is an effective way to improve the performance of many neural text generation models.
no code implementations • 27 May 2021 • Guoping Huang, Lemao Liu, Xing Wang, Longyue Wang, Huayang Li, Zhaopeng Tu, Chengyan Huang, Shuming Shi
Automatic machine translation is super efficient to produce translations yet their quality is not guaranteed.
1 code implementation • ACL 2021 • Deng Cai, Yan Wang, Huayang Li, Wai Lam, Lemao Liu
Second, the memory retriever and NMT model can be jointly optimized for the ultimate translation goal.
1 code implementation • Findings (ACL) 2021 • Jiannan Xiang, Yahui Liu, Deng Cai, Huayang Li, Defu Lian, Lemao Liu
An important aspect of developing dialogue systems is how to evaluate and compare the performance of different systems.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Huayang Li, Lemao Liu, Guoping Huang, Shuming Shi
Many efforts have been devoted to extracting constituency trees from pre-trained language models, often proceeding in two stages: feature definition and parsing.
1 code implementation • 10 Aug 2020 • Yahui Liu, Marco De Nadai, Deng Cai, Huayang Li, Xavier Alameda-Pineda, Nicu Sebe, Bruno Lepri
Our proposed model disentangles the image content from the visual attributes, and it learns to modify the latter using the textual description, before generating a new image from the content and the modified attribute representation.
no code implementations • ACL 2020 • Jierui Li, Lemao Liu, Huayang Li, Guanlin Li, Guoping Huang, Shuming Shi
Recently many efforts have been devoted to interpreting the black-box NMT models, but little progress has been made on metrics to evaluate explanation methods.
no code implementations • 13 Aug 2019 • Huayang Li, Guoping Huang, Deng Cai, Lemao Liu
Experiments show that our approach can indeed improve the translation quality with the automatically generated constraints.