1 code implementation • Findings (NAACL) 2022 • Huan Lin, Baosong Yang, Liang Yao, Dayiheng Liu, Haibo Zhang, Jun Xie, Min Zhang, Jinsong Su
Diverse NMT aims at generating multiple diverse yet faithful translations given a source sentence.
1 code implementation • COLING 2022 • Zhongjian Miao, Xiang Li, Liyan Kang, Wen Zhang, Chulun Zhou, Yidong Chen, Bin Wang, Min Zhang, Jinsong Su
Most existing methods on robust neural machine translation (NMT) construct adversarial examples by injecting noise into authentic examples and indiscriminately exploit two types of examples.
no code implementations • 30 Mar 2025 • Hui Li, Ante Wang, kunquan li, Zhihao Wang, Liang Zhang, Delai Qiu, Qingsong Liu, Jinsong Su
To address these issues, we propose a MultiAgent Framework for cross-domain misinformation detection with Automated Decision Rule Optimization (MARO).
no code implementations • 4 Mar 2025 • Zhibin Lan, LiQiang Niu, Fandong Meng, Jie zhou, Jinsong Su
To deal with this issue, we propose a simple yet effective framework that dynamically improves the embedding model's representation learning for negative pairs based on their discriminative difficulty.
no code implementations • 18 Feb 2025 • Liangying Shao, Yanfu Yan, Denys Poshyvanyk, Jinsong Su
Deep learning-based code generation has completely transformed the way developers write programs today.
no code implementations • 17 Feb 2025 • Yujie Lin, Ante Wang, Moye Chen, Jingyao Liu, Hao liu, Jinsong Su, Xinyan Xiao
Recently, inference-time scaling of chain-of-thought (CoT) has been demonstrated as a promising approach for addressing multi-modal reasoning tasks.
1 code implementation • 16 Feb 2025 • Ante Wang, Linfeng Song, Ye Tian, Dian Yu, Haitao Mi, Xiangyu Duan, Zhaopeng Tu, Jinsong Su, Dong Yu
Recent advancements in tree search algorithms guided by verifiers have significantly enhanced the reasoning capabilities of large language models (LLMs), but at the cost of increased computational resources.
no code implementations • 8 Jan 2025 • Yaoxiang Wang, Haoling Li, Xin Zhang, Jie Wu, Xiao Liu, Wenxiang Hu, Zhongxin Guo, Yangyu Huang, Ying Xin, Yujiu Yang, Jinsong Su, Qi Chen, Scarlett Li
Effective instruction tuning is indispensable for optimizing code LLMs, aligning model behavior with user expectations and enhancing model performance in real-world applications.
no code implementations • 26 Dec 2024 • Jiawei Yu, Xiang Geng, Yuang Li, Mengxin Ren, Wei Tang, Jiahuan Li, Zhibin Lan, Min Zhang, Hao Yang, ShuJian Huang, Jinsong Su
Spoken named entity recognition (NER) aims to identify named entities from speech, playing an important role in speech processing.
1 code implementation • 23 Dec 2024 • Yujie Lin, Jingyao Liu, Yan Gao, Ante Wang, Jinsong Su
Metaphor detection, a critical task in natural language processing, involves identifying whether a particular word in a sentence is used metaphorically.
no code implementations • 20 Nov 2024 • Jiawei Yu, Yuang Li, Xiaosong Qiao, Huan Zhao, Xiaofeng Zhao, Wei Tang, Min Zhang, Hao Yang, Jinsong Su
Existing research primarily utilizes additional text data and predefined speech styles supported by TTS models.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+3
no code implementations • 4 Nov 2024 • Zijun Min, Bingshuai Liu, Liang Zhang, Jia Song, Jinsong Su, Song He, Xiaochen Bo
In this work, we introduce the Optimal TRansport-based Multi-grained Alignments model (ORMA), a novel approach that facilitates multi-grained alignments between textual descriptions and molecules.
Ranked #2 on
Cross-Modal Retrieval
on ChEBI-20
1 code implementation • 31 Oct 2024 • Jia Song, Wanru Zhuang, Yujie Lin, Liang Zhang, Chunyan Li, Jinsong Su, Song He, Xiaochen Bo
Cross-modal text-molecule retrieval model aims to learn a shared feature space of the text and molecule modalities for accurate similarity calculation, which facilitates the rapid screening of molecules with specific properties and activities in drug design.
Ranked #3 on
Cross-Modal Retrieval
on ChEBI-20
1 code implementation • 29 Oct 2024 • Suhang Wu, Jialong Tang, Baosong Yang, Ante Wang, Kaidi Jia, Jiawei Yu, Junfeng Yao, Jinsong Su
Experimental results reveal linguistic inequalities: 1) high-resource languages stand out in Monolingual Knowledge Extraction; 2) Indo-European languages lead RALMs to provide answers directly from documents, alleviating the challenge of expressing answers across languages; 3) English benefits from RALMs' selection bias and speaks louder in multilingual knowledge selection.
no code implementations • 6 Oct 2024 • Wenbo Li, Guohao Li, Zhibin Lan, Xue Xu, Wanru Zhuang, Jiachen Liu, Xinyan Xiao, Jinsong Su
Diffusion-based text-to-image models have demonstrated impressive achievements in diversity and aesthetics but struggle to generate images with legible visual texts.
no code implementations • 5 Oct 2024 • Zhihao Wang, Shiyu Liu, Jianheng Huang, Zheng Wang, Yixuan Liao, Xiaoxin Chen, Junfeng Yao, Jinsong Su
Preliminary experiments demonstrate that PTFS achieves better pre-training performance, while CPT has lower training cost.
1 code implementation • 4 Oct 2024 • Liangying Shao, Liang Zhang, Minlong Peng, Guoqi Ma, Hao Yue, Mingming Sun, Jinsong Su
Further analysis shows that: 1) the one2set paradigm owns the advantage of high recall, but suffers from improper assignments of supervision signals during training; 2) LLMs are powerful in keyphrase selection, but existing selection methods often make redundant selections.
1 code implementation • 29 Sep 2024 • Ante Wang, Linfeng Song, Zijun Min, Ge Xu, Xiaoli Wang, Junfeng Yao, Jinsong Su
We carefully analyze the negative effects of this phenomenon on pretrained Seq2seq query producers and then propose effective instance-level weighting strategies for training to mitigate these issues from multiple perspectives.
1 code implementation • 20 Sep 2024 • Zhibin Lan, LiQiang Niu, Fandong Meng, Wenbo Li, Jie zhou, Jinsong Su
Recently, when dealing with high-resolution images, dominant LMMs usually divide them into multiple local images and one global image, which will lead to a large number of visual tokens.
1 code implementation • 5 Sep 2024 • Yihang Zheng, Bo Li, Zhenghao Lin, Yi Luo, Xuanhe Zhou, Chen Lin, Jinsong Su, Guoliang Li, Shifu Li
However, there is still a lack of a comprehensive benchmark to evaluate the capabilities of different LLMs and their modular components in database QA.
1 code implementation • 3 Jul 2024 • Zhibin Lan, LiQiang Niu, Fandong Meng, Jie zhou, Min Zhang, Jinsong Su
Among them, the target text decoder is used to alleviate the language alignment burden, and the image tokenizer converts long sequences of pixels into shorter sequences of visual tokens, preventing the model from focusing on low-level visual features.
no code implementations • 29 Jun 2024 • Ante Wang, Linfeng Song, Ye Tian, Baolin Peng, Dian Yu, Haitao Mi, Jinsong Su, Dong Yu
Recent research suggests that tree search algorithms (e. g. Monte Carlo Tree Search) can dramatically boost LLM performance on complex mathematical reasoning tasks.
1 code implementation • 24 Jun 2024 • Hao Yue, Shaopeng Lai, Chengyi Yang, Liang Zhang, Junfeng Yao, Jinsong Su
However, these studies ignore the non-bridge entities, each of which co-occurs with only one target entity and offers the semantic association between target entities for relation prediction.
1 code implementation • 10 Jun 2024 • Yan Gao, Zhiwei Cao, Zhongjian Miao, Baosong Yang, Shiyu Liu, Min Zhang, Jinsong Su
In this paper, we first conduct a preliminary study to reveal two key limitations of $k$NN-MT-AR: 1) the optimization gap leads to inaccurate estimation of $\lambda$ for determining $k$NN retrieval skipping, and 2) using a fixed threshold fails to accommodate the dynamic demands for $k$NN retrieval at different timesteps.
1 code implementation • 4 Jun 2024 • Zhiwei Cao, Qian Cao, Yu Lu, Ningxin Peng, Luyang Huang, Shanbo Cheng, Jinsong Su
This decline can be attributed to the loss of key information during the compression process.
no code implementations • 21 May 2024 • Huangjun Shen, Liangying Shao, Wenbo Li, Zhibin Lan, Zhanyu Liu, Jinsong Su
In recent years, multi-modal machine translation has attracted significant interest in both academia and industry due to its superior performance.
no code implementations • 4 May 2024 • Zhihao Wang, Longyue Wang, Jinsong Su, Junfeng Yao, Zhaopeng Tu
By manually annotating the NAT outputs, we identify two types of information redundancy errors that correspond well to lexical and reordering multi-modality problems.
no code implementations • 14 Mar 2024 • Ante Wang, Linfeng Song, Ye Tian, Baolin Peng, Lifeng Jin, Haitao Mi, Jinsong Su, Dong Yu
Calibration, which establishes the correlation between accuracy and model confidence, is important for LLM development.
1 code implementation • 2 Mar 2024 • Jianheng Huang, Leyang Cui, Ante Wang, Chengyi Yang, Xinting Liao, Linfeng Song, Junfeng Yao, Jinsong Su
When conducting continual learning based on a publicly-released LLM checkpoint, the availability of the original training data may be non-existent.
no code implementations • 23 Feb 2024 • Ante Wang, Linfeng Song, Baolin Peng, Ye Tian, Lifeng Jin, Haitao Mi, Jinsong Su, Dong Yu
Experiments on Biographies show that our method can effectively improve the factuality of generations with simple and intuitive prompts across different scales of LLMs.
1 code implementation • 15 Feb 2024 • Yaoxiang Wang, Zhiyong Wu, Junfeng Yao, Jinsong Su
The emergence of Large Language Models (LLMs) like ChatGPT has inspired the development of LLM-based agents capable of addressing complex, real-world tasks.
no code implementations • 21 Jan 2024 • Yuang Li, Jiawei Yu, Min Zhang, Mengxin Ren, Yanqing Zhao, Xiaofeng Zhao, Shimin Tao, Jinsong Su, Hao Yang
In this work, we connect the Whisper encoder with ChatGLM3 and provide in-depth comparisons of these two approaches using Chinese automatic speech recognition (ASR) and name entity recognition (NER) tasks.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+6
1 code implementation • 25 Dec 2023 • Rui Zhao, Liang Zhang, Biao Fu, Cong Hu, Jinsong Su, Yidong Chen
The first KL divergence optimizes the conditional variational autoencoder and regularizes the encoder outputs, while the second KL divergence performs a self-distillation from the posterior path to the prior path, ensuring the consistency of decoder outputs.
1 code implementation • 20 Dec 2023 • Jianheng Huang, Ante Wang, Linfeng Gao, Linfeng Song, Jinsong Su
Based on the observation that the search query is typically related to the topic of dialogue response, we train a response-augmented query producer (RA) to provide rich and effective training signals for QP.
no code implementations • 4 Dec 2023 • Bingshuai Liu, Chenyang Lyu, Zijun Min, Zhanyu Wang, Jinsong Su, Longyue Wang
The advancement of Large Language Models (LLMs) has brought substantial attention to the Chain of Thought (CoT) approach, primarily due to its ability to enhance the capability of LLMs on complex reasoning tasks.
no code implementations • 22 Nov 2023 • Yang Li, Qi'ao Zhao, Chen Lin, Zhenjie Zhang, Xiaomin Zhu, Jinsong Su
(2) The diverse semantics of side information that describes items and users from multi-level in a context different from recommendation systems.
no code implementations • 1 Nov 2023 • Xiaoyue Wang, Xin Liu, Lijie Wang, Yaoxiang Wang, Jinsong Su, Hua Wu
Then, we pair each sample with a bias indicator representing its bias degree, and use these extended samples to train a sample generator.
1 code implementation • 9 Sep 2023 • Yifan Dong, Suhang Wu, Fandong Meng, Jie zhou, Xiaoli Wang, Jianxin Lin, Jinsong Su
2) the input text and image are often not perfectly matched, and thus the image may introduce noise into the model.
no code implementations • 19 Aug 2023 • Suhang Wu, Minlong Peng, Yue Chen, Jinsong Su, Mingming Sun
In this paper, we propose Eva-KELLM, a new benchmark for evaluating knowledge editing of LLMs.
no code implementations • 6 Jul 2023 • Bingshuai Liu, Longyue Wang, Chenyang Lyu, Yong Zhang, Jinsong Su, Shuming Shi, Zhaopeng Tu
Accordingly, we propose a novel multi-modal metric that considers object-text alignment to filter the fine-tuning data in the target culture, which is used to fine-tune a T2I model to improve cross-cultural generation.
1 code implementation • 2 Jun 2023 • Xiaoyue Wang, Lijie Wang, Xin Liu, Suhang Wu, Jinsong Su, Hua Wu
In this way, the top-layer sentence representation will be trained to ignore the common biased features encoded by the low-layer sentence representation and focus on task-relevant unbiased features.
1 code implementation • 31 May 2023 • Tong Li, Zhihao Wang, Liangying Shao, Xuling Zheng, Xiaoli Wang, Jinsong Su
Specifically, in addition to a text encoder encoding the input text, our model is equipped with a table header generator to first output a table header, i. e., the first row of the table, in the manner of sequence generation.
1 code implementation • 27 May 2023 • Zhibin Lan, Jiawei Yu, Xiang Li, Wen Zhang, Jian Luan, Bin Wang, Degen Huang, Jinsong Su
Text image translation (TIT) aims to translate the source texts embedded in the image to target translations, which has a wide range of applications and thus has important research value.
1 code implementation • 26 May 2023 • Zhiwei Cao, Baosong Yang, Huan Lin, Suhang Wu, Xiangpeng Wei, Dayiheng Liu, Jun Xie, Min Zhang, Jinsong Su
$k$-Nearest neighbor machine translation ($k$NN-MT) has attracted increasing attention due to its ability to non-parametrically adapt to new translation domains.
1 code implementation • 25 May 2023 • Zhihao Wang, Longyue Wang, Jinsong Su, Junfeng Yao, Zhaopeng Tu
Experimental results on the large-scale WMT20 En-De show that the asymmetric architecture (e. g. bigger encoder and smaller decoder) can achieve comparable performance with the scaling model, while maintaining the superiority of decoding speed with standard NAT models.
1 code implementation • 23 May 2023 • Liyan Kang, Luyang Huang, Ningxin Peng, Peihao Zhu, Zewei Sun, Shanbo Cheng, Mingxuan Wang, Degen Huang, Jinsong Su
We also introduce two deliberately designed test sets to verify the necessity of visual information: Ambiguous with the presence of ambiguous words, and Unambiguous in which the text context is self-contained for translation.
1 code implementation • 13 May 2023 • Chulun Zhou, Yunlong Liang, Fandong Meng, Jinan Xu, Jinsong Su, Jie zhou
In this paper, we propose Regularized Contrastive Cross-lingual Cross-modal (RC^3) pre-training, which further exploits more abundant weakly-aligned multilingual image-text pairs.
no code implementations • 4 May 2023 • Binbin Xie, Jia Song, Liangying Shao, Suhang Wu, Xiangpeng Wei, Baosong Yang, Huan Lin, Jun Xie, Jinsong Su
In this paper, we comprehensively summarize representative studies from the perspectives of dominant models, datasets and evaluation metrics.
1 code implementation • 16 Feb 2023 • Ante Wang, Linfeng Song, Qi Liu, Haitao Mi, Longyue Wang, Zhaopeng Tu, Jinsong Su, Dong Yu
We propose a dialogue model that can access the vast and dynamic information from any search engine for response generation.
no code implementations • 27 Jan 2023 • Chulun Zhou, Yunlong Liang, Fandong Meng, Jie zhou, Jinan Xu, Hongji Wang, Min Zhang, Jinsong Su
To address these issues, in this paper, we propose a multi-task multi-stage transitional (MMT) training framework, where an NCT model is trained using the bilingual chat translation dataset and additional monolingual dialogues.
1 code implementation • 26 Nov 2022 • Liang Zhang, Jinsong Su, Yidong Chen, Zhongjian Miao, Zijun Min, Qingguo Hu, Xiaodong Shi
Existing methods usually directly predict the relations of all entity pairs of input document in a one-pass manner, ignoring the fact that predictions of some entity pairs heavily depend on the predicted results of other pairs.
1 code implementation • 13 Nov 2022 • Binbin Xie, Xiangpeng Wei, Baosong Yang, Huan Lin, Jun Xie, Xiaoli Wang, Min Zhang, Jinsong Su
Keyphrase generation aims to automatically generate short phrases summarizing an input document.
no code implementations • 11 Nov 2022 • Xiaoyue Wang, Linfeng Song, Xin Liu, Chulun Zhou, Jinsong Su
Simile recognition involves two subtasks: simile sentence classification that discriminates whether a sentence contains simile, and simile component extraction that locates the corresponding objects (i. e., tenors and vehicles).
1 code implementation • 18 Oct 2022 • Shuai Fan, Chen Lin, Haonan Li, Zhenghao Lin, Jinsong Su, Hang Zhang, Yeyun Gong, Jian Guo, Nan Duan
Most existing pre-trained language representation models (PLMs) are sub-optimal in sentiment analysis tasks, as they capture the sentiment information from word-level while under-considering sentence-level information.
3 code implementations • 17 Oct 2022 • Hui Jiang, Ziyao Lu, Fandong Meng, Chulun Zhou, Jie zhou, Degen Huang, Jinsong Su
Meanwhile we inject two types of perturbations into the retrieved pairs for robust training.
1 code implementation • Findings (ACL) 2022 • Shaopeng Lai, Qingyu Zhou, Jiali Zeng, Zhongli Li, Chao Li, Yunbo Cao, Jinsong Su
First, they simply mix additionally-constructed training instances and original ones to train models, which fails to help models be explicitly aware of the procedure of gradual corrections.
1 code implementation • ACL 2022 • Yunlong Liang, Fandong Meng, Chulun Zhou, Jinan Xu, Yufeng Chen, Jinsong Su, Jie zhou
The goal of the cross-lingual summarization (CLS) is to convert a document in one language (e. g., English) to a summary in another one (e. g., Chinese).
no code implementations • ACL 2022 • Chulun Zhou, Fandong Meng, Jie zhou, Min Zhang, Hongji Wang, Jinsong Su
Most dominant neural machine translation (NMT) models are restricted to make predictions only according to the local context of preceding words in a left-to-right manner.
1 code implementation • 22 Dec 2021 • Changxing Wu, Liuwen Cao, Yubin Ge, Yang Liu, Min Zhang, Jinsong Su
Then, we employ a label sequence decoder to output the predicted labels in a top-down manner, where the predicted higher-level labels are directly used to guide the label prediction at the current level.
1 code implementation • 15 Dec 2021 • Xin Liu, Dayiheng Liu, Baosong Yang, Haibo Zhang, Junwei Ding, Wenqing Yao, Weihua Luo, Haiying Zhang, Jinsong Su
Generative commonsense reasoning requires machines to generate sentences describing an everyday scenario given several concepts, which has attracted much attention recently.
1 code implementation • EMNLP 2021 • Shaopeng Lai, Ante Wang, Fandong Meng, Jie zhou, Yubin Ge, Jiali Zeng, Junfeng Yao, Degen Huang, Jinsong Su
Dominant sentence ordering models can be classified into pairwise ordering models and set-to-sequence models.
no code implementations • 14 Sep 2021 • Zhe Hu, Zhiwei Cao, Hou Pong Chan, Jiachen Liu, Xinyan Xiao, Jinsong Su, Hua Wu
Controllable text generation is an appealing but challenging task, which allows users to specify particular attributes of the generated outputs.
1 code implementation • EMNLP 2021 • Yunlong Liang, Chulun Zhou, Fandong Meng, Jinan Xu, Yufeng Chen, Jinsong Su, Jie zhou
Neural Chat Translation (NCT) aims to translate conversational text between speakers of different languages.
1 code implementation • Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence 2021 • An-Hui Wang, Linfeng Song, Hui Jiang, Shaopeng Lai, Junfeng Yao, Min Zhang, Jinsong Su
Conversational discourse structures aim to describe how a dialogue is organised, thus they are helpful for dialogue understanding and response generation.
Ranked #3 on
Discourse Parsing
on STAC
no code implementations • ACL 2021 • Yubin Ge, Ly Dinh, Xiaofeng Liu, Jinsong Su, Ziyao Lu, Ante Wang, Jana Diesner
In this paper, we focus on the problem of citing sentence generation, which entails generating a short text to capture the salient information in a cited paper and the connection between the citing and cited paper.
no code implementations • ACL 2021 • Xin Liu, Baosong Yang, Dayiheng Liu, Haibo Zhang, Weihua Luo, Min Zhang, Haiying Zhang, Jinsong Su
A well-known limitation in pretrain-finetune paradigm lies in its inflexibility caused by the one-size-fits-all vocabulary.
1 code implementation • ACL 2021 • Huan Lin, Liang Yao, Baosong Yang, Dayiheng Liu, Haibo Zhang, Weihua Luo, Degen Huang, Jinsong Su
Furthermore, we contribute the first Chinese-English parallel corpus annotated with user behavior called UDT-Corpus.
1 code implementation • ACL 2021 • Hui Jiang, Chulun Zhou, Fandong Meng, Biao Zhang, Jie zhou, Degen Huang, Qingqiang Wu, Jinsong Su
Due to the great potential in facilitating software development, code generation has attracted increasing attention recently.
no code implementations • 31 May 2021 • Binbin Xie, Jinsong Su, Yubin Ge, Xiang Li, Jianwei Cui, Junfeng Yao, Bin Wang
However, such a decoder only exploits the preorder traversal based preceding actions, which are insufficient to ensure correct action predictions.
1 code implementation • 5 Mar 2021 • Jinsong Su, Jialong Tang, Hui Jiang, Ziyao Lu, Yubin Ge, Linfeng Song, Deyi Xiong, Le Sun, Jiebo Luo
In aspect-based sentiment analysis (ABSA), many neural models are equipped with an attention mechanism to quantify the contribution of each context word to sentiment prediction.
Aspect-Based Sentiment Analysis
Aspect-Based Sentiment Analysis (ABSA)
+1
1 code implementation • ACL 2020 • Linfeng Song, Ante Wang, Jinsong Su, Yue Zhang, Kun Xu, Yubin Ge, Dong Yu
The task of graph-to-text generation aims at producing sentences that preserve the meaning of input graphs.
Ranked #14 on
Data-to-Text Generation
on WebNLG
1 code implementation • 4 Sep 2020 • Huan Lin, Fandong Meng, Jinsong Su, Yongjing Yin, Zhengyuan Yang, Yubin Ge, Jie zhou, Jiebo Luo
Particularly, we represent the input image with global and regional visual features, we introduce two parallel DCCNs to model multimodal context vectors with visual features at different granularities.
Ranked #3 on
Multimodal Machine Translation
on Multi30K
1 code implementation • ACL 2020 • Yongjing Yin, Fandong Meng, Jinsong Su, Chulun Zhou, Zhengyuan Yang, Jie zhou, Jiebo Luo
Multi-modal neural machine translation (NMT) aims to translate source sentences into a target language paired with images.
no code implementations • WS 2020 • Junxuan Chen, Xiang Li, Jiarui Zhang, Chulun Zhou, Jianwei Cui, Bin Wang, Jinsong Su
Finally, we combine the discourse structure information with the word embedding before it is fed into the encoder.
1 code implementation • ACL 2020 • Chulun Zhou, Liang-Yu Chen, Jiachen Liu, Xinyan Xiao, Jinsong Su, Sheng Guo, Hua Wu
Unsupervised style transfer aims to change the style of an input sentence while preserving its original content without using parallel training data.
1 code implementation • 19 Dec 2019 • Jiali Zeng, Linfeng Song, Jinsong Su, Jun Xie, Wei Song, Jiebo Luo
Simile recognition is to detect simile sentences and to extract simile components, i. e., tenors and vehicles.
no code implementations • IJCNLP 2019 • Jiali Zeng, Yang Liu, Jinsong Su, Yubin Ge, Yaojie Lu, Yongjing Yin, Jiebo Luo
Previous studies on the domain adaptation for neural machine translation (NMT) mainly focus on the one-pass transferring out-of-domain translation knowledge to in-domain NMT model.
1 code implementation • 16 Dec 2019 • Yongjing Yin, Linfeng Song, Jinsong Su, Jiali Zeng, Chulun Zhou, Jiebo Luo
Sentence ordering is to restore the original paragraph from a set of sentences.
no code implementations • 13 Dec 2019 • Zhengyuan Yang, Tushar Kumar, Tianlang Chen, Jinsong Su, Jiebo Luo
In this paper, we study Tracking by Language that localizes the target box sequence in a video based on a language query.
1 code implementation • 7 Dec 2019 • Yiyi Zhou, Rongrong Ji, Gen Luo, Xiaoshuai Sun, Jinsong Su, Xinghao Ding, Chia-Wen Lin, Qi Tian
Referring Expression Comprehension (REC) is an emerging research spot in computer vision, which refers to detecting the target region in an image given an text description.
1 code implementation • IJCNLP 2019 • Linfeng Song, Yue Zhang, Daniel Gildea, Mo Yu, Zhiguo Wang, Jinsong Su
Medical relation extraction discovers relations between entity mentions in text, such as research articles.
1 code implementation • 11 Aug 2019 • Songyang Zhang, Jinsong Su, Jiebo Luo
We address the problem of video moment localization with natural language, i. e. localizing a video segment described by a natural language sentence.
1 code implementation • 20 Jun 2019 • Mengge Xue, Weiming Cai, Jinsong Su, Linfeng Song, Yubin Ge, Yubao Liu, Bin Wang
However, most neural collective EL methods depend entirely upon neural networks to automatically model the semantic dependencies between different EL decisions, which lack of the guidance from external knowledge.
1 code implementation • ACL 2019 • Jialong Tang, Ziyao Lu, Jinsong Su, Yubin Ge, Linfeng Song, Le Sun, Jiebo Luo
In aspect-level sentiment classification (ASC), it is prevalent to equip dominant neural models with attention mechanisms, for the sake of acquiring the importance of each context word on the given aspect.
Aspect-Based Sentiment Analysis (ABSA)
Sentiment Classification
1 code implementation • TACL 2019 • Linfeng Song, Daniel Gildea, Yue Zhang, Zhiguo Wang, Jinsong Su
It is intuitive that semantic representations can be useful for machine translation, mainly because they can help in enforcing meaning preservation and handling data sparsity (many sentences correspond to one meaning) of machine translation models.
no code implementations • IJCNLP 2019 • Mingxuan Wang, Jun Xie, Zhixing Tan, Jinsong Su, Deyi Xiong, Lei LI
In this study, we first investigate a novel capsule network with dynamic routing for linear time Neural Machine Translation (NMT), referred as \textsc{CapsNMT}.
3 code implementations • EMNLP 2018 • Biao Zhang, Deyi Xiong, Jinsong Su, Qian Lin, Huiji Zhang
Experiments on WMT14 translation tasks demonstrate that ATR-based neural machine translation can yield competitive performance on English- German and English-French language pairs in terms of both translation quality and speed.
1 code implementation • EMNLP 2018 • Jiali Zeng, Jinsong Su, Huating Wen, Yang Liu, Jun Xie, Yongjing Yin, Jianqiang Zhao
Based on this intuition, in this paper, we devote to distinguishing and exploiting word-level domain contexts for multi-domain NMT.
no code implementations • COLING 2018 • Mingxuan Wang, Jun Xie, Zhixing Tan, Jinsong Su, Deyi Xiong, Chao Bian
Neural machine translation with source-side attention have achieved remarkable performance.
1 code implementation • 24 Jul 2018 • Jing Yang, Biao Zhang, Yue Qin, Xiangwen Zhang, Qian Lin, Jinsong Su
Although neural machine translation(NMT) yields promising translation performance, it unfortunately suffers from over- and under-translation is- sues [Tu et al., 2016], of which studies have become research hotspots in NMT.
1 code implementation • COLING 2018 • Junyang Lin, Xu sun, Xuancheng Ren, Shuming Ma, Jinsong Su, Qi Su
A great proportion of sequence-to-sequence (Seq2Seq) models for Neural Machine Translation (NMT) adopt Recurrent Neural Network (RNN) to generate translation word by word following a sequential order.
Ranked #9 on
Machine Translation
on IWSLT2015 English-Vietnamese
no code implementations • CVPR 2018 • Fuhai Chen, Rongrong Ji, Xiaoshuai Sun, Yongjian Wu, Jinsong Su
In offline optimization, we adopt an end-to-end formulation, which jointly trains the visual tree parser, the structured relevance and diversity constraints, as well as the LSTM based captioning model.
1 code implementation • ACL 2018 • Biao Zhang, Deyi Xiong, Jinsong Su
To alleviate this issue, we propose an average attention network as an alternative to the self-attention network in the decoder of the neural Transformer.
Ranked #60 on
Machine Translation
on WMT2014 English-German
no code implementations • E2E NLG Challenge System Descriptions 2018 • Biao Zhang, Jing Yang, Qian Lin, Jinsong Su
This paper describes our system used for the end-to-end (E2E) natural language generation (NLG) challenge.
Ranked #8 on
Data-to-Text Generation
on E2E NLG Challenge
no code implementations • 16 Jan 2018 • Jinsong Su, Shan Wu, Deyi Xiong, Yaojie Lu, Xianpei Han, Biao Zhang
Partially inspired by successful applications of variational recurrent neural networks, we propose a novel variational recurrent neural machine translation (VRNMT) model in this paper.
2 code implementations • 16 Jan 2018 • Xiangwen Zhang, Jinsong Su, Yue Qin, Yang Liu, Rongrong Ji, Hongji Wang
The dominant neural machine translation (NMT) models apply unified attentional encoder-decoder neural networks for translation.
no code implementations • ACL 2017 • Changxing Wu, Xiaodong Shi, Yidong Chen, Jinsong Su, Boli Wang
We introduce a simple and effective method to learn discourse-specific word embeddings (DSWE) for implicit discourse relation recognition.
no code implementations • 27 Apr 2017 • Biao Zhang, Deyi Xiong, Jinsong Su
In this paper, we propose a novel GRU-gated attention model (GAtt) for NMT which enhances the degree of discrimination of context vectors by enabling source representations to be sensitive to the partial translation generated by the decoder.
no code implementations • COLING 2016 • Jinsong Su, Biao Zhang, Deyi Xiong, Ruochen Li, Jianmin Yin
After that, we fully incorporate information of different linguistic units into a bilinear semantic similarity model.
no code implementations • COLING 2016 • Biao Zhang, Deyi Xiong, Jinsong Su, Hong Duan, Min Zhang
Parallel sentence representations are important for bilingual and cross-lingual tasks in natural language processing.
no code implementations • 25 Sep 2016 • Jinsong Su, Zhixing Tan, Deyi Xiong, Rongrong Ji, Xiaodong Shi, Yang Liu
Neural machine translation (NMT) heavily relies on word-level modelling to learn semantic representations of input sentences.
no code implementations • 29 Jul 2016 • Biao Zhang, Deyi Xiong, Jinsong Su
The vanilla sequence-to-sequence learning (seq2seq) reads and encodes a source sequence into a fixed-length vector only once, suffering from its insufficiency in modeling structural correspondence between the source and target sequence.
1 code implementation • 25 May 2016 • Biao Zhang, Deyi Xiong, Jinsong Su
In this paper, we propose a bidimensional attention based recursive autoencoder (BattRAE) to integrate clues and sourcetarget interactions at multiple levels of granularity into bilingual phrase representations.
1 code implementation • EMNLP 2016 • Biao Zhang, Deyi Xiong, Jinsong Su, Hong Duan, Min Zhang
Models of neural machine translation are often from a discriminative family of encoderdecoders that learn a conditional distribution of a target sentence given a source sentence.
1 code implementation • EMNLP 2016 • Biao Zhang, Deyi Xiong, Jinsong Su, Qun Liu, Rongrong Ji, Hong Duan, Min Zhang
In order to perform efficient inference and learning, we introduce neural discourse relation models to approximate the prior and posterior distributions of the latent variable, and employ these approximated distributions to optimize a reparameterized variational lower bound.
no code implementations • 12 Mar 2016 • Biao Zhang, Deyi Xiong, Jinsong Su
Inspired by this, we propose a neural recognizer for implicit discourse relation analysis, which builds upon a semantic memory that stores knowledge in a distributed fashion.