5 code implementations • ACL 2021 • Yang Xu, Yiheng Xu, Tengchao Lv, Lei Cui, Furu Wei, Guoxin Wang, Yijuan Lu, Dinei Florencio, Cha Zhang, Wanxiang Che, Min Zhang, Lidong Zhou
Pre-training of text and layout has proved effective in a variety of visually-rich document understanding tasks due to its effective model architecture and the advantage of large-scale unlabeled scanned/digital-born documents.
Ranked #1 on Key Information Extraction on SROIE
2 code implementations • 19 Oct 2022 • Hongqiu Wu, Ruixue Ding, Hai Zhao, Boli Chen, Pengjun Xie, Fei Huang, Min Zhang
Multiple pre-training objectives fill the vacancy of the understanding capability of single-objective language modeling, which serves the ultimate purpose of pre-trained language models (PrLMs), generalizing well on a mass of scenarios.
2 code implementations • 27 Oct 2022 • Peijie Jiang, Dingkun Long, Yanzhao Zhang, Pengjun Xie, Meishan Zhang, Min Zhang
We apply BABERT for feature induction of Chinese sequence labeling tasks.
Ranked #1 on Chinese Word Segmentation on MSRA
Chinese Named Entity Recognition Chinese Word Segmentation +3
2 code implementations • WWW 2020 • Chong Chen, Min Zhang, Weizhi Ma, Yiqun Liu, and Shaoping Ma Department of Computer Science and Technology, Institute for Articial Intelligence, Beijing National Research Center for Information Science and Technology, Tsinghua University cc17@mails.tsinghua.edu.cn, z-m@tsinghua.edu.cn
Factorization Machines (FM) with negative sampling is a popular solution for context-aware recommendation.
1 code implementation • EMNLP 2021 • Kun Wu, Lijie Wang, Zhenghua Li, Ao Zhang, Xinyan Xiao, Hua Wu, Min Zhang, Haifeng Wang
For better distribution matching, we require that at least 80% of SQL patterns in the training data are covered by generated queries.
9 code implementations • NeurIPS 2021 • Xiaobo Liang, Lijun Wu, Juntao Li, Yue Wang, Qi Meng, Tao Qin, Wei Chen, Min Zhang, Tie-Yan Liu
Dropout is a powerful and widely used technique to regularize the training of deep neural networks.
Ranked #4 on Machine Translation on WMT2014 English-French
2 code implementations • ACL 2020 • Yu Zhang, Zhenghua Li, Min Zhang
Experiments and analysis on 27 datasets from 13 languages clearly show that techniques developed before the DL era, such as structural learning (global TreeCRF loss) and high-order modeling are still useful, and can further boost parsing performance over the state-of-the-art biaffine parser, especially for partially annotated training data.
Ranked #1 on Dependency Parsing on CoNLL-2009
3 code implementations • EMNLP 2018 • Jiacheng Zhang, Huanbo Luan, Maosong Sun, FeiFei Zhai, Jingfang Xu, Min Zhang, Yang Liu
Although the Transformer translation model (Vaswani et al., 2017) has achieved state-of-the-art performance in a variety of translation tasks, how to use document-level context to deal with discourse phenomena problematic for Transformer still remains a challenge.
2 code implementations • NAACL 2022 • Yue Zhang, Zhenghua Li, Zuyi Bao, Jiacheng Li, Bo Zhang, Chen Li, Fei Huang, Min Zhang
This paper presents MuCGEC, a multi-reference multi-source evaluation dataset for Chinese Grammatical Error Correction (CGEC), consisting of 7, 063 sentences collected from three Chinese-as-a-Second-Language (CSL) learner sources.
2 code implementations • 8 Apr 2024 • Longhui Zhang, Dingkun Long, Meishan Zhang, Yanzhao Zhang, Pengjun Xie, Min Zhang
Experimental results on Chinese sequence labeling datasets demonstrate that the improved BABERT variant outperforms the vanilla version, not only on these tasks but also more broadly across a range of Chinese natural language understanding tasks.
1 code implementation • 25 Jan 2024 • Haoze Wu, Omri Isac, Aleksandar Zeljić, Teruhiro Tagomori, Matthew Daggitt, Wen Kokke, Idan Refaeli, Guy Amir, Kyle Julian, Shahaf Bassan, Pei Huang, Ori Lahav, Min Wu, Min Zhang, Ekaterina Komendantskaya, Guy Katz, Clark Barrett
This paper serves as a comprehensive system description of version 2. 0 of the Marabou framework for formal analysis of neural networks.
1 code implementation • 11 Dec 2021 • Tong Zhu, Xiaoye Qu, Wenliang Chen, Zhefeng Wang, Baoxing Huai, Nicholas Jing Yuan, Min Zhang
Most previous studies of document-level event extraction mainly focus on building argument chains in an autoregressive way, which achieves a certain success but is inefficient in both training and inference.
Ranked #3 on Document-level Event Extraction on ChFinAnn
2 code implementations • 11 Jun 2021 • Bin Hao, Min Zhang, Weizhi Ma, Shaoyun Shi, Xinxing Yu, Houzhi Shan, Yiqun Liu, Shaoping Ma
To the best of our knowledge, this is the largest real-world interaction dataset for personalized recommendation.
1 code implementation • 21 Mar 2024 • Han Zhao, Min Zhang, Wei Zhao, Pengxiang Ding, Siteng Huang, Donglin Wang
In recent years, the application of multimodal large language models (MLLM) in various fields has achieved remarkable success.
1 code implementation • 20 Apr 2022 • Yisheng Xiao, Lijun Wu, Junliang Guo, Juntao Li, Min Zhang, Tao Qin, Tie-Yan Liu
While NAR generation can significantly accelerate inference speed for machine translation, the speedup comes at the cost of sacrificed translation accuracy compared to its counterpart, autoregressive (AR) generation.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +11
1 code implementation • COLING 2018 • Yaosheng Yang, Wenliang Chen, Zhenghua Li, Zhengqiu He, Min Zhang
A bottleneck problem with Chinese named entity recognition (NER) in new domains is the lack of annotated data.
Chinese Named Entity Recognition named-entity-recognition +5
2 code implementations • 20 Oct 2020 • Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Min Zhang, Shaoping Ma
Through this process, it teaches the DR model how to retrieve relevant documents from the entire corpus instead of how to rerank a potentially biased sample of documents.
4 code implementations • 16 Apr 2021 • Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma
ADORE replaces the widely-adopted static hard negative sampling method with a dynamic one to directly optimize the ranking performance.
5 code implementations • 2 Aug 2021 • Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma
Compared with previous DR models that use brute-force search, JPQ almost matches the best retrieval performance with 30x compression on index size.
4 code implementations • 12 Oct 2021 • Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma
However, the efficiency of most existing DR models is limited by the large memory cost of storing dense vectors and the time-consuming nearest neighbor search (NNS) in vector space.
1 code implementation • 7 Nov 2023 • Dongfang Li, Zetian Sun, Xinshuo Hu, Zhenyu Liu, Ziyang Chen, Baotian Hu, Aiguo Wu, Min Zhang
Open-domain generative systems have gained significant attention in the field of conversational AI (e. g., generative search engines).
3 code implementations • 28 Jun 2020 • Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Min Zhang, Shaoping Ma
Although exact term match between queries and documents is the dominant method to perform first-stage retrieval, we propose a different approach, called RepBERT, to represent documents and queries with fixed-length contextualized embeddings.
1 code implementation • 19 Sep 2023 • Juntao Li, Zecheng Tang, Yuyang Ding, Pinzheng Wang, Pei Guo, Wangjie You, Dan Qiao, Wenliang Chen, Guohong Fu, Qiaoming Zhu, Guodong Zhou, Min Zhang
This report provides the main details to pre-train an analogous model, including pre-training data processing, Bilingual Flan data collection, the empirical observations that inspire our model architecture design, training objectives of different stages, and other enhancement techniques.
1 code implementation • 30 Jul 2019 • Haitao Wang, Zhengqiu He, Jin Ma, Wenliang Chen, Min Zhang
Our data is the first dataset for inter-personal relationship extraction.
1 code implementation • 9 Nov 2023 • Tong Zhu, Junfei Ren, Zijian Yu, Mengsong Wu, Guoliang Zhang, Xiaoye Qu, Wenliang Chen, Zhefeng Wang, Baoxing Huai, Min Zhang
Sharing knowledge between information extraction tasks has always been a challenge due to the diverse data formats and task variations.
1 code implementation • IJCNLP 2019 • Jie Zhu, Junhui Li, Muhua Zhu, Longhua Qian, Min Zhang, Guodong Zhou
Recent studies on AMR-to-text generation often formalize the task as a sequence-to-sequence (seq2seq) learning problem by converting an Abstract Meaning Representation (AMR) graph into a word sequence.
1 code implementation • 22 Oct 2022 • Yue Zhang, Bo Zhang, Zhenghua Li, Zuyi Bao, Chen Li, Min Zhang
Then, we obtain parse trees of the source incorrect sentences by projecting trees of the target correct sentences.
1 code implementation • 24 Mar 2023 • Keqin Peng, Liang Ding, Qihuang Zhong, Li Shen, Xuebo Liu, Min Zhang, Yuanxin Ouyang, DaCheng Tao
We show that: 1) The performance of ChatGPT depends largely on temperature, and a lower temperature usually can achieve better performance; 2) Emphasizing the task information can further improve ChatGPT's performance, particularly in complex MT tasks; 3) Introducing domain information can elicit ChatGPT's generalization ability and improve its performance in the specific domain; 4) ChatGPT tends to generate hallucinations for non-English-centric MT tasks, which can be partially addressed by our proposed prompts but still need to be highlighted for the MT/NLP community.
1 code implementation • 25 May 2023 • Yue Zhang, Bo Zhang, Haochen Jiang, Zhenghua Li, Chen Li, Fei Huang, Min Zhang
We introduce NaSGEC, a new dataset to facilitate research on Chinese grammatical error correction (CGEC) for native speaker texts from multiple domains.
2 code implementations • 26 Jun 2022 • Chenyang Wang, Yuanqing Yu, Weizhi Ma, Min Zhang, Chong Chen, Yiqun Liu, Shaoping Ma
Then, we empirically analyze the learning dynamics of typical CF methods in terms of quantified alignment and uniformity, which shows that better alignment or uniformity both contribute to higher recommendation performance.
1 code implementation • 9 Mar 2019 • Weizhi Ma, Min Zhang, Yue Cao, Woojeong, Jin, Chenyang Wang, Yiqun Liu, Shaoping Ma, Xiang Ren
The framework encourages two modules to complement each other in generating effective and explainable recommendation: 1) inductive rules, mined from item-centric knowledge graphs, summarize common multi-hop relational patterns for inferring different item associations and provide human-readable explanation for model prediction; 2) recommendation module can be augmented by induced rules and thus have better generalization ability dealing with the cold-start issue.
1 code implementation • COLING 2022 • Yu Zhang, Qingrong Xia, Shilin Zhou, Yong Jiang, Guohong Fu, Min Zhang
Semantic role labeling (SRL) is a fundamental yet challenging task in the NLP community.
Dependency Parsing Semantic Role Labeling (predicted predicates)
1 code implementation • 1 Oct 2020 • Ding-Nan Zo, Song-Hai Zhang, Tai-Jiang M, Min Zhang
It is currently the largest dataset for fine-grained classification of dogs, including130 dog breeds and 70, 428 real-world images.
1 code implementation • EMNLP 2016 • Biao Zhang, Deyi Xiong, Jinsong Su, Hong Duan, Min Zhang
Models of neural machine translation are often from a discriminative family of encoderdecoders that learn a conditional distribution of a target sentence given a source sentence.
1 code implementation • 11 Aug 2022 • Jingtao Zhan, Qingyao Ai, Yiqun Liu, Jiaxin Mao, Xiaohui Xie, Min Zhang, Shaoping Ma
By making the REM and DAMs disentangled, DDR enables a flexible training paradigm in which REM is trained with supervision once and DAMs are trained with unsupervised data.
3 code implementations • 20 Aug 2020 • Shaoyun Shi, Hanxiong Chen, Weizhi Ma, Jiaxin Mao, Min Zhang, Yongfeng Zhang
Both reasoning and generalization ability are important for prediction tasks such as recommender systems, where reasoning provides strong connection between user history and target items for accurate prediction, and generalization helps the model to draw a robust user portrait over noisy inputs.
2 code implementations • 22 Nov 2023 • Yilun Liu, Shimin Tao, Xiaofeng Zhao, Ming Zhu, Wenbing Ma, Junhao Zhu, Chang Su, Yutai Hou, Miao Zhang, Min Zhang, Hongxia Ma, Li Zhang, Hao Yang, Yanfei Jiang
Instruction tuning is crucial for enabling Language Learning Models (LLMs) in responding to human instructions.
1 code implementation • 29 Aug 2019 • Haitao Wang, Zhengqiu He, Tong Zhu, Hao Shao, Wenliang Chen, Min Zhang
In this paper, we present the task definition, the description of data and the evaluation methodology used during this shared task.
1 code implementation • 5 May 2023 • Yunxin Li, Baotian Hu, Xinyu Chen, Lin Ma, Yong Xu, Min Zhang
LMEye addresses this issue by allowing the LLM to request the desired visual information aligned with various human instructions, which we term as the dynamic visual information interaction.
1 code implementation • 13 Apr 2023 • Hao Fei, Shengqiong Wu, Jingye Li, Bobo Li, Fei Li, Libo Qin, Meishan Zhang, Min Zhang, Tat-Seng Chua
Universally modeling all typical information extraction tasks (UIE) with one generative language model (GLM) has revealed great potential by the latest study, where various IE predictions are unified into a linearized hierarchical expression under a GLM.
1 code implementation • EMNLP 2020 • Dongqin Xu, Junhui Li, Muhua Zhu, Min Zhang, Guodong Zhou
In the literature, the research on abstract meaning representation (AMR) parsing is much restricted by the size of human-curated dataset which is critical to build an AMR parser with good performance.
Ranked #15 on AMR Parsing on LDC2017T10 (using extra training data)
1 code implementation • IJCNLP 2019 • Xiangyu Duan, Hoongfei Yu, Mingming Yin, Min Zhang, Weihua Luo, Yue Zhang
We propose a contrastive attention mechanism to extend the sequence-to-sequence framework for abstractive sentence summarization task, which aims to generate a brief summary of a given source sentence.
1 code implementation • 23 Feb 2024 • Yuanqing Yu, Chongming Gao, Jiawei Chen, Heng Tang, Yuefeng Sun, Qian Chen, Weizhi Ma, Min Zhang
EasyRL4Rec seeks to facilitate the model development and experimental process in the domain of RL-based RSs.
1 code implementation • 23 May 2023 • Yang Bai, Min Cao, Daming Gao, Ziqiang Cao, Chen Chen, Zhenfeng Fan, Liqiang Nie, Min Zhang
RA offsets the overfitting risk by introducing a novel positive relation detection task (i. e., learning to distinguish strong and weak positive pairs).
Ranked #2 on Text based Person Retrieval on RSTPReid
1 code implementation • 15 Mar 2021 • Pranav Kadam, Min Zhang, Shan Liu, C. -C. Jay Kuo
Inspired by the recent PointHop classification method, an unsupervised 3D point cloud registration method, called R-PointHop, is proposed in this work.
1 code implementation • 18 Jan 2022 • Chong Chen, Fei Sun, Min Zhang, Bolin Ding
From the perspective of utility, if a system's utility is damaged by some bad data, the system needs to forget these data to regain utility.
1 code implementation • EMNLP 2021 • Xincheng Ju, Dong Zhang, Rong Xiao, Junhui Li, Shoushan Li, Min Zhang, Guodong Zhou
Therefore, in this paper, we are the first to jointly perform multi-modal ATE (MATE) and multi-modal ASC (MASC), and we propose a multi-modal joint learning approach with auxiliary cross-modal relation detection for multi-modal aspect-level sentiment analysis (MALSA).
1 code implementation • 19 Aug 2023 • Min Cao, Yang Bai, Ziyin Zeng, Mang Ye, Min Zhang
TPBS, as a fine-grained cross-modal retrieval task, is also facing the rise of research on the CLIP-based TBPS.
Ranked #4 on Text based Person Retrieval on RSTPReid
1 code implementation • 8 May 2023 • Zecheng Tang, Pinzheng Wang, Keyan Zhou, Juntao Li, Ziqiang Cao, Min Zhang
Diffusion models have been successfully adapted to text generation tasks by mapping the discrete text into the continuous space.
1 code implementation • COLING 2020 • Tong Zhu, Haitao Wang, Junjie Yu, Xiabing Zhou, Wenliang Chen, Wei zhang, Min Zhang
The experimental results show that the ranking lists of the comparison systems on the DS-labelled test data and human-annotated test data are different.
1 code implementation • 23 Oct 2023 • Houquan Zhou, Yumeng Liu, Zhenghua Li, Min Zhang, Bo Zhang, Chen Li, Ji Zhang, Fei Huang
In this paper, we propose a unified decoding intervention framework that employs an external critic to assess the appropriateness of the token to be generated incrementally, and then dynamically influence the choice of the next token.
Ranked #1 on Grammatical Error Correction on MuCGEC
1 code implementation • ACL 2019 • Xiangyu Duan, Mingming Yin, Min Zhang, Boxing Chen, Weihua Luo
But there is no cross-lingual parallel corpus, whose source sentence language is different to the summary language, to directly train a cross-lingual ASSUM system.
1 code implementation • COLING 2022 • Dan Qiao, Chenchen Dai, Yuyang Ding, Juntao Li, Qiang Chen, Wenliang Chen, Min Zhang
The conventional success of textual classification relies on annotated data, and the new paradigm of pre-trained language models (PLMs) still requires a few labeled data for downstream tasks.
2 code implementations • 1 Jul 2020 • Chong Chen, Min Zhang, Weizhi Ma, Yiqun Liu, and Shaoping Ma
However, existing KG enhanced recommendation methods have largely focused on exploring advanced neural network architectures to better investigate the structural information of KG.
1 code implementation • NAACL 2022 • Linzhi Wu, Pengjun Xie, Jie zhou, Meishan Zhang, Chunping Ma, Guangwei Xu, Min Zhang
Prior research has mainly resorted to heuristic rule-based constraints to reduce the noise for specific self-augmentation methods individually.
1 code implementation • ACL 2022 • Bei Li, Quan Du, Tao Zhou, Yi Jing, Shuhan Zhou, Xin Zeng, Tong Xiao, Jingbo Zhu, Xuebo Liu, Min Zhang
Inspired by this, we design a new architecture, {\it ODE Transformer}, which is analogous to the Runge-Kutta method that is well motivated in ODE.
1 code implementation • 28 Nov 2023 • Longhui Zhang, Yanzhao Zhang, Dingkun Long, Pengjun Xie, Meishan Zhang, Min Zhang
Text ranking is a critical task in various information retrieval applications, and the recent success of pre-trained language models (PLMs), especially large language models (LLMs), has sparked interest in their application to text ranking.
1 code implementation • 26 Feb 2024 • Yuyang Ding, Juntao Li, Pinzheng Wang, Zecheng Tang, Bowen Yan, Min Zhang
In the Named Entity Recognition (NER) task, recent advancements have seen the remarkable improvement of LLMs in a broad range of entity domains via instruction tuning, by adopting entity-centric schema.
1 code implementation • Findings (ACL) 2021 • Hongqiu Wu, Hai Zhao, Min Zhang
Code summarization (CS) is becoming a promising area in recent language understanding, which aims to generate sensible human language automatically for programming language in the format of source code, serving in the most convenience of programmer developing.
1 code implementation • CVPR 2022 • Yifeng Xiong, Jiadong Lin, Min Zhang, John E. Hopcroft, Kun He
The black-box adversarial attack has attracted impressive attention for its practical use in the field of deep learning security.
1 code implementation • 20 Oct 2022 • Yu Zhao, Jianguo Wei, Zhichao Lin, Yueheng Sun, Meishan Zhang, Min Zhang
Accordingly, we manually annotate a dataset to facilitate the investigation of the newly-introduced task and build several benchmark encoder-decoder models by using VL-BART and VL-T5 as backbones.
1 code implementation • 19 May 2023 • Yu Zhao, Hao Fei, Wei Ji, Jianguo Wei, Meishan Zhang, Min Zhang, Tat-Seng Chua
With an external 3D scene extractor, we obtain the 3D objects and scene features for input images, based on which we construct a target object-centered 3D spatial scene graph (Go3D-S2G), such that we model the spatial semantics of target objects within the holistic 3D scenes.
1 code implementation • 17 Nov 2023 • Shenghao Yang, Chenyang Wang, Yankai Liu, Kangping Xu, Weizhi Ma, Yiqun Liu, Min Zhang, Haitao Zeng, Junlan Feng, Chao Deng
In this paper, we propose CoWPiRec, an approach of Collaborative Word-based Pre-trained item representation for Recommendation.
3 code implementations • 30 Jul 2019 • Min Zhang, Haoxuan You, Pranav Kadam, Shan Liu, C. -C. Jay Kuo
In the attribute building stage, we address the problem of unordered point cloud data using a space partitioning procedure and developing a robust descriptor that characterizes the relationship between a point and its one-hop neighbor in a PointHop unit.
1 code implementation • 23 Nov 2022 • Zhijun Wang, Xuebo Liu, Min Zhang
Existing research generally treats Chinese character as a minimum unit for representation.
Ranked #1 on Machine Translation on WMT2017 Chinese-English
1 code implementation • Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence 2021 • An-Hui Wang, Linfeng Song, Hui Jiang, Shaopeng Lai, Junfeng Yao, Min Zhang, Jinsong Su
Conversational discourse structures aim to describe how a dialogue is organised, thus they are helpful for dialogue understanding and response generation.
Ranked #3 on Discourse Parsing on STAC
1 code implementation • 3 Nov 2022 • Peiyuan Gong, Xuebo Liu, Heyan Huang, Min Zhang
Pretraining-based (PT-based) automatic evaluation metrics (e. g., BERTScore and BARTScore) have been widely used in several sentence generation tasks (e. g., machine translation and text summarization) due to their better correlation with human judgments over traditional overlap-based methods.
1 code implementation • 22 Jun 2023 • Senbao Shi, Zhenran Xu, Baotian Hu, Min Zhang
Multimodal Entity Linking (MEL) is the task of mapping mentions with multimodal contexts to the referent entities from a knowledge base.
2 code implementations • 15 Apr 2023 • Jiayu Li, Peijie Sun, Zhefan Wang, Weizhi Ma, Yangkun Li, Min Zhang, Zhoutian Feng, Daiyue Xue
To address such a task, we propose an Intent-aware ranking Ensemble Learning~(IntEL) model to fuse multiple single-objective item lists with various user intents, in which item-level personalized weights are learned.
1 code implementation • EMNLP 2016 • Biao Zhang, Deyi Xiong, Jinsong Su, Qun Liu, Rongrong Ji, Hong Duan, Min Zhang
In order to perform efficient inference and learning, we introduce neural discourse relation models to approximate the prior and posterior distributions of the latent variable, and employ these approximated distributions to optimize a reparameterized variational lower bound.
1 code implementation • 22 Jul 2019 • Qingrong Xia, Zhenghua Li, Min Zhang, Meishan Zhang, Guohong Fu, Rui Wang, Luo Si
Semantic role labeling (SRL), also known as shallow semantic parsing, is an important yet challenging task in NLP.
1 code implementation • COLING 2022 • Shilin Zhou, Qingrong Xia, Zhenghua Li, Yu Zhang, Yu Hong, Min Zhang
Moreover, we propose a simple constrained Viterbi procedure to ensure the legality of the output graph according to the constraints of the SRL structure.
1 code implementation • 8 Dec 2022 • Zhaocong Li, Xuebo Liu, Derek F. Wong, Lidia S. Chao, Min Zhang
In this paper, we propose a novel transfer learning method for NMT, namely ConsistTL, which can continuously transfer knowledge from the parent model during the training of the child model.
1 code implementation • 15 Dec 2023 • Shuanghao Bai, Min Zhang, Wanqi Zhou, Siteng Huang, Zhirong Luan, Donglin Wang, Badong Chen
Therefore, in this paper, we first experimentally demonstrate that the unsupervised-trained VLMs can significantly reduce the distribution discrepancy between source and target domains, thereby improving the performance of UDA.
1 code implementation • 19 Dec 2023 • Pengxiang Ding, Qiongjie Cui, Min Zhang, Mengyuan Liu, Haofan Wang, Donglin Wang
Human motion forecasting, with the goal of estimating future human behavior over a period of time, is a fundamental task in many real-world applications.
1 code implementation • 22 Apr 2024 • Jiayin Wang, Fengran Mo, Weizhi Ma, Peijie Sun, Min Zhang, Jian-Yun Nie
To address this oversight, we propose benchmarking LLMs from a user perspective in both dataset construction and evaluation designs.
2 code implementations • 9 Feb 2020 • Min Zhang, Yifan Wang, Pranav Kadam, Shan Liu, C. -C. Jay Kuo
The PointHop method was recently proposed by Zhang et al. for 3D point cloud classification with unsupervised feature extraction.
1 code implementation • ACL 2020 • Xiangyu Duan, Baijun Ji, Hao Jia, Min Tan, Min Zhang, Boxing Chen, Weihua Luo, Yue Zhang
In this paper, we propose a new task of machine translation (MT), which is based on no parallel sentences but can refer to a ground-truth bilingual dictionary.
1 code implementation • NAACL 2021 • Qingrong Xia, Bo Zhang, Rui Wang, Zhenghua Li, Yue Zhang, Fei Huang, Luo Si, Min Zhang
Fine-grained opinion mining (OM) has achieved increasing attraction in the natural language processing (NLP) community, which aims to find the opinion structures of {``}Who expressed what opinions towards what{''} in one sentence.
1 code implementation • ACL 2019 • Zhenghua Li, Xue Peng, Min Zhang, Rui Wang, Luo Si
During the past decades, due to the lack of sufficient labeled data, most studies on cross-domain parsing focus on unsupervised domain adaptation, assuming there is no target-domain training data.
1 code implementation • IJCNLP 2019 • Qingrong Xia, Zhenghua Li, Min Zhang
In this paper, we adopt a simple unified span-based model for both span-based and word-based Chinese SRL as a strong baseline.
1 code implementation • 12 Jul 2020 • Feiyu Yang, Zhan Song, Zhenzhong Xiao, Yu Chen, Zhe Pan, Min Zhang, Min Xue, Yaoyang Mo, Yao Zhang, Guoxiong Guan, Beibei Qian
Recently, the leading performance of human pose estimation is dominated by heatmap based methods.
1 code implementation • 23 Feb 2024 • Zhefan Wang, Yuanqing Yu, Wendi Zheng, Weizhi Ma, Min Zhang
LLM-based agents have gained considerable attention for their decision-making skills and ability to handle complex tasks.
1 code implementation • Findings (ACL) 2021 • Jinpeng Zhang, Baijun Ji, Nini Xiao, Xiangyu Duan, Min Zhang, Yangbin Shi, Weihua Luo
Bilingual Lexicon Induction (BLI) aims to map words in one language to their translations in another, and is typically through learning linear projections to align monolingual word representation spaces.
1 code implementation • Findings (ACL) 2022 • Yiming Zhang, Min Zhang, Sai Wu, Junbo Zhao
The aspect-based sentiment analysis (ABSA) is a fine-grained task that aims to determine the sentiment polarity towards targeted aspect terms occurring in the sentence.
Ranked #4 on Aspect-Based Sentiment Analysis (ABSA) on SemEval-2014 Task-4 (using extra training data)
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +3
1 code implementation • 8 May 2023 • Yunxin Li, Baotian Hu, Xinyu Chen, Yuxin Ding, Lin Ma, Min Zhang
This makes the language model well-suitable for such multi-modal reasoning scenario on joint textual and visual clues.
1 code implementation • 12 Oct 2023 • Xin Zhang, Zehan Li, Yanzhao Zhang, Dingkun Long, Pengjun Xie, Meishan Zhang, Min Zhang
As such cases span from English to other natural or programming languages, from retrieval to classification and beyond, it is desirable to build a unified embedding model rather than dedicated ones for each scenario.
1 code implementation • ACL 2021 • Chen Gong, Saihao Huang, Houquan Zhou, Zhenghua Li, Min Zhang, Zhefeng Wang, Baoxing Huai, Nicholas Jing Yuan
Several previous works on syntactic parsing propose to annotate shallow word-internal structures for better utilizing character-level information.
1 code implementation • ACL 2022 • Nan Yu, Meishan Zhang, Guohong Fu, Min Zhang
Pre-trained language models (PLMs) have shown great potentials in natural language processing (NLP) including rhetorical structure theory (RST) discourse parsing. Current PLMs are obtained by sentence-level pre-training, which is different from the basic processing unit, i. e. element discourse unit (EDU). To this end, we propose a second-stage EDU-level pre-training approach in this work, which presents two novel tasks to learn effective EDU representations continually based on well pre-trained language models. Concretely, the two tasks are (1) next EDU prediction (NEP) and (2) discourse marker prediction (DMP). We take a state-of-the-art transition-based neural parser as baseline, and adopt it with a light bi-gram EDU modification to effectively explore the EDU-level pre-trained EDU representation. Experimental results on a benckmark dataset show that our method is highly effective, leading a 2. 1-point improvement in F1-score. All codes and pre-trained models will be released publicly to facilitate future studies.
1 code implementation • 17 Aug 2022 • Ziyi Ye, Xiaohui Xie, Yiqun Liu, Zhihong Wang, Xuesong Chen, Min Zhang, Shaoping Ma
We explore the effectiveness of BTA for satisfaction modeling in two popular information access scenarios, i. e., search and recommendation.
1 code implementation • 31 Oct 2022 • Zhaochen Su, Zecheng Tang, Xinyan Guan, Juntao Li, Lijun Wu, Min Zhang
Existing methods mainly perform continual training to mitigate such a misalignment.
1 code implementation • 4 Oct 2023 • Murong Yue, Jie Zhao, Min Zhang, Liang Du, Ziyu Yao
Large language models (LLMs) such as GPT-4 have exhibited remarkable performance in a variety of tasks, but this strong performance often comes with the high expense of using paid API services.
1 code implementation • 6 Apr 2022 • Zhumin Chu, Qingyao Ai, Zhihong Wang, Yiqun Liu, Yingye Huang, Rui Zhang, Min Zhang, Shaoping Ma
This raises the question of how to accurately model user satisfaction in conversational search scenarios.
1 code implementation • 2 Dec 2022 • Hexuan Deng, Liang Ding, Xuebo Liu, Meishan Zhang, DaCheng Tao, Min Zhang
Preliminary experiments on En-Zh and En-Ja news domain corpora demonstrate that monolingual data can significantly improve translation quality (e. g., +3. 15 BLEU on En-Zh).
1 code implementation • 12 Mar 2023 • Zhengrui Ma, Chenze Shao, Shangtong Gui, Min Zhang, Yang Feng
Non-autoregressive translation (NAT) reduces the decoding latency but suffers from performance degradation due to the multi-modality problem.
2 code implementations • 5 Oct 2018 • C. -C. Jay Kuo, Min Zhang, Siyang Li, Jiali Duan, Yueru Chen
To construct convolutional layers, we develop a new signal transform, called the Saab (Subspace Approximation with Adjusted Bias) transform.
1 code implementation • 11 Sep 2023 • Linhan Wang, Shuo Lei, Jianfeng He, Shengkun Wang, Min Zhang, Chang-Tien Lu
To tackle these challenges, we propose a Self-Correlation and Cross-Correlation Learning Network for the few-shot remote sensing image semantic segmentation.
1 code implementation • 20 Feb 2024 • Shoutao Guo, Shaolei Zhang, Zhengrui Ma, Min Zhang, Yang Feng
We propose SiLLM, which delegates the two sub-tasks to separate agents, thereby incorporating LLM into SiMT.
1 code implementation • 10 Sep 2020 • Siteng Huang, Min Zhang, Yachen Kang, Donglin Wang
However, these approaches only augment the representations of samples with available semantics while ignoring the query set, which loses the potential for the improvement and may lead to a shift between the modalities combination and the pure-visual representation.
2 code implementations • 13 Dec 2021 • Chong Liu, Xiaoyang Liu, Rongqin Zheng, Lixin Zhang, Xiaobo Liang, Juntao Li, Lijun Wu, Min Zhang, Leyu Lin
State-of-the-art sequential recommendation models proposed very recently combine contrastive learning techniques for obtaining high-quality user representations.
1 code implementation • 14 Jul 2022 • Min Zhang, Siteng Huang, Wenbin Li, Donglin Wang
To solve this problem, we present a plug-in Hierarchical Tree Structure-aware (HTS) method, which not only learns the relationship of FSL and pretext tasks, but more importantly, can adaptively select and aggregate feature representations generated by pretext tasks to maximize the performance of FSL tasks.
1 code implementation • 16 Dec 2022 • Qian Yang, Qian Chen, Wen Wang, Baotian Hu, Min Zhang
Moreover, the pipelined approaches of retrieval and generation might result in poor generation performance when retrieval performance is low.
1 code implementation • 27 Mar 2023 • Siteng Huang, Biao Gong, Yutong Feng, Min Zhang, Yiliang Lv, Donglin Wang
Recent compositional zero-shot learning (CZSL) methods adapt pre-trained vision-language models (VLMs) by constructing trainable prompts only for composed state-object pairs.
2 code implementations • 16 Aug 2023 • Zecheng Tang, Keyan Zhou, Juntao Li, Yuyang Ding, Pinzheng Wang, Bowen Yan, Min Zhang
In view of this, we introduce a Context-aware Model self-Detoxification~(CMD) framework that pays attention to both the context and the detoxification process, i. e., first detoxifying the context and then making the language model generate along the safe context.
1 code implementation • 13 Dec 2021 • Shiping Li, Min Cao, Min Zhang
In this paper, we propose a semantic-aligned embedding method for text-based person search, in which the feature alignment across modalities is achieved by automatically learning the semantic-aligned visual features and textual features.
Ranked #9 on Text based Person Retrieval on CUHK-PEDES
1 code implementation • 25 Apr 2022 • Jingtao Zhan, Xiaohui Xie, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma
For example, representation-based retrieval models perform almost as well as interaction-based retrieval models in terms of interpolation but not extrapolation.
1 code implementation • COLING 2022 • Nan Yu, Guohong Fu, Min Zhang
It is believed that speaker interactions are helpful for this task.
Ranked #2 on Discourse Parsing on STAC
1 code implementation • 26 May 2023 • Zhiwei Cao, Baosong Yang, Huan Lin, Suhang Wu, Xiangpeng Wei, Dayiheng Liu, Jun Xie, Min Zhang, Jinsong Su
$k$-Nearest neighbor machine translation ($k$NN-MT) has attracted increasing attention due to its ability to non-parametrically adapt to new translation domains.
1 code implementation • 23 Oct 2023 • Zhengrui Ma, Shaolei Zhang, Shoutao Guo, Chenze Shao, Min Zhang, Yang Feng
Simultaneous machine translation (SiMT) models are trained to strike a balance between latency and translation quality.
1 code implementation • 14 Nov 2023 • Zhenran Xu, Senbao Shi, Baotian Hu, Jindi Yu, Dongfang Li, Min Zhang, Yuxiang Wu
Large Language Models (LLMs) have shown remarkable capabilities in general natural language processing tasks but often fall short in complex reasoning tasks.
1 code implementation • 26 Feb 2024 • Liangxin Liu, Xuebo Liu, Derek F. Wong, Dongfang Li, Ziyi Wang, Baotian Hu, Min Zhang
In this work, we propose a novel approach, termed SelectIT, that capitalizes on the foundational capabilities of the LLM itself.
1 code implementation • 30 Mar 2024 • Hongqiu Wu, Y. Wang, XingYuan Liu, Hai Zhao, Min Zhang
The Instruction-Driven Game Engine (IDGE) project aims to democratize game development by enabling a large language model (LLM) to follow free-form game rules and autonomously generate game-play processes.
1 code implementation • 8 May 2023 • Hongqiu Wu, Yongxiang Liu, Hanwen Shi, Hai Zhao, Min Zhang
Based on the observation, we propose simple yet effective \textit{Contextualized representation-Adversarial Training} (CreAT), in which the attack is explicitly optimized to deviate the contextualized representation of the encoder.
1 code implementation • 27 Mar 2024 • Shenghao Yang, Weizhi Ma, Peijie Sun, Qingyao Ai, Yiqun Liu, Mingchen Cai, Min Zhang
Different from previous relation-aware models that rely on predefined rules, we propose to leverage the Large Language Model (LLM) to provide new types of relations and connections between items.
1 code implementation • NAACL 2019 • Kai Song, Yue Zhang, Heng Yu, Weihua Luo, Kun Wang, Min Zhang
Leveraging user-provided translation to constrain NMT has practical significance.
1 code implementation • 22 Dec 2021 • Changxing Wu, Liuwen Cao, Yubin Ge, Yang Liu, Min Zhang, Jinsong Su
Then, we employ a label sequence decoder to output the predicted labels in a top-down manner, where the predicted higher-level labels are directly used to guide the label prediction at the current level.
1 code implementation • Findings (ACL) 2022 • Houquan Zhou, Yang Li, Zhenghua Li, Min Zhang
In recent years, large-scale pre-trained language models (PLMs) have made extraordinary progress in most NLP tasks.
1 code implementation • 13 Nov 2022 • Binbin Xie, Xiangpeng Wei, Baosong Yang, Huan Lin, Jun Xie, Xiaoli Wang, Min Zhang, Jinsong Su
Keyphrase generation aims to automatically generate short phrases summarizing an input document.
1 code implementation • CVPR 2023 • Zhaodi Zhang, Zhiyi Xue, Yang Chen, Si Liu, Yueling Zhang, Jing Liu, Min Zhang
Via abstraction, all perturbed images are mapped into intervals before feeding into neural networks for training.
1 code implementation • 3 May 2023 • Yunxin Li, Baotian Hu, Yuxin Ding, Lin Ma, Min Zhang
Inspired by the Divide-and-Conquer algorithm and dual-process theory, in this paper, we regard linguistically complex texts as compound proposition texts composed of multiple simple proposition sentences and propose an end-to-end Neural Divide-and-Conquer Reasoning framework, dubbed NDCR.
1 code implementation • 20 May 2023 • Hao Fei, Qian Liu, Meishan Zhang, Min Zhang, Tat-Seng Chua
In this work, we investigate a more realistic unsupervised multimodal machine translation (UMMT) setup, inference-time image-free UMMT, where the model is trained with source-text image pairs, and tested with only source-text inputs.
1 code implementation • 16 Nov 2023 • Ziyi Ye, Qingyao Ai, Yiqun Liu, Maarten de Rijke, Min Zhang, Christina Lioma, Tuukka Ruotsalo
Inspired by recent research that revealed associations between the brain and the large computational language models, we propose a generative language BCI that utilizes the capacity of a large language model (LLM) jointly with a semantic brain decoder to directly generate language from functional magnetic resonance imaging (fMRI) input.
1 code implementation • COLING 2020 • Junjie Yu, Tong Zhu, Wenliang Chen, Wei zhang, Min Zhang
In this paper, we propose an alternative approach to improve RE systems via enriching diverse expressions by relational paraphrase sentences.
1 code implementation • Findings (NAACL) 2022 • Huan Lin, Baosong Yang, Liang Yao, Dayiheng Liu, Haibo Zhang, Jun Xie, Min Zhang, Jinsong Su
Diverse NMT aims at generating multiple diverse yet faithful translations given a source sentence.
1 code implementation • 8 Nov 2022 • Jinpeng Zhang, Chuanqi Dong, Xiangyu Duan, Yuqi Zhang, Min Zhang
Word alignment is to find translationally equivalent words between source and target sentences.
1 code implementation • 14 Nov 2023 • Houquan Zhou, Yang Hou, Zhenghua Li, Xuebin Wang, Zhefeng Wang, Xinyu Duan, Min Zhang
While recent advancements in large language models (LLMs) bring us closer to achieving artificial general intelligence, the question persists: Do LLMs truly understand language, or do they merely mimic comprehension through pattern recognition?
1 code implementation • 18 Mar 2024 • Xiaojie Li, Yibo Yang, Xiangtai Li, Jianlong Wu, Yue Yu, Bernard Ghanem, Min Zhang
To tackle these challenges, we present GenView, a controllable framework that augments the diversity of positive views leveraging the power of pretrained generative models while preserving semantics.
1 code implementation • 17 Apr 2024 • Dongfang Li, Zhenyu Liu, Xinshuo Hu, Zetian Sun, Baotian Hu, Min Zhang
In this paper, we address this gap by presenting a comprehensive analysis of these compressed vectors, drawing parallels to the parameters trained with gradient descent, and introduce the concept of state vector.
1 code implementation • COLING 2020 • Huaao Zhang, Shigui Qiu, Xiangyu Duan, Min Zhang
Neural machine translation with millions of parameters is vulnerable to unfamiliar inputs.
1 code implementation • COLING 2020 • Qingrong Xia, Rui Wang, Zhenghua Li, Yue Zhang, Min Zhang
Recently, due to the interplay between syntax and semantics, incorporating syntactic knowledge into neural semantic role labeling (SRL) has achieved much attention.
1 code implementation • 25 Jun 2022 • Hongqiu Wu, Ruixue Ding, Hai Zhao, Pengjun Xie, Fei Huang, Min Zhang
Deep neural models (e. g. Transformer) naturally learn spurious features, which create a ``shortcut'' between the labels and inputs, thus impairing the generalization and robustness.
Ranked #1 on Machine Reading Comprehension on DREAM
Machine Reading Comprehension Named Entity Recognition (NER) +4
1 code implementation • COLING 2022 • Zhongjian Miao, Xiang Li, Liyan Kang, Wen Zhang, Chulun Zhou, Yidong Chen, Bin Wang, Min Zhang, Jinsong Su
Most existing methods on robust neural machine translation (NMT) construct adversarial examples by injecting noise into authentic examples and indiscriminately exploit two types of examples.
1 code implementation • 13 Mar 2023 • Yisheng Xiao, Ruiyang Xu, Lijun Wu, Juntao Li, Tao Qin, Yan-Tie Liu, Min Zhang
Experiments on \textbf{3} different tasks (neural machine translation, summarization, and code generation) with \textbf{15} datasets in total confirm that our proposed simple method achieves significant performance improvement over the strong CMLM model.
1 code implementation • 22 May 2023 • Dongfang Li, Jindi Yu, Baotian Hu, Zhenran Xu, Min Zhang
As ChatGPT and GPT-4 spearhead the development of Large Language Models (LLMs), more researchers are investigating their performance across various tasks.
1 code implementation • 19 Oct 2023 • Zhenran Xu, Yulin Chen, Baotian Hu, Min Zhang
Zero-shot entity linking (EL) aims at aligning entity mentions to unseen entities to challenge the generalization ability.
1 code implementation • 19 Oct 2023 • Yulin Chen, Zhenran Xu, Baotian Hu, Min Zhang
Entity linking aims to link ambiguous mentions to their corresponding entities in a knowledge base.
1 code implementation • 21 Feb 2024 • Yunxin Li, Baotian Hu, Wenhan Luo, Lin Ma, Yuxin Ding, Min Zhang
For this setting, previous methods utilize visual and textual encoders to encode the image and keywords and employ a language model-based decoder to generate the product description.
1 code implementation • 8 May 2024 • Yunxin Li, Baotian Hu, Haoyuan Shi, Wei Wang, Longyue Wang, Min Zhang
Large Multimodal Models (LMMs) have achieved impressive success in visual understanding and reasoning, remarkably improving the performance of mathematical reasoning in a visual context.
1 code implementation • NAACL 2022 • Yahui Liu, Haoping Yang, Chen Gong, Qingrong Xia, Zhenghua Li, Min Zhang
1) Based on a frame-free annotation methodology, we avoid writing complex frames for new predicates.
1 code implementation • 23 Jul 2022 • Qian Yang, Yunxin Li, Baotian Hu, Lin Ma, Yuxing Ding, Min Zhang
CSI), a relation inferrer, and a Lexical Constraint-aware Generator (arr.
1 code implementation • 25 Jul 2023 • Hexuan Deng, Xin Zhang, Meishan Zhang, Xuebo Liu, Min Zhang
In this paper, we conduct a holistic exploration of the Universal Decompositional Semantic (UDS) Parsing.
1 code implementation • 9 Dec 2023 • Ziyi Ye, Xiaohui Xie, Qingyao Ai, Yiqun Liu, Zhihong Wang, Weihang Su, Min Zhang
To explore the effectiveness of brain signals in the context of RF, we propose a novel RF framework that combines BCI-based relevance feedback with pseudo-relevance signals and implicit signals to improve the performance of document re-ranking.
1 code implementation • 9 May 2024 • Dan Qiao, Yi Su, Pinzheng Wang, Jing Ye, Wenjing Xie, Yuechi Zhou, Yuyang Ding, Zecheng Tang, Jikai Wang, Yixin Ji, Yue Wang, Pei Guo, Zechen Sun, Zikang Zhang, Juntao Li, Pingfu Chao, Wenliang Chen, Guohong Fu, Guodong Zhou, Qiaoming Zhu, Min Zhang
Large Language Models (LLMs) have played an important role in many fields due to their powerful capabilities. However, their massive number of parameters leads to high deployment requirements and incurs significant inference costs, which impedes their practical applications.
1 code implementation • EMNLP 2018 • Yang Xu, Yu Hong, Huibin Ruan, Jianmin Yao, Min Zhang, Guodong Zhou
We tackle discourse-level relation recognition, a problem of determining semantic relations between text spans.
1 code implementation • 30 Oct 2020 • Yangyang Guo, Liqiang Nie, Zhiyong Cheng, Qi Tian, Min Zhang
Concretely, we design a novel interpretation scheme whereby the loss of mis-predicted frequent and sparse answers of the same question type is distinctly exhibited during the late training phase.
1 code implementation • CoNLL (EMNLP) 2021 • Yang Hou, Houquan Zhou, Zhenghua Li, Yu Zhang, Min Zhang, Zhefeng Wang, Baoxing Huai, Nicholas Jing Yuan
In the coarse labeling stage, the joint model outputs a bracketed tree, in which each node corresponds to one of four labels (i. e., phrase, subphrase, word, subword).
1 code implementation • 24 Jul 2022 • Min Zhang, Zhihong Pan, Xin Zhou, C. -C. Jay Kuo
Normalizing flow models have been used successfully for generative image super-resolution (SR) by approximating complex distribution of natural images to simple tractable distribution in latent space through Invertible Neural Networks (INN).
1 code implementation • 22 Dec 2022 • Zhaoxin Fan, Kaixing Yang, Min Zhang, Zhenbo Song, Hongyan Liu, Jun He
In stage 1, a novel devices detection and tracking scheme is introduced, which accurately locate the height limit devices in the left or right image.
1 code implementation • 21 May 2023 • Gongyao Jiang, Shuang Liu, Meishan Zhang, Min Zhang
Dialogue-level dependency parsing has received insufficient attention, especially for Chinese.
1 code implementation • 24 Jul 2023 • Yifan Wang, Peijie Sun, Min Zhang, Qinglin Jia, Jingjie Li, Shaoping Ma
To directly introduce the correct feedback label information, we propose an Unbiased delayed feedback Label Correction framework (ULC), which uses an auxiliary model to correct labels for observed negative feedback samples.
1 code implementation • 16 Aug 2023 • Xinshuo Hu, Dongfang Li, Baotian Hu, Zihao Zheng, Zhenyu Liu, Min Zhang
To evaluate the effectiveness of our approach in terms of truthfulness and detoxification, we conduct extensive experiments on LLMs, encompassing additional abilities such as language modeling and mathematical reasoning.
1 code implementation • 16 Oct 2023 • Haoke Zhang, Yue Wang, Juntao Li, Xiabing Zhou, Min Zhang
Large Language Models~(LLMs) have demonstrated incredible capabilities in understanding, generating, and manipulating languages.
1 code implementation • 20 Oct 2023 • Zecheng Tang, Kaifeng Qi, Juntao Li, Min Zhang
By leveraging the augmenting data from the GEC models themselves in the post-training process and introducing regularization data for cycle training, our proposed method can effectively improve the model robustness of well-trained GEC models with only a few more training epochs as an extra cost.
1 code implementation • 5 Dec 2023 • Xinyu Ma, Xuebo Liu, Min Zhang
In multilingual translation research, the comprehension and utilization of language families are of paramount importance.
1 code implementation • 27 Mar 2024 • Zhefan Wang, Weizhi Ma, Min Zhang
First, we propose and define the recommendability identification task, which investigates the need for recommendations in the current conversational context.
1 code implementation • 27 Mar 2024 • Shenghao Yang, Weizhi Ma, Peijie Sun, Min Zhang, Qingyao Ai, Yiqun Liu, Mingchen Cai
Knowledge-based recommendation models effectively alleviate the data sparsity issue leveraging the side information in the knowledge graph, and have achieved considerable performance.
1 code implementation • COLING 2018 • Yachao Li, Junhui Li, Min Zhang
In the popular sequence to sequence (seq2seq) neural machine translation (NMT), there exist many weighted sum models (WSMs), each of which takes a set of input and generates one output.
1 code implementation • 6 Mar 2020 • Houquan Zhou, Yu Zhang, Zhenghua Li, Min Zhang
In the pre deep learning era, part-of-speech tags have been considered as indispensable ingredients for feature engineering in dependency parsing.
1 code implementation • 3 Aug 2021 • Ziyi Ye, Xiaohui Xie, Yiqun Liu, Zhihong Wang, Xuesong Chen, Min Zhang, Shaoping Ma
In this paper, we carefully design a lab-based user study to investigate brain activities during reading comprehension.
1 code implementation • 25 May 2022 • Yang Xu, Yutai Hou, Wanxiang Che, Min Zhang
On the newly defined cross-lingual model editing task, we empirically demonstrate the failure of monolingual baselines in propagating the edit to multiple languages and the effectiveness of the proposed language anisotropic model editing.
1 code implementation • 17 Jul 2023 • Jiayin Wang, Weizhi Ma, Chumeng Jiang, Min Zhang, Yuan Zhang, Biao Li, Peng Jiang
In this paper, we call for a shift of attention from modeling user preferences to developing fair exposure mechanisms for items.
1 code implementation • ICCV 2023 • Min Zhang, Junkun Yuan, Yue He, Wenbin Li, Zhengyu Chen, Kun Kuang
To achieve this goal, we apply a bilevel optimization to explicitly model and optimize the coupling relationship between the OOD model and auxiliary adapter layers.
1 code implementation • 5 Nov 2023 • Jianling Li, Meishan Zhang, Peiming Guo, Min Zhang, Yue Zhang
Our experimental results demonstrate that self-training for constituency parsing, equipped with an LLM, outperforms traditional methods regardless of the LLM's performance.
1 code implementation • 13 Nov 2023 • Meizhi Zhong, Lemao Liu, Kehai Chen, Mingming Yang, Min Zhang
Simultaneous Machine Translation (SiMT) aims to yield a real-time partial translation with a monotonically growing the source-side context.
1 code implementation • 27 Mar 2024 • Jiayu Li, Peijie Sun, Chumeng Jiang, Weizhi Ma, Qingyao Ai, Min Zhang
In this paper, we provide a new perspective that takes situations as the preconditions for users' interactions.
1 code implementation • 29 Apr 2024 • Xinyu Ma, Xuebo Liu, Derek F. Wong, Jun Rao, Bei Li, Liang Ding, Lidia S. Chao, DaCheng Tao, Min Zhang
Experimental results show that MMT models trained on our dataset exhibit a greater ability to exploit visual information than those trained on other MMT datasets.
1 code implementation • Findings (EMNLP) 2021 • Qingrong Xia, Zhenghua Li, Rui Wang, Min Zhang
In particular, one recent seq-to-seq work directly fine-tunes AMR graph sequences on the encoder-decoder pre-trained language model and achieves new state-of-the-art results, outperforming previous works by a large margin.
1 code implementation • ACL 2022 • Xin Zhang, Guangwei Xu, Yueheng Sun, Meishan Zhang, Xiaobin Wang, Min Zhang
Recent works of opinion expression identification (OEI) rely heavily on the quality and scale of the manually-constructed training corpus, which could be extremely difficult to satisfy.
1 code implementation • 23 Oct 2022 • Panzhong Lu, Xin Zhang, Meishan Zhang, Min Zhang
First, we construct a dataset of phrase grounding with both noun phrases and pronouns to image regions.
2 code implementations • ACM Multimedia 2022 • Meiyu Liang, Junping Du, Xiaowen Cao, Yang Yu, Kangkang Lu, Zhe Xue, Min Zhang
Secondly, for further improving learning ability of implicit cross-media semantic associations, a semantic label association graph is constructed, and the graph convolutional network is utilized to mine the implicit semantic structures, thus guiding learning of discriminative features of different modalities.
1 code implementation • 24 May 2023 • Qihuang Zhong, Liang Ding, Juhua Liu, Xuebo Liu, Min Zhang, Bo Du, DaCheng Tao
Token dropping is a recently-proposed strategy to speed up the pretraining of masked language models, such as BERT, by skipping the computation of a subset of the input tokens at several middle layers.
1 code implementation • 9 Oct 2023 • Hongqiu Wu, Linfeng Liu, Hai Zhao, Min Zhang
Beyond the great cognitive powers showcased by language models, it is crucial to scrutinize whether their reasoning capabilities stem from strong generalization or merely exposure to relevant data.
1 code implementation • 1 Nov 2023 • Weihang Su, Qingyao Ai, Yueyue Wu, Yixiao Ma, Haitao Li, Yiqun Liu, Zhijing Wu, Min Zhang
Legal case retrieval aims to help legal workers find relevant cases related to their cases at hand, which is important for the guarantee of fairness and justice in legal judgments.
1 code implementation • 2 Jan 2024 • Xiongri Shen, Zhenxi Song, Linling Li, Min Zhang, Lingyan Liang Honghai Liu, Demao Deng, Zhiguo Zhang
Early diagnosis of mild cognitive impairment (MCI) and subjective cognitive decline (SCD) utilizing multi-modal magnetic resonance imaging (MRI) is a pivotal area of research.
1 code implementation • 19 Feb 2024 • Junru Lu, Siyu An, Min Zhang, Yulan He, Di Yin, Xing Sun
In the quest to facilitate the deep intelligence of Large Language Models (LLMs) accessible in final-end user-bot interactions, the art of prompt crafting emerges as a critical yet complex task for the average user.
no code implementations • 16 Jan 2018 • YaoSheng Yang, Meishan Zhang, Wenliang Chen, Wei zhang, Haofen Wang, Min Zhang
To quickly obtain new labeled data, we can choose crowdsourcing as an alternative way at lower cost in a short time.
Chinese Named Entity Recognition named-entity-recognition +2
no code implementations • 11 Jan 2018 • Kai Song, Yue Zhang, Min Zhang, Weihua Luo
Neural machine translation (NMT) suffers a performance deficiency when a limited vocabulary fails to cover the source or target side adequately, which happens frequently when dealing with morphologically rich languages.
no code implementations • 11 Jan 2018 • Zhengqiu He, Wenliang Chen, Zhenghua Li, Meishan Zhang, Wei zhang, Min Zhang
First, we encode the context of entities on a dependency tree as sentence-level entity embedding based on tree-GRU.
no code implementations • EMNLP 2017 • Xing Wang, Zhaopeng Tu, Deyi Xiong, Min Zhang
Otherwise, the NMT decoder generates a word from the vocabulary as the general NMT decoder does.
no code implementations • ACL 2017 • Junhui Li, Deyi Xiong, Zhaopeng Tu, Muhua Zhu, Min Zhang, Guodong Zhou
Even though a linguistics-free sequence to sequence model in neural machine translation (NMT) has certain capability of implicitly learning syntactic information of source sentences, this paper shows that source syntax can be explicitly incorporated into NMT effectively to provide further improvements.
no code implementations • 17 Oct 2016 • Xing Wang, Zhengdong Lu, Zhaopeng Tu, Hang Li, Deyi Xiong, Min Zhang
Neural Machine Translation (NMT) is a new approach to machine translation that has made great progress in recent years.
no code implementations • 4 Aug 2016 • Qingrong Xia, Zhenghua Li, Jiayuan Chao, Min Zhang
This paper describes our system designed for the NLPCC 2016 shared task on word segmentation on micro-blog texts.
no code implementations • 29 Sep 2016 • Zhenghua Li, Yue Zhang, Jiayuan Chao, Min Zhang
The first approach is previously proposed to directly train a log-linear graph-based parser (LLGPar) with PA based on a forest-based objective.
no code implementations • 11 Feb 2015 • Yongfeng Zhang, Min Zhang, Yiqun Liu, Shaoping Ma
In this paper, we focus on the problem of phrase-level sentiment polarity labelling and attempt to bridge the gap between phrase-level and review-level sentiment analysis.
no code implementations • 18 Nov 2013 • Min Zhang, Lei Yang, Zheng-Hai Huang
Additionally, combining an effective heuristic for determining $n$-rank, we can also apply the proposed algorithm to solve MnRA when $n$-rank is unknown in advance.
no code implementations • 30 Jun 2018 • Min Zhang, Qianli Ma, Chengfeng Wen, Hai Chen, Deruo Liu, Xianfeng GU, Jie He, Xiaoyin Xu
The Wasserstein distance between the nodules is calculated based on our new spherical optimal mass transport, this new algorithm works directly on sphere by using spherical metric, which is much more accurate and efficient than previous methods.
no code implementations • 16 Jul 2018 • Dun Liang, Yuanchen Guo, Shaokui Zhang, Song-Hai Zhang, Peter Hall, Min Zhang, Shi-Min Hu
Combining LineNet and TTLane, we proposed a pipeline to model HD maps with crowdsourced data for the first time.
no code implementations • EMNLP 2018 • Chenlin Shen, Changlong Sun, Jingjing Wang, Yangyang Kang, Shoushan Li, Xiaozhong Liu, Luo Si, Min Zhang, Guodong Zhou
On the basis, we propose a three-stage hierarchical matching network to explore deep sentiment information in a QA text pair.
no code implementations • ACL 2018 • Xinzhou Jiang, Zhenghua Li, Bo Zhang, Min Zhang, Sheng Li, Luo Si
Treebank conversion is a straightforward and effective way to exploit various heterogeneous treebanks for boosting parsing performance.
no code implementations • EMNLP 2017 • Chen Gong, Zhenghua Li, Min Zhang, Xinzhou Jiang
Traditionally, word segmentation (WS) adopts the single-grained formalism, where a sentence corresponds to a single word sequence.
no code implementations • WS 2018 • Nancy Chen, Xiangyu Duan, Min Zhang, Rafael E. Banchs, Haizhou Li
Transliteration is defined as phonetic translation of names across languages.
no code implementations • WS 2018 • Nancy Chen, Rafael E. Banchs, Min Zhang, Xiangyu Duan, Haizhou Li
This report presents the results from the Named Entity Transliteration Shared Task conducted as part of The Seventh Named Entities Workshop (NEWS 2018) held at ACL 2018 in Melbourne, Australia.
no code implementations • COLING 2018 • Lu Wang, Shoushan Li, Changlong Sun, Luo Si, Xiaozhong Liu, Min Zhang, Guodong Zhou
Question-Answer (QA) matching is a fundamental task in the Natural Language Processing community.
no code implementations • COLING 2016 • Fangyuan Li, Ruihong Huang, Deyi Xiong, Min Zhang
Aiming to resolve high complexities of event descriptions, previous work (Huang and Riloff, 2013) proposes multi-faceted event recognition and a bootstrapping method to automatically acquire both event facet phrases and event expressions from unannotated texts.
no code implementations • COLING 2016 • Wenliang Chen, Zhenjie Zhang, Zhenghua Li, Min Zhang
In this paper, we propose an approach to learn distributed representations of users and items from text comments for recommendation systems.