Search Results for author: Min Zhang

Found 398 papers, 177 papers with code

LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document Understanding

5 code implementations ACL 2021 Yang Xu, Yiheng Xu, Tengchao Lv, Lei Cui, Furu Wei, Guoxin Wang, Yijuan Lu, Dinei Florencio, Cha Zhang, Wanxiang Che, Min Zhang, Lidong Zhou

Pre-training of text and layout has proved effective in a variety of visually-rich document understanding tasks due to its effective model architecture and the advantage of large-scale unlabeled scanned/digital-born documents.

Document Image Classification Document Layout Analysis +6

Forging Multiple Training Objectives for Pre-trained Language Models via Meta-Learning

2 code implementations19 Oct 2022 Hongqiu Wu, Ruixue Ding, Hai Zhao, Boli Chen, Pengjun Xie, Fei Huang, Min Zhang

Multiple pre-training objectives fill the vacancy of the understanding capability of single-objective language modeling, which serves the ultimate purpose of pre-trained language models (PrLMs), generalizing well on a mass of scenarios.

Language Modelling Meta-Learning

Efficient Second-Order TreeCRF for Neural Dependency Parsing

2 code implementations ACL 2020 Yu Zhang, Zhenghua Li, Min Zhang

Experiments and analysis on 27 datasets from 13 languages clearly show that techniques developed before the DL era, such as structural learning (global TreeCRF loss) and high-order modeling are still useful, and can further boost parsing performance over the state-of-the-art biaffine parser, especially for partially annotated training data.

Chinese Dependency Parsing Dependency Parsing

Improving the Transformer Translation Model with Document-Level Context

3 code implementations EMNLP 2018 Jiacheng Zhang, Huanbo Luan, Maosong Sun, FeiFei Zhai, Jingfang Xu, Min Zhang, Yang Liu

Although the Transformer translation model (Vaswani et al., 2017) has achieved state-of-the-art performance in a variety of translation tasks, how to use document-level context to deal with discourse phenomena problematic for Transformer still remains a challenge.

Decoder Sentence +1

MuCGEC: a Multi-Reference Multi-Source Evaluation Dataset for Chinese Grammatical Error Correction

2 code implementations NAACL 2022 Yue Zhang, Zhenghua Li, Zuyi Bao, Jiacheng Li, Bo Zhang, Chen Li, Fei Huang, Min Zhang

This paper presents MuCGEC, a multi-reference multi-source evaluation dataset for Chinese Grammatical Error Correction (CGEC), consisting of 7, 063 sentences collected from three Chinese-as-a-Second-Language (CSL) learner sources.

Grammatical Error Correction Sentence

Chinese Sequence Labeling with Semi-Supervised Boundary-Aware Language Model Pre-training

2 code implementations8 Apr 2024 Longhui Zhang, Dingkun Long, Meishan Zhang, Yanzhao Zhang, Pengjun Xie, Min Zhang

Experimental results on Chinese sequence labeling datasets demonstrate that the improved BABERT variant outperforms the vanilla version, not only on these tasks but also more broadly across a range of Chinese natural language understanding tasks.

Language Modelling Natural Language Understanding

Efficient Document-level Event Extraction via Pseudo-Trigger-aware Pruned Complete Graph

1 code implementation11 Dec 2021 Tong Zhu, Xiaoye Qu, Wenliang Chen, Zhefeng Wang, Baoxing Huai, Nicholas Jing Yuan, Min Zhang

Most previous studies of document-level event extraction mainly focus on building argument chains in an autoregressive way, which achieves a certain success but is inefficient in both training and inference.

Document-level Event Extraction Event Extraction

Cobra: Extending Mamba to Multi-Modal Large Language Model for Efficient Inference

1 code implementation21 Mar 2024 Han Zhao, Min Zhang, Wei Zhao, Pengxiang Ding, Siteng Huang, Donglin Wang

In recent years, the application of multimodal large language models (MLLM) in various fields has achieved remarkable success.

Language Modelling Large Language Model

A Survey on Non-Autoregressive Generation for Neural Machine Translation and Beyond

1 code implementation20 Apr 2022 Yisheng Xiao, Lijun Wu, Junliang Guo, Juntao Li, Min Zhang, Tao Qin, Tie-Yan Liu

While NAR generation can significantly accelerate inference speed for machine translation, the speedup comes at the cost of sacrificed translation accuracy compared to its counterpart, autoregressive (AR) generation.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +11

Learning To Retrieve: How to Train a Dense Retrieval Model Effectively and Efficiently

2 code implementations20 Oct 2020 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Min Zhang, Shaoping Ma

Through this process, it teaches the DR model how to retrieve relevant documents from the entire corpus instead of how to rerank a potentially biased sample of documents.

Passage Retrieval Retrieval

Optimizing Dense Retrieval Model Training with Hard Negatives

4 code implementations16 Apr 2021 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma

ADORE replaces the widely-adopted static hard negative sampling method with a dynamic one to directly optimize the ranking performance.

Information Retrieval Representation Learning +1

Jointly Optimizing Query Encoder and Product Quantization to Improve Retrieval Performance

5 code implementations2 Aug 2021 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma

Compared with previous DR models that use brute-force search, JPQ almost matches the best retrieval performance with 30x compression on index size.

Information Retrieval Quantization +1

Learning Discrete Representations via Constrained Clustering for Effective and Efficient Dense Retrieval

4 code implementations12 Oct 2021 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma

However, the efficiency of most existing DR models is limited by the large memory cost of storing dense vectors and the time-consuming nearest neighbor search (NNS) in vector space.

Constrained Clustering Information Retrieval +3

A Survey of Large Language Models Attribution

1 code implementation7 Nov 2023 Dongfang Li, Zetian Sun, Xinshuo Hu, Zhenyu Liu, Ziyang Chen, Baotian Hu, Aiguo Wu, Min Zhang

Open-domain generative systems have gained significant attention in the field of conversational AI (e. g., generative search engines).

RepBERT: Contextualized Text Embeddings for First-Stage Retrieval

3 code implementations28 Jun 2020 Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Min Zhang, Shaoping Ma

Although exact term match between queries and documents is the dominant method to perform first-stage retrieval, we propose a different approach, called RepBERT, to represent documents and queries with fixed-length contextualized embeddings.

Passage Ranking Retrieval

OpenBA: An Open-sourced 15B Bilingual Asymmetric seq2seq Model Pre-trained from Scratch

1 code implementation19 Sep 2023 Juntao Li, Zecheng Tang, Yuyang Ding, Pinzheng Wang, Pei Guo, Wangjie You, Dan Qiao, Wenliang Chen, Guohong Fu, Qiaoming Zhu, Guodong Zhou, Min Zhang

This report provides the main details to pre-train an analogous model, including pre-training data processing, Bilingual Flan data collection, the empirical observations that inspire our model architecture design, training objectives of different stages, and other enhancement techniques.

Mirror: A Universal Framework for Various Information Extraction Tasks

1 code implementation9 Nov 2023 Tong Zhu, Junfei Ren, Zijian Yu, Mengsong Wu, Guoliang Zhang, Xiaoye Qu, Wenliang Chen, Zhefeng Wang, Baoxing Huai, Min Zhang

Sharing knowledge between information extraction tasks has always been a challenge due to the diverse data formats and task variations.

Machine Reading Comprehension

Modeling Graph Structure in Transformer for Better AMR-to-Text Generation

1 code implementation IJCNLP 2019 Jie Zhu, Junhui Li, Muhua Zhu, Longhua Qian, Min Zhang, Guodong Zhou

Recent studies on AMR-to-text generation often formalize the task as a sequence-to-sequence (seq2seq) learning problem by converting an Abstract Meaning Representation (AMR) graph into a word sequence.

AMR-to-Text Generation Text Generation

Towards Making the Most of ChatGPT for Machine Translation

1 code implementation24 Mar 2023 Keqin Peng, Liang Ding, Qihuang Zhong, Li Shen, Xuebo Liu, Min Zhang, Yuanxin Ouyang, DaCheng Tao

We show that: 1) The performance of ChatGPT depends largely on temperature, and a lower temperature usually can achieve better performance; 2) Emphasizing the task information can further improve ChatGPT's performance, particularly in complex MT tasks; 3) Introducing domain information can elicit ChatGPT's generalization ability and improve its performance in the specific domain; 4) ChatGPT tends to generate hallucinations for non-English-centric MT tasks, which can be partially addressed by our proposed prompts but still need to be highlighted for the MT/NLP community.

In-Context Learning Machine Translation +2

NaSGEC: a Multi-Domain Chinese Grammatical Error Correction Dataset from Native Speaker Texts

1 code implementation25 May 2023 Yue Zhang, Bo Zhang, Haochen Jiang, Zhenghua Li, Chen Li, Fei Huang, Min Zhang

We introduce NaSGEC, a new dataset to facilitate research on Chinese grammatical error correction (CGEC) for native speaker texts from multiple domains.

Grammatical Error Correction

Towards Representation Alignment and Uniformity in Collaborative Filtering

2 code implementations26 Jun 2022 Chenyang Wang, Yuanqing Yu, Weizhi Ma, Min Zhang, Chong Chen, Yiqun Liu, Shaoping Ma

Then, we empirically analyze the learning dynamics of typical CF methods in terms of quantified alignment and uniformity, which shows that better alignment or uniformity both contribute to higher recommendation performance.

Collaborative Filtering Recommendation Systems

Jointly Learning Explainable Rules for Recommendation with Knowledge Graph

1 code implementation9 Mar 2019 Weizhi Ma, Min Zhang, Yue Cao, Woojeong, Jin, Chenyang Wang, Yiqun Liu, Shaoping Ma, Xiang Ren

The framework encourages two modules to complement each other in generating effective and explainable recommendation: 1) inductive rules, mined from item-centric knowledge graphs, summarize common multi-hop relational patterns for inferring different item associations and provide human-readable explanation for model prediction; 2) recommendation module can be augmented by induced rules and thus have better generalization ability dealing with the cold-start issue.

Explainable Recommendation Knowledge Graphs +1

A new dataset of dog breed images and a benchmark for fine-grained classification

1 code implementation1 Oct 2020 Ding-Nan Zo, Song-Hai Zhang, Tai-Jiang M, Min Zhang

It is currently the largest dataset for fine-grained classification of dogs, including130 dog breeds and 70, 428 real-world images.

Benchmarking Classification +3

Variational Neural Machine Translation

1 code implementation EMNLP 2016 Biao Zhang, Deyi Xiong, Jinsong Su, Hong Duan, Min Zhang

Models of neural machine translation are often from a discriminative family of encoderdecoders that learn a conditional distribution of a target sentence given a source sentence.

Decoder Machine Translation +2

Disentangled Modeling of Domain and Relevance for Adaptable Dense Retrieval

1 code implementation11 Aug 2022 Jingtao Zhan, Qingyao Ai, Yiqun Liu, Jiaxin Mao, Xiaohui Xie, Min Zhang, Shaoping Ma

By making the REM and DAMs disentangled, DDR enables a flexible training paradigm in which REM is trained with supervision once and DAMs are trained with unsupervised data.

Ad-Hoc Information Retrieval Domain Adaptation +1

Neural Logic Reasoning

3 code implementations20 Aug 2020 Shaoyun Shi, Hanxiong Chen, Weizhi Ma, Jiaxin Mao, Min Zhang, Yongfeng Zhang

Both reasoning and generalization ability are important for prediction tasks such as recommender systems, where reasoning provides strong connection between user history and target items for accurate prediction, and generalization helps the model to draw a robust user portrait over noisy inputs.

Logical Reasoning Recommendation Systems

CCKS 2019 Shared Task on Inter-Personal Relationship Extraction

1 code implementation29 Aug 2019 Haitao Wang, Zhengqiu He, Tong Zhu, Hao Shao, Wenliang Chen, Min Zhang

In this paper, we present the task definition, the description of data and the evaluation methodology used during this shared task.

Sentence

LMEye: An Interactive Perception Network for Large Language Models

1 code implementation5 May 2023 Yunxin Li, Baotian Hu, Xinyu Chen, Lin Ma, Yong Xu, Min Zhang

LMEye addresses this issue by allowing the LLM to request the desired visual information aligned with various human instructions, which we term as the dynamic visual information interaction.

Language Modelling Large Language Model +1

LasUIE: Unifying Information Extraction with Latent Adaptive Structure-aware Generative Language Model

1 code implementation13 Apr 2023 Hao Fei, Shengqiong Wu, Jingye Li, Bobo Li, Fei Li, Libo Qin, Meishan Zhang, Min Zhang, Tat-Seng Chua

Universally modeling all typical information extraction tasks (UIE) with one generative language model (GLM) has revealed great potential by the latest study, where various IE predictions are unified into a linearized hierarchical expression under a GLM.

Language Modelling UIE

Improving AMR Parsing with Sequence-to-Sequence Pre-training

1 code implementation EMNLP 2020 Dongqin Xu, Junhui Li, Muhua Zhu, Min Zhang, Guodong Zhou

In the literature, the research on abstract meaning representation (AMR) parsing is much restricted by the size of human-curated dataset which is critical to build an AMR parser with good performance.

Ranked #15 on AMR Parsing on LDC2017T10 (using extra training data)

AMR Parsing Machine Translation +1

Contrastive Attention Mechanism for Abstractive Sentence Summarization

1 code implementation IJCNLP 2019 Xiangyu Duan, Hoongfei Yu, Mingming Yin, Min Zhang, Weihua Luo, Yue Zhang

We propose a contrastive attention mechanism to extend the sequence-to-sequence framework for abstractive sentence summarization task, which aims to generate a brief summary of a given source sentence.

Abstractive Text Summarization Sentence +1

RaSa: Relation and Sensitivity Aware Representation Learning for Text-based Person Search

1 code implementation23 May 2023 Yang Bai, Min Cao, Daming Gao, Ziqiang Cao, Chen Chen, Zhenfeng Fan, Liqiang Nie, Min Zhang

RA offsets the overfitting risk by introducing a novel positive relation detection task (i. e., learning to distinguish strong and weak positive pairs).

Person Search Relation +2

R-PointHop: A Green, Accurate, and Unsupervised Point Cloud Registration Method

1 code implementation15 Mar 2021 Pranav Kadam, Min Zhang, Shan Liu, C. -C. Jay Kuo

Inspired by the recent PointHop classification method, an unsupervised 3D point cloud registration method, called R-PointHop, is proposed in this work.

Attribute Dimensionality Reduction +2

Recommendation Unlearning

1 code implementation18 Jan 2022 Chong Chen, Fei Sun, Min Zhang, Bolin Ding

From the perspective of utility, if a system's utility is damaged by some bad data, the system needs to forget these data to regain utility.

Machine Unlearning Recommendation Systems

Joint Multi-modal Aspect-Sentiment Analysis with Auxiliary Cross-modal Relation Detection

1 code implementation EMNLP 2021 Xincheng Ju, Dong Zhang, Rong Xiao, Junhui Li, Shoushan Li, Min Zhang, Guodong Zhou

Therefore, in this paper, we are the first to jointly perform multi-modal ATE (MATE) and multi-modal ASC (MASC), and we propose a multi-modal joint learning approach with auxiliary cross-modal relation detection for multi-modal aspect-level sentiment analysis (MALSA).

Relation Sentiment Analysis +1

An Empirical Study of CLIP for Text-based Person Search

1 code implementation19 Aug 2023 Min Cao, Yang Bai, Ziyin Zeng, Mang Ye, Min Zhang

TPBS, as a fine-grained cross-modal retrieval task, is also facing the rise of research on the CLIP-based TBPS.

Cross-Modal Retrieval Data Augmentation +5

Can Diffusion Model Achieve Better Performance in Text Generation? Bridging the Gap between Training and Inference!

1 code implementation8 May 2023 Zecheng Tang, Pinzheng Wang, Keyan Zhou, Juntao Li, Ziqiang Cao, Min Zhang

Diffusion models have been successfully adapted to text generation tasks by mapping the discrete text into the continuous space.

Text Generation

Towards Accurate and Consistent Evaluation: A Dataset for Distantly-Supervised Relation Extraction

1 code implementation COLING 2020 Tong Zhu, Haitao Wang, Junjie Yu, Xiabing Zhou, Wenliang Chen, Wei zhang, Min Zhang

The experimental results show that the ranking lists of the comparison systems on the DS-labelled test data and human-annotated test data are different.

Relation Relation Extraction

Improving Seq2Seq Grammatical Error Correction via Decoding Interventions

1 code implementation23 Oct 2023 Houquan Zhou, Yumeng Liu, Zhenghua Li, Min Zhang, Bo Zhang, Chen Li, Ji Zhang, Fei Huang

In this paper, we propose a unified decoding intervention framework that employs an external critic to assess the appropriateness of the token to be generated incrementally, and then dynamically influence the choice of the next token.

Decoder Grammatical Error Correction +1

Zero-Shot Cross-Lingual Abstractive Sentence Summarization through Teaching Generation and Attention

1 code implementation ACL 2019 Xiangyu Duan, Mingming Yin, Min Zhang, Boxing Chen, Weihua Luo

But there is no cross-lingual parallel corpus, whose source sentence language is different to the summary language, to directly train a cross-lingual ASSUM system.

Sentence Sentence Summarization +1

SelfMix: Robust Learning Against Textual Label Noise with Self-Mixup Training

1 code implementation COLING 2022 Dan Qiao, Chenchen Dai, Yuyang Ding, Juntao Li, Qiang Chen, Wenliang Chen, Min Zhang

The conventional success of textual classification relies on annotated data, and the new paradigm of pre-trained language models (PLMs) still requires a few labeled data for downstream tasks.

text-classification Text Classification

Jointly Non-Sampling Learning for Knowledge Graph Enhanced Recommendation

2 code implementations1 Jul 2020 Chong Chen, Min Zhang, Weizhi Ma, Yiqun Liu, and Shaoping Ma

However, existing KG enhanced recommendation methods have largely focused on exploring advanced neural network architectures to better investigate the structural information of KG.

Knowledge Graph Embedding Knowledge Graphs +2

TSRankLLM: A Two-Stage Adaptation of LLMs for Text Ranking

1 code implementation28 Nov 2023 Longhui Zhang, Yanzhao Zhang, Dingkun Long, Pengjun Xie, Meishan Zhang, Min Zhang

Text ranking is a critical task in various information retrieval applications, and the recent success of pre-trained language models (PLMs), especially large language models (LLMs), has sparked interest in their application to text ranking.

Decoder Information Retrieval +1

Rethinking Negative Instances for Generative Named Entity Recognition

1 code implementation26 Feb 2024 Yuyang Ding, Juntao Li, Pinzheng Wang, Zecheng Tang, Bowen Yan, Min Zhang

In the Named Entity Recognition (NER) task, recent advancements have seen the remarkable improvement of LLMs in a broad range of entity domains via instruction tuning, by adopting entity-centric schema.

named-entity-recognition Named Entity Recognition +2

Code Summarization with Structure-induced Transformer

1 code implementation Findings (ACL) 2021 Hongqiu Wu, Hai Zhao, Min Zhang

Code summarization (CS) is becoming a promising area in recent language understanding, which aims to generate sensible human language automatically for programming language in the format of source code, serving in the most convenience of programmer developing.

Code Summarization Natural Language Understanding +1

Stochastic Variance Reduced Ensemble Adversarial Attack for Boosting the Adversarial Transferability

1 code implementation CVPR 2022 Yifeng Xiong, Jiadong Lin, Min Zhang, John E. Hopcroft, Kun He

The black-box adversarial attack has attracted impressive attention for its practical use in the field of deep learning security.

Adversarial Attack

Visual Spatial Description: Controlled Spatial-Oriented Image-to-Text Generation

1 code implementation20 Oct 2022 Yu Zhao, Jianguo Wei, Zhichao Lin, Yueheng Sun, Meishan Zhang, Min Zhang

Accordingly, we manually annotate a dataset to facilitate the investigation of the newly-introduced task and build several benchmark encoder-decoder models by using VL-BART and VL-T5 as backbones.

Decoder Image Captioning +1

Generating Visual Spatial Description via Holistic 3D Scene Understanding

1 code implementation19 May 2023 Yu Zhao, Hao Fei, Wei Ji, Jianguo Wei, Meishan Zhang, Min Zhang, Tat-Seng Chua

With an external 3D scene extractor, we obtain the 3D objects and scene features for input images, based on which we construct a target object-centered 3D spatial scene graph (Go3D-S2G), such that we model the spatial semantics of target objects within the holistic 3D scenes.

Scene Understanding Text Generation

PointHop: An Explainable Machine Learning Method for Point Cloud Classification

3 code implementations30 Jul 2019 Min Zhang, Haoxuan You, Pranav Kadam, Shan Liu, C. -C. Jay Kuo

In the attribute building stage, we address the problem of unordered point cloud data using a space partitioning procedure and developing a robust descriptor that characterizes the relationship between a point and its one-hop neighbor in a PointHop unit.

Attribute BIG-bench Machine Learning +3

Revisiting Grammatical Error Correction Evaluation and Beyond

1 code implementation3 Nov 2022 Peiyuan Gong, Xuebo Liu, Heyan Huang, Min Zhang

Pretraining-based (PT-based) automatic evaluation metrics (e. g., BERTScore and BARTScore) have been widely used in several sentence generation tasks (e. g., machine translation and text summarization) due to their better correlation with human judgments over traditional overlap-based methods.

Grammatical Error Correction Machine Translation +2

Generative Multimodal Entity Linking

1 code implementation22 Jun 2023 Senbao Shi, Zhenran Xu, Baotian Hu, Min Zhang

Multimodal Entity Linking (MEL) is the task of mapping mentions with multimodal contexts to the referent entities from a knowledge base.

Entity Linking In-Context Learning +1

Intent-aware Ranking Ensemble for Personalized Recommendation

2 code implementations15 Apr 2023 Jiayu Li, Peijie Sun, Zhefan Wang, Weizhi Ma, Yangkun Li, Min Zhang, Zhoutian Feng, Daiyue Xue

To address such a task, we propose an Intent-aware ranking Ensemble Learning~(IntEL) model to fuse multiple single-objective item lists with various user intents, in which item-level personalized weights are learned.

Ensemble Learning Recommendation Systems

Variational Neural Discourse Relation Recognizer

1 code implementation EMNLP 2016 Biao Zhang, Deyi Xiong, Jinsong Su, Qun Liu, Rongrong Ji, Hong Duan, Min Zhang

In order to perform efficient inference and learning, we introduce neural discourse relation models to approximate the prior and posterior distributions of the latent variable, and employ these approximated distributions to optimize a reparameterized variational lower bound.

Relation

Syntax-aware Neural Semantic Role Labeling

1 code implementation22 Jul 2019 Qingrong Xia, Zhenghua Li, Min Zhang, Meishan Zhang, Guohong Fu, Rui Wang, Luo Si

Semantic role labeling (SRL), also known as shallow semantic parsing, is an important yet challenging task in NLP.

Semantic Parsing Semantic Role Labeling +1

Fast and Accurate End-to-End Span-based Semantic Role Labeling as Word-based Graph Parsing

1 code implementation COLING 2022 Shilin Zhou, Qingrong Xia, Zhenghua Li, Yu Zhang, Yu Hong, Min Zhang

Moreover, we propose a simple constrained Viterbi procedure to ensure the legality of the output graph according to the constraints of the SRL structure.

Chinese Word Segmentation named-entity-recognition +3

ConsistTL: Modeling Consistency in Transfer Learning for Low-Resource Neural Machine Translation

1 code implementation8 Dec 2022 Zhaocong Li, Xuebo Liu, Derek F. Wong, Lidia S. Chao, Min Zhang

In this paper, we propose a novel transfer learning method for NMT, namely ConsistTL, which can continuously transfer knowledge from the parent model during the training of the child model.

Low-Resource Neural Machine Translation NMT +2

Prompt-based Distribution Alignment for Unsupervised Domain Adaptation

1 code implementation15 Dec 2023 Shuanghao Bai, Min Zhang, Wanqi Zhou, Siteng Huang, Zhirong Luan, Donglin Wang, Badong Chen

Therefore, in this paper, we first experimentally demonstrate that the unsupervised-trained VLMs can significantly reduce the distribution discrepancy between source and target domains, thereby improving the performance of UDA.

Prompt Engineering Unsupervised Domain Adaptation

Expressive Forecasting of 3D Whole-body Human Motions

1 code implementation19 Dec 2023 Pengxiang Ding, Qiongjie Cui, Min Zhang, Mengyuan Liu, Haofan Wang, Donglin Wang

Human motion forecasting, with the goal of estimating future human behavior over a period of time, is a fundamental task in many real-world applications.

Human Pose Forecasting Motion Forecasting

A User-Centric Benchmark for Evaluating Large Language Models

1 code implementation22 Apr 2024 Jiayin Wang, Fengran Mo, Weizhi Ma, Peijie Sun, Min Zhang, Jian-Yun Nie

To address this oversight, we propose benchmarking LLMs from a user perspective in both dataset construction and evaluation designs.

Benchmarking

PointHop++: A Lightweight Learning Model on Point Sets for 3D Classification

2 code implementations9 Feb 2020 Min Zhang, Yifan Wang, Pranav Kadam, Shan Liu, C. -C. Jay Kuo

The PointHop method was recently proposed by Zhang et al. for 3D point cloud classification with unsupervised feature extraction.

3D Classification 3D Point Cloud Classification +2

Bilingual Dictionary Based Neural Machine Translation without Using Parallel Sentences

1 code implementation ACL 2020 Xiangyu Duan, Baijun Ji, Hao Jia, Min Tan, Min Zhang, Boxing Chen, Weihua Luo, Yue Zhang

In this paper, we propose a new task of machine translation (MT), which is based on no parallel sentences but can refer to a ground-truth bilingual dictionary.

Machine Translation Translation +1

A Unified Span-Based Approach for Opinion Mining with Syntactic Constituents

1 code implementation NAACL 2021 Qingrong Xia, Bo Zhang, Rui Wang, Zhenghua Li, Yue Zhang, Fei Huang, Luo Si, Min Zhang

Fine-grained opinion mining (OM) has achieved increasing attraction in the natural language processing (NLP) community, which aims to find the opinion structures of {``}Who expressed what opinions towards what{''} in one sentence.

Multi-Task Learning Opinion Mining +1

Semi-supervised Domain Adaptation for Dependency Parsing

1 code implementation ACL 2019 Zhenghua Li, Xue Peng, Min Zhang, Rui Wang, Luo Si

During the past decades, due to the lack of sufficient labeled data, most studies on cross-domain parsing focus on unsupervised domain adaptation, assuming there is no target-domain training data.

Chinese Dependency Parsing Dependency Parsing +3

Multi-Agent Collaboration Framework for Recommender Systems

1 code implementation23 Feb 2024 Zhefan Wang, Yuanqing Yu, Wendi Zheng, Weizhi Ma, Min Zhang

LLM-based agents have gained considerable attention for their decision-making skills and ability to handle complex tasks.

Decision Making Explanation Generation +1

Combining Static Word Embeddings and Contextual Representations for Bilingual Lexicon Induction

1 code implementation Findings (ACL) 2021 Jinpeng Zhang, Baijun Ji, Nini Xiao, Xiangyu Duan, Min Zhang, Yangbin Shi, Weihua Luo

Bilingual Lexicon Induction (BLI) aims to map words in one language to their translations in another, and is typically through learning linear projections to align monolingual word representation spaces.

Bilingual Lexicon Induction Word Embeddings

Towards Unifying the Label Space for Aspect- and Sentence-based Sentiment Analysis

1 code implementation Findings (ACL) 2022 Yiming Zhang, Min Zhang, Sai Wu, Junbo Zhao

The aspect-based sentiment analysis (ABSA) is a fine-grained task that aims to determine the sentiment polarity towards targeted aspect terms occurring in the sentence.

Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +3

A Multi-Modal Context Reasoning Approach for Conditional Inference on Joint Textual and Visual Clues

1 code implementation8 May 2023 Yunxin Li, Baotian Hu, Xinyu Chen, Yuxin Ding, Lin Ma, Min Zhang

This makes the language model well-suitable for such multi-modal reasoning scenario on joint textual and visual clues.

Language Modelling

Language Models are Universal Embedders

1 code implementation12 Oct 2023 Xin Zhang, Zehan Li, Yanzhao Zhang, Dingkun Long, Pengjun Xie, Meishan Zhang, Min Zhang

As such cases span from English to other natural or programming languages, from retrieval to classification and beyond, it is desirable to build a unified embedding model rather than dedicated ones for each scenario.

Code Search Language Modelling +2

An In-depth Study on Internal Structure of Chinese Words

1 code implementation ACL 2021 Chen Gong, Saihao Huang, Houquan Zhou, Zhenghua Li, Min Zhang, Zhefeng Wang, Baoxing Huai, Nicholas Jing Yuan

Several previous works on syntactic parsing propose to annotate shallow word-internal structures for better utilizing character-level information.

Sentence

RST Discourse Parsing with Second-Stage EDU-Level Pre-training

1 code implementation ACL 2022 Nan Yu, Meishan Zhang, Guohong Fu, Min Zhang

Pre-trained language models (PLMs) have shown great potentials in natural language processing (NLP) including rhetorical structure theory (RST) discourse parsing. Current PLMs are obtained by sentence-level pre-training, which is different from the basic processing unit, i. e. element discourse unit (EDU). To this end, we propose a second-stage EDU-level pre-training approach in this work, which presents two novel tasks to learn effective EDU representations continually based on well pre-trained language models. Concretely, the two tasks are (1) next EDU prediction (NEP) and (2) discourse marker prediction (DMP). We take a state-of-the-art transition-based neural parser as baseline, and adopt it with a light bi-gram EDU modification to effectively explore the EDU-level pre-trained EDU representation. Experimental results on a benckmark dataset show that our method is highly effective, leading a 2. 1-point improvement in F1-score. All codes and pre-trained models will be released publicly to facilitate future studies.

Discourse Marker Prediction Discourse Parsing +1

Brain Topography Adaptive Network for Satisfaction Modeling in Interactive Information Access System

1 code implementation17 Aug 2022 Ziyi Ye, Xiaohui Xie, Yiqun Liu, Zhihong Wang, Xuesong Chen, Min Zhang, Shaoping Ma

We explore the effectiveness of BTA for satisfaction modeling in two popular information access scenarios, i. e., search and recommendation.

EEG Recommendation Systems +1

Large Language Model Cascades with Mixture of Thoughts Representations for Cost-efficient Reasoning

1 code implementation4 Oct 2023 Murong Yue, Jie Zhao, Min Zhang, Liang Du, Ziyu Yao

Large language models (LLMs) such as GPT-4 have exhibited remarkable performance in a variety of tasks, but this strong performance often comes with the high expense of using paid API services.

Decision Making Language Modelling +1

Improving Simultaneous Machine Translation with Monolingual Data

1 code implementation2 Dec 2022 Hexuan Deng, Liang Ding, Xuebo Liu, Meishan Zhang, DaCheng Tao, Min Zhang

Preliminary experiments on En-Zh and En-Ja news domain corpora demonstrate that monolingual data can significantly improve translation quality (e. g., +3. 15 BLEU on En-Zh).

Hallucination Knowledge Distillation +4

Fuzzy Alignments in Directed Acyclic Graph for Non-Autoregressive Machine Translation

1 code implementation12 Mar 2023 Zhengrui Ma, Chenze Shao, Shangtong Gui, Min Zhang, Yang Feng

Non-autoregressive translation (NAT) reduces the decoding latency but suffers from performance degradation due to the multi-modality problem.

Machine Translation Sentence +1

Interpretable Convolutional Neural Networks via Feedforward Design

2 code implementations5 Oct 2018 C. -C. Jay Kuo, Min Zhang, Siyang Li, Jiali Duan, Yueru Chen

To construct convolutional layers, we develop a new signal transform, called the Saab (Subspace Approximation with Adjusted Bias) transform.

Self-Correlation and Cross-Correlation Learning for Few-Shot Remote Sensing Image Semantic Segmentation

1 code implementation11 Sep 2023 Linhan Wang, Shuo Lei, Jianfeng He, Shengkun Wang, Min Zhang, Chang-Tien Lu

To tackle these challenges, we propose a Self-Correlation and Cross-Correlation Learning Network for the few-shot remote sensing image semantic segmentation.

Few-Shot Learning Segmentation +1

SiLLM: Large Language Models for Simultaneous Machine Translation

1 code implementation20 Feb 2024 Shoutao Guo, Shaolei Zhang, Zhengrui Ma, Min Zhang, Yang Feng

We propose SiLLM, which delegates the two sub-tasks to separate agents, thereby incorporating LLM into SiMT.

Machine Translation Sentence +1

Attributes-Guided and Pure-Visual Attention Alignment for Few-Shot Recognition

1 code implementation10 Sep 2020 Siteng Huang, Min Zhang, Yachen Kang, Donglin Wang

However, these approaches only augment the representations of samples with available semantics while ignoring the query set, which loses the potential for the improvement and may lead to a shift between the modalities combination and the pure-visual representation.

feature selection Metric Learning

CT4Rec: Simple yet Effective Consistency Training for Sequential Recommendation

2 code implementations13 Dec 2021 Chong Liu, Xiaoyang Liu, Rongqin Zheng, Lixin Zhang, Xiaobo Liang, Juntao Li, Lijun Wu, Min Zhang, Leyu Lin

State-of-the-art sequential recommendation models proposed very recently combine contrastive learning techniques for obtaining high-quality user representations.

Click-Through Rate Prediction Contrastive Learning +2

Tree Structure-Aware Few-Shot Image Classification via Hierarchical Aggregation

1 code implementation14 Jul 2022 Min Zhang, Siteng Huang, Wenbin Li, Donglin Wang

To solve this problem, we present a plug-in Hierarchical Tree Structure-aware (HTS) method, which not only learns the relationship of FSL and pretext tasks, but more importantly, can adaptively select and aggregate feature representations generated by pretext tasks to maximize the performance of FSL tasks.

Few-Shot Image Classification Few-Shot Learning

Enhancing Multi-modal and Multi-hop Question Answering via Structured Knowledge and Unified Retrieval-Generation

1 code implementation16 Dec 2022 Qian Yang, Qian Chen, Wen Wang, Baotian Hu, Min Zhang

Moreover, the pipelined approaches of retrieval and generation might result in poor generation performance when retrieval performance is low.

Answer Generation Decoder +4

Troika: Multi-Path Cross-Modal Traction for Compositional Zero-Shot Learning

1 code implementation27 Mar 2023 Siteng Huang, Biao Gong, Yutong Feng, Min Zhang, Yiliang Lv, Donglin Wang

Recent compositional zero-shot learning (CZSL) methods adapt pre-trained vision-language models (VLMs) by constructing trainable prompts only for composed state-object pairs.

Compositional Zero-Shot Learning Object

CMD: a framework for Context-aware Model self-Detoxification

2 code implementations16 Aug 2023 Zecheng Tang, Keyan Zhou, Juntao Li, Yuyang Ding, Pinzheng Wang, Bowen Yan, Min Zhang

In view of this, we introduce a Context-aware Model self-Detoxification~(CMD) framework that pays attention to both the context and the detoxification process, i. e., first detoxifying the context and then making the language model generate along the safe context.

Language Modelling

Learning Semantic-Aligned Feature Representation for Text-based Person Search

1 code implementation13 Dec 2021 Shiping Li, Min Cao, Min Zhang

In this paper, we propose a semantic-aligned embedding method for text-based person search, in which the feature alignment across modalities is achieved by automatically learning the semantic-aligned visual features and textual features.

Person Search Text based Person Retrieval +1

Evaluating Interpolation and Extrapolation Performance of Neural Retrieval Models

1 code implementation25 Apr 2022 Jingtao Zhan, Xiaohui Xie, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, Shaoping Ma

For example, representation-based retrieval models perform almost as well as interaction-based retrieval models in terms of interpolation but not extrapolation.

Retrieval

Bridging the Domain Gaps in Context Representations for k-Nearest Neighbor Neural Machine Translation

1 code implementation26 May 2023 Zhiwei Cao, Baosong Yang, Huan Lin, Suhang Wu, Xiangpeng Wei, Dayiheng Liu, Jun Xie, Min Zhang, Jinsong Su

$k$-Nearest neighbor machine translation ($k$NN-MT) has attracted increasing attention due to its ability to non-parametrically adapt to new translation domains.

Domain Adaptation Machine Translation +3

Non-autoregressive Streaming Transformer for Simultaneous Translation

1 code implementation23 Oct 2023 Zhengrui Ma, Shaolei Zhang, Shoutao Guo, Chenze Shao, Min Zhang, Yang Feng

Simultaneous machine translation (SiMT) models are trained to strike a balance between latency and translation quality.

Decoder Machine Translation +1

Towards Reasoning in Large Language Models via Multi-Agent Peer Review Collaboration

1 code implementation14 Nov 2023 Zhenran Xu, Senbao Shi, Baotian Hu, Jindi Yu, Dongfang Li, Min Zhang, Yuxiang Wu

Large Language Models (LLMs) have shown remarkable capabilities in general natural language processing tasks but often fall short in complex reasoning tasks.

Math

SelectIT: Selective Instruction Tuning for Large Language Models via Uncertainty-Aware Self-Reflection

1 code implementation26 Feb 2024 Liangxin Liu, Xuebo Liu, Derek F. Wong, Dongfang Li, Ziyi Wang, Baotian Hu, Min Zhang

In this work, we propose a novel approach, termed SelectIT, that capitalizes on the foundational capabilities of the LLM itself.

Instruction-Driven Game Engines on Large Language Models

1 code implementation30 Mar 2024 Hongqiu Wu, Y. Wang, XingYuan Liu, Hai Zhao, Min Zhang

The Instruction-Driven Game Engine (IDGE) project aims to democratize game development by enabling a large language model (LLM) to follow free-form game rules and autonomously generate game-play processes.

Language Modelling Large Language Model

Toward Adversarial Training on Contextualized Language Representation

1 code implementation8 May 2023 Hongqiu Wu, Yongxiang Liu, Hanwen Shi, Hai Zhao, Min Zhang

Based on the observation, we propose simple yet effective \textit{Contextualized representation-Adversarial Training} (CreAT), in which the attack is explicitly optimized to deviate the contextualized representation of the encoder.

Decoder named-entity-recognition +1

Sequential Recommendation with Latent Relations based on Large Language Model

1 code implementation27 Mar 2024 Shenghao Yang, Weizhi Ma, Peijie Sun, Qingyao Ai, Yiqun Liu, Mingchen Cai, Min Zhang

Different from previous relation-aware models that rely on predefined rules, we propose to leverage the Large Language Model (LLM) to provide new types of relations and connections between items.

Collaborative Filtering Knowledge Graphs +4

A Label Dependence-aware Sequence Generation Model for Multi-level Implicit Discourse Relation Recognition

1 code implementation22 Dec 2021 Changxing Wu, Liuwen Cao, Yubin Ge, Yang Liu, Min Zhang, Jinsong Su

Then, we employ a label sequence decoder to output the predicted labels in a top-down manner, where the predicted higher-level labels are directly used to guide the label prediction at the current level.

Decoder Relation

WR-ONE2SET: Towards Well-Calibrated Keyphrase Generation

1 code implementation13 Nov 2022 Binbin Xie, Xiangpeng Wei, Baosong Yang, Huan Lin, Jun Xie, Xiaoli Wang, Min Zhang, Jinsong Su

Keyphrase generation aims to automatically generate short phrases summarizing an input document.

Keyphrase Generation

Boosting Verified Training for Robust Image Classifications via Abstraction

1 code implementation CVPR 2023 Zhaodi Zhang, Zhiyi Xue, Yang Chen, Si Liu, Yueling Zhang, Jing Liu, Min Zhang

Via abstraction, all perturbed images are mapped into intervals before feeding into neural networks for training.

A Neural Divide-and-Conquer Reasoning Framework for Image Retrieval from Linguistically Complex Text

1 code implementation3 May 2023 Yunxin Li, Baotian Hu, Yuxin Ding, Lin Ma, Min Zhang

Inspired by the Divide-and-Conquer algorithm and dual-process theory, in this paper, we regard linguistically complex texts as compound proposition texts composed of multiple simple proposition sentences and propose an end-to-end Neural Divide-and-Conquer Reasoning framework, dubbed NDCR.

Image Retrieval Logical Reasoning +1

Scene Graph as Pivoting: Inference-time Image-free Unsupervised Multimodal Machine Translation with Visual Scene Hallucination

1 code implementation20 May 2023 Hao Fei, Qian Liu, Meishan Zhang, Min Zhang, Tat-Seng Chua

In this work, we investigate a more realistic unsupervised multimodal machine translation (UMMT) setup, inference-time image-free UMMT, where the model is trained with source-text image pairs, and tested with only source-text inputs.

Hallucination Multimodal Machine Translation +1

Language Generation from Brain Recordings

1 code implementation16 Nov 2023 Ziyi Ye, Qingyao Ai, Yiqun Liu, Maarten de Rijke, Min Zhang, Christina Lioma, Tuukka Ruotsalo

Inspired by recent research that revealed associations between the brain and the large computational language models, we propose a generative language BCI that utilizes the capacity of a large language model (LLM) jointly with a semantic brain decoder to directly generate language from functional magnetic resonance imaging (fMRI) input.

Decoder Language Modelling +3

Improving Relation Extraction with Relational Paraphrase Sentences

1 code implementation COLING 2020 Junjie Yu, Tong Zhu, Wenliang Chen, Wei zhang, Min Zhang

In this paper, we propose an alternative approach to improve RE systems via enriching diverse expressions by relational paraphrase sentences.

Relation Relation Extraction

Third-Party Aligner for Neural Word Alignments

1 code implementation8 Nov 2022 Jinpeng Zhang, Chuanqi Dong, Xiangyu Duan, Yuqi Zhang, Min Zhang

Word alignment is to find translationally equivalent words between source and target sentences.

Language Modelling Word Alignment

How Well Do Large Language Models Understand Syntax? An Evaluation by Asking Natural Language Questions

1 code implementation14 Nov 2023 Houquan Zhou, Yang Hou, Zhenghua Li, Xuebin Wang, Zhefeng Wang, Xinyu Duan, Min Zhang

While recent advancements in large language models (LLMs) bring us closer to achieving artificial general intelligence, the question persists: Do LLMs truly understand language, or do they merely mimic comprehension through pattern recognition?

Prepositional Phrase Attachment Question Answering +1

GenView: Enhancing View Quality with Pretrained Generative Model for Self-Supervised Learning

1 code implementation18 Mar 2024 Xiaojie Li, Yibo Yang, Xiangtai Li, Jianlong Wu, Yue Yu, Bernard Ghanem, Min Zhang

To tackle these challenges, we present GenView, a controllable framework that augments the diversity of positive views leveraging the power of pretrained generative models while preserving semantics.

Contrastive Learning Data Augmentation +1

In-Context Learning State Vector with Inner and Momentum Optimization

1 code implementation17 Apr 2024 Dongfang Li, Zhenyu Liu, Xinshuo Hu, Zetian Sun, Baotian Hu, Min Zhang

In this paper, we address this gap by presenting a comprehensive analysis of these compressed vectors, drawing parallels to the parameters trained with gradient descent, and introduce the concept of state vector.

In-Context Learning Test-time Adaptation

Semantic Role Labeling with Heterogeneous Syntactic Knowledge

1 code implementation COLING 2020 Qingrong Xia, Rui Wang, Zhenghua Li, Yue Zhang, Min Zhang

Recently, due to the interplay between syntax and semantics, incorporating syntactic knowledge into neural semantic role labeling (SRL) has achieved much attention.

Semantic Role Labeling

Adversarial Self-Attention for Language Understanding

1 code implementation25 Jun 2022 Hongqiu Wu, Ruixue Ding, Hai Zhao, Pengjun Xie, Fei Huang, Min Zhang

Deep neural models (e. g. Transformer) naturally learn spurious features, which create a ``shortcut'' between the labels and inputs, thus impairing the generalization and robustness.

Machine Reading Comprehension Named Entity Recognition (NER) +4

Towards Robust Neural Machine Translation with Iterative Scheduled Data-Switch Training

1 code implementation COLING 2022 Zhongjian Miao, Xiang Li, Liyan Kang, Wen Zhang, Chulun Zhou, Yidong Chen, Bin Wang, Min Zhang, Jinsong Su

Most existing methods on robust neural machine translation (NMT) construct adversarial examples by injecting noise into authentic examples and indiscriminately exploit two types of examples.

Machine Translation NMT +2

AMOM: Adaptive Masking over Masking for Conditional Masked Language Model

1 code implementation13 Mar 2023 Yisheng Xiao, Ruiyang Xu, Lijun Wu, Juntao Li, Tao Qin, Yan-Tie Liu, Min Zhang

Experiments on \textbf{3} different tasks (neural machine translation, summarization, and code generation) with \textbf{15} datasets in total confirm that our proposed simple method achieves significant performance improvement over the strong CMLM model.

Code Generation Decoder +3

ExplainCPE: A Free-text Explanation Benchmark of Chinese Pharmacist Examination

1 code implementation22 May 2023 Dongfang Li, Jindi Yu, Baotian Hu, Zhenran Xu, Min Zhang

As ChatGPT and GPT-4 spearhead the development of Large Language Models (LLMs), more researchers are investigating their performance across various tasks.

General Knowledge In-Context Learning

A Read-and-Select Framework for Zero-shot Entity Linking

1 code implementation19 Oct 2023 Zhenran Xu, Yulin Chen, Baotian Hu, Min Zhang

Zero-shot entity linking (EL) aims at aligning entity mentions to unseen entities to challenge the generalization ability.

Entity Disambiguation Entity Linking +1

Revisiting Sparse Retrieval for Few-shot Entity Linking

1 code implementation19 Oct 2023 Yulin Chen, Zhenran Xu, Baotian Hu, Min Zhang

Entity linking aims to link ambiguous mentions to their corresponding entities in a knowledge base.

Entity Linking Retrieval

A Multimodal In-Context Tuning Approach for E-Commerce Product Description Generation

1 code implementation21 Feb 2024 Yunxin Li, Baotian Hu, Wenhan Luo, Lin Ma, Yuxin Ding, Min Zhang

For this setting, previous methods utilize visual and textual encoders to encode the image and keywords and employ a language model-based decoder to generate the product description.

In-Context Learning Language Modelling +2

VisionGraph: Leveraging Large Multimodal Models for Graph Theory Problems in Visual Context

1 code implementation8 May 2024 Yunxin Li, Baotian Hu, Haoyuan Shi, Wei Wang, Longyue Wang, Min Zhang

Large Multimodal Models (LMMs) have achieved impressive success in visual understanding and reasoning, remarkably improving the performance of mathematical reasoning in a visual context.

Math Mathematical Reasoning

Relevance Feedback with Brain Signals

1 code implementation9 Dec 2023 Ziyi Ye, Xiaohui Xie, Qingyao Ai, Yiqun Liu, Zhihong Wang, Weihang Su, Min Zhang

To explore the effectiveness of brain signals in the context of RF, we propose a novel RF framework that combines BCI-based relevance feedback with pseudo-relevance signals and implicit signals to improve the performance of document re-ranking.

Brain Computer Interface Re-Ranking

OpenBA-V2: Reaching 77.3% High Compression Ratio with Fast Multi-Stage Pruning

1 code implementation9 May 2024 Dan Qiao, Yi Su, Pinzheng Wang, Jing Ye, Wenjing Xie, Yuechi Zhou, Yuyang Ding, Zecheng Tang, Jikai Wang, Yixin Ji, Yue Wang, Pei Guo, Zechen Sun, Zikang Zhang, Juntao Li, Pingfu Chao, Wenliang Chen, Guohong Fu, Guodong Zhou, Qiaoming Zhu, Min Zhang

Large Language Models (LLMs) have played an important role in many fields due to their powerful capabilities. However, their massive number of parameters leads to high deployment requirements and incurs significant inference costs, which impedes their practical applications.

Loss re-scaling VQA: Revisiting the LanguagePrior Problem from a Class-imbalance View

1 code implementation30 Oct 2020 Yangyang Guo, Liqiang Nie, Zhiyong Cheng, Qi Tian, Min Zhang

Concretely, we design a novel interpretation scheme whereby the loss of mis-predicted frequent and sparse answers of the same question type is distinctly exhibited during the late training phase.

Face Recognition Image Classification +2

A Coarse-to-Fine Labeling Framework for Joint Word Segmentation, POS Tagging, and Constituent Parsing

1 code implementation CoNLL (EMNLP) 2021 Yang Hou, Houquan Zhou, Zhenghua Li, Yu Zhang, Min Zhang, Zhefeng Wang, Baoxing Huai, Nicholas Jing Yuan

In the coarse labeling stage, the joint model outputs a bracketed tree, in which each node corresponds to one of four labels (i. e., phrase, subphrase, word, subword).

Part-Of-Speech Tagging POS +2

Enhancing Image Rescaling using Dual Latent Variables in Invertible Neural Network

1 code implementation24 Jul 2022 Min Zhang, Zhihong Pan, Xin Zhou, C. -C. Jay Kuo

Normalizing flow models have been used successfully for generative image super-resolution (SR) by approximating complex distribution of natural images to simple tractable distribution in latent space through Invertible Neural Networks (INN).

Image Restoration Image Super-Resolution

SHLE: Devices Tracking and Depth Filtering for Stereo-based Height Limit Estimation

1 code implementation22 Dec 2022 Zhaoxin Fan, Kaixing Yang, Min Zhang, Zhenbo Song, Hongyan Liu, Jun He

In stage 1, a novel devices detection and tracking scheme is introduced, which accurately locate the height limit devices in the left or right image.

Unbiased Delayed Feedback Label Correction for Conversion Rate Prediction

1 code implementation24 Jul 2023 Yifan Wang, Peijie Sun, Min Zhang, Qinglin Jia, Jingjie Li, Shaoping Ma

To directly introduce the correct feedback label information, we propose an Unbiased delayed feedback Label Correction framework (ULC), which uses an auxiliary model to correct labels for observed negative feedback samples.

counterfactual

Separate the Wheat from the Chaff: Model Deficiency Unlearning via Parameter-Efficient Module Operation

1 code implementation16 Aug 2023 Xinshuo Hu, Dongfang Li, Baotian Hu, Zihao Zheng, Zhenyu Liu, Min Zhang

To evaluate the effectiveness of our approach in terms of truthfulness and detoxification, we conduct extensive experiments on LLMs, encompassing additional abilities such as language modeling and mathematical reasoning.

Language Modelling Mathematical Reasoning

G-SPEED: General SParse Efficient Editing MoDel

1 code implementation16 Oct 2023 Haoke Zhang, Yue Wang, Juntao Li, Xiabing Zhou, Min Zhang

Large Language Models~(LLMs) have demonstrated incredible capabilities in understanding, generating, and manipulating languages.

Beyond Hard Samples: Robust and Effective Grammatical Error Correction with Cycle Self-Augmenting

1 code implementation20 Oct 2023 Zecheng Tang, Kaifeng Qi, Juntao Li, Min Zhang

By leveraging the augmenting data from the GEC models themselves in the post-training process and introducing regularization data for cycle training, our proposed method can effectively improve the model robustness of well-trained GEC models with only a few more training epochs as an extra cost.

Adversarial Attack Grammatical Error Correction

Clustering Pseudo Language Family in Multilingual Translation Models with Fisher Information Matrix

1 code implementation5 Dec 2023 Xinyu Ma, Xuebo Liu, Min Zhang

In multilingual translation research, the comprehension and utilization of language families are of paramount importance.

Clustering Translation

To Recommend or Not: Recommendability Identification in Conversations with Pre-trained Language Models

1 code implementation27 Mar 2024 Zhefan Wang, Weizhi Ma, Min Zhang

First, we propose and define the recommendability identification task, which investigates the need for recommendations in the current conversational context.

Recommendation Systems

Common Sense Enhanced Knowledge-based Recommendation with Large Language Model

1 code implementation27 Mar 2024 Shenghao Yang, Weizhi Ma, Peijie Sun, Min Zhang, Qingyao Ai, Yiqun Liu, Mingchen Cai

Knowledge-based recommendation models effectively alleviate the data sparsity issue leveraging the side information in the knowledge graph, and have achieved considerable performance.

Common Sense Reasoning Knowledge Graphs +2

Adaptive Weighting for Neural Machine Translation

1 code implementation COLING 2018 Yachao Li, Junhui Li, Min Zhang

In the popular sequence to sequence (seq2seq) neural machine translation (NMT), there exist many weighted sum models (WSMs), each of which takes a set of input and generates one output.

Decoder Machine Translation +2

Is POS Tagging Necessary or Even Helpful for Neural Dependency Parsing?

1 code implementation6 Mar 2020 Houquan Zhou, Yu Zhang, Zhenghua Li, Min Zhang

In the pre deep learning era, part-of-speech tags have been considered as indispensable ingredients for feature engineering in dependency parsing.

Dependency Parsing Feature Engineering +4

Towards a Better Understanding Human Reading Comprehension with Brain Signals

1 code implementation3 Aug 2021 Ziyi Ye, Xiaohui Xie, Yiqun Liu, Zhihong Wang, Xuesong Chen, Min Zhang, Shaoping Ma

In this paper, we carefully design a lab-based user study to investigate brain activities during reading comprehension.

EEG Information Retrieval +4

Language Anisotropic Cross-Lingual Model Editing

1 code implementation25 May 2022 Yang Xu, Yutai Hou, Wanxiang Che, Min Zhang

On the newly defined cross-lingual model editing task, we empirically demonstrate the failure of monolingual baselines in propagating the edit to multiple languages and the effectiveness of the proposed language anisotropic model editing.

Model Editing

Measuring Item Global Residual Value for Fair Recommendation

1 code implementation17 Jul 2023 Jiayin Wang, Weizhi Ma, Chumeng Jiang, Min Zhang, Yuan Zhang, Biao Li, Peng Jiang

In this paper, we call for a shift of attention from modeling user preferences to developing fair exposure mechanisms for items.

Recommendation Systems

MAP: Towards Balanced Generalization of IID and OOD through Model-Agnostic Adapters

1 code implementation ICCV 2023 Min Zhang, Junkun Yuan, Yue He, Wenbin Li, Zhengyu Chen, Kun Kuang

To achieve this goal, we apply a bilevel optimization to explicitly model and optimize the coupling relationship between the OOD model and auxiliary adapter layers.

Bilevel Optimization Inductive Bias

LLM-enhanced Self-training for Cross-domain Constituency Parsing

1 code implementation5 Nov 2023 Jianling Li, Meishan Zhang, Peiming Guo, Min Zhang, Yue Zhang

Our experimental results demonstrate that self-training for constituency parsing, equipped with an LLM, outperforms traditional methods regardless of the LLM's performance.

Constituency Parsing Language Modelling +1

Context Consistency between Training and Testing in Simultaneous Machine Translation

1 code implementation13 Nov 2023 Meizhi Zhong, Lemao Liu, Kehai Chen, Mingming Yang, Min Zhang

Simultaneous Machine Translation (SiMT) aims to yield a real-time partial translation with a monotonically growing the source-side context.

Machine Translation Translation

A Situation-aware Enhancer for Personalized Recommendation

1 code implementation27 Mar 2024 Jiayu Li, Peijie Sun, Chumeng Jiang, Weizhi Ma, Qingyao Ai, Min Zhang

In this paper, we provide a new perspective that takes situations as the preconditions for users' interactions.

Recommendation Systems

3AM: An Ambiguity-Aware Multi-Modal Machine Translation Dataset

1 code implementation29 Apr 2024 Xinyu Ma, Xuebo Liu, Derek F. Wong, Jun Rao, Bei Li, Liang Ding, Lidia S. Chao, DaCheng Tao, Min Zhang

Experimental results show that MMT models trained on our dataset exhibit a greater ability to exploit visual information than those trained on other MMT datasets.

Multimodal Machine Translation Sentence +2

Stacked AMR Parsing with Silver Data

1 code implementation Findings (EMNLP) 2021 Qingrong Xia, Zhenghua Li, Rui Wang, Min Zhang

In particular, one recent seq-to-seq work directly fine-tunes AMR graph sequences on the encoder-decoder pre-trained language model and achieves new state-of-the-art results, outperforming previous works by a large margin.

AMR Parsing Decoder +1

Identifying Chinese Opinion Expressions with Extremely-Noisy Crowdsourcing Annotations

1 code implementation ACL 2022 Xin Zhang, Guangwei Xu, Yueheng Sun, Meishan Zhang, Xiaobin Wang, Min Zhang

Recent works of opinion expression identification (OEI) rely heavily on the quality and scale of the manually-constructed training corpus, which could be extremely difficult to satisfy.

Extending Phrase Grounding with Pronouns in Visual Dialogues

1 code implementation23 Oct 2022 Panzhong Lu, Xin Zhang, Meishan Zhang, Min Zhang

First, we construct a dataset of phrase grounding with both noun phrases and pronouns to image regions.

Phrase Grounding

Semantic Structure Enhanced Contrastive Adversarial Hash Network for Cross-media Representation Learning

2 code implementations ACM Multimedia 2022 Meiyu Liang, Junping Du, Xiaowen Cao, Yang Yu, Kangkang Lu, Zhe Xue, Min Zhang

Secondly, for further improving learning ability of implicit cross-media semantic associations, a semantic label association graph is constructed, and the graph convolutional network is utilized to mine the implicit semantic structures, thus guiding learning of discriminative features of different modalities.

Representation Learning

Revisiting Token Dropping Strategy in Efficient BERT Pretraining

1 code implementation24 May 2023 Qihuang Zhong, Liang Ding, Juhua Liu, Xuebo Liu, Min Zhang, Bo Du, DaCheng Tao

Token dropping is a recently-proposed strategy to speed up the pretraining of masked language models, such as BERT, by skipping the computation of a subset of the input tokens at several middle layers.

Empower Nested Boolean Logic via Self-Supervised Curriculum Learning

1 code implementation9 Oct 2023 Hongqiu Wu, Linfeng Liu, Hai Zhao, Min Zhang

Beyond the great cognitive powers showcased by language models, it is crucial to scrutinize whether their reasoning capabilities stem from strong generalization or merely exposure to relevant data.

Logical Reasoning Self-Supervised Learning

Caseformer: Pre-training for Legal Case Retrieval Based on Inter-Case Distinctions

1 code implementation1 Nov 2023 Weihang Su, Qingyao Ai, Yueyue Wu, Yixiao Ma, Haitao Li, Yiqun Liu, Zhijing Wu, Min Zhang

Legal case retrieval aims to help legal workers find relevant cases related to their cases at hand, which is important for the guarantee of fairness and justice in legal judgments.

Fairness Retrieval

HA-HI: Synergising fMRI and DTI through Hierarchical Alignments and Hierarchical Interactions for Mild Cognitive Impairment Diagnosis

1 code implementation2 Jan 2024 Xiongri Shen, Zhenxi Song, Linling Li, Min Zhang, Lingyan Liang Honghai Liu, Demao Deng, Zhiguo Zhang

Early diagnosis of mild cognitive impairment (MCI) and subjective cognitive decline (SCD) utilizing multi-modal magnetic resonance imaging (MRI) is a pivotal area of research.

FIPO: Free-form Instruction-oriented Prompt Optimization with Preference Dataset and Modular Fine-tuning Schema

1 code implementation19 Feb 2024 Junru Lu, Siyu An, Min Zhang, Yulan He, Di Yin, Xing Sun

In the quest to facilitate the deep intelligence of Large Language Models (LLMs) accessible in final-end user-bot interactions, the art of prompt crafting emerges as a critical yet complex task for the average user.

Improved English to Russian Translation by Neural Suffix Prediction

no code implementations11 Jan 2018 Kai Song, Yue Zhang, Min Zhang, Weihua Luo

Neural machine translation (NMT) suffers a performance deficiency when a limited vocabulary fails to cover the source or target side adequately, which happens frequently when dealing with morphologically rich languages.

Machine Translation NMT +1

SEE: Syntax-aware Entity Embedding for Neural Relation Extraction

no code implementations11 Jan 2018 Zhengqiu He, Wenliang Chen, Zhenghua Li, Meishan Zhang, Wei zhang, Min Zhang

First, we encode the context of entities on a dependency tree as sentence-level entity embedding based on tree-GRU.

Relation Relation Classification +3

Modeling Source Syntax for Neural Machine Translation

no code implementations ACL 2017 Junhui Li, Deyi Xiong, Zhaopeng Tu, Muhua Zhu, Min Zhang, Guodong Zhou

Even though a linguistics-free sequence to sequence model in neural machine translation (NMT) has certain capability of implicitly learning syntactic information of source sentences, this paper shows that source syntax can be explicitly incorporated into NMT effectively to provide further improvements.

Machine Translation NMT +1

Neural Machine Translation Advised by Statistical Machine Translation

no code implementations17 Oct 2016 Xing Wang, Zhengdong Lu, Zhaopeng Tu, Hang Li, Deyi Xiong, Min Zhang

Neural Machine Translation (NMT) is a new approach to machine translation that has made great progress in recent years.

Machine Translation NMT +1

Word Segmentation on Micro-blog Texts with External Lexicon and Heterogeneous Data

no code implementations4 Aug 2016 Qingrong Xia, Zhenghua Li, Jiayuan Chao, Min Zhang

This paper describes our system designed for the NLPCC 2016 shared task on word segmentation on micro-blog texts.

Segmentation

Training Dependency Parsers with Partial Annotation

no code implementations29 Sep 2016 Zhenghua Li, Yue Zhang, Jiayuan Chao, Min Zhang

The first approach is previously proposed to directly train a log-linear graph-based parser (LLGPar) with PA based on a forest-based objective.

Dependency Parsing

Boost Phrase-level Polarity Labelling with Review-level Sentiment Classification

no code implementations11 Feb 2015 Yongfeng Zhang, Min Zhang, Yiqun Liu, Shaoping Ma

In this paper, we focus on the problem of phrase-level sentiment polarity labelling and attempt to bridge the gap between phrase-level and review-level sentiment analysis.

Classification General Classification +2

Minimum $n$-Rank Approximation via Iterative Hard Thresholding

no code implementations18 Nov 2013 Min Zhang, Lei Yang, Zheng-Hai Huang

Additionally, combining an effective heuristic for determining $n$-rank, we can also apply the proposed algorithm to solve MnRA when $n$-rank is unknown in advance.

Image Inpainting Video Inpainting

Classification of lung nodules in CT images based on Wasserstein distance in differential geometry

no code implementations30 Jun 2018 Min Zhang, Qianli Ma, Chengfeng Wen, Hai Chen, Deruo Liu, Xianfeng GU, Jie He, Xiaoyin Xu

The Wasserstein distance between the nodules is calculated based on our new spherical optimal mass transport, this new algorithm works directly on sphere by using spherical metric, which is much more accurate and efficient than previous methods.

Computed Tomography (CT) General Classification +2

Supervised Treebank Conversion: Data and Approaches

no code implementations ACL 2018 Xinzhou Jiang, Zhenghua Li, Bo Zhang, Min Zhang, Sheng Li, Luo Si

Treebank conversion is a straightforward and effective way to exploit various heterogeneous treebanks for boosting parsing performance.

Dependency Parsing Multi-Task Learning +1

Multi-Grained Chinese Word Segmentation

no code implementations EMNLP 2017 Chen Gong, Zhenghua Li, Min Zhang, Xinzhou Jiang

Traditionally, word segmentation (WS) adopts the single-grained formalism, where a sentence corresponds to a single word sequence.

Chinese Word Segmentation Language Modelling +2

Report of NEWS 2018 Named Entity Transliteration Shared Task

no code implementations WS 2018 Nancy Chen, Rafael E. Banchs, Min Zhang, Xiangyu Duan, Haizhou Li

This report presents the results from the Named Entity Transliteration Shared Task conducted as part of The Seventh Named Entities Workshop (NEWS 2018) held at ACL 2018 in Melbourne, Australia.

Information Retrieval Transliteration

Learning Event Expressions via Bilingual Structure Projection

no code implementations COLING 2016 Fangyuan Li, Ruihong Huang, Deyi Xiong, Min Zhang

Aiming to resolve high complexities of event descriptions, previous work (Huang and Riloff, 2013) proposes multi-faceted event recognition and a bootstrapping method to automatically acquire both event facet phrases and event expressions from unannotated texts.

Distributed Representations for Building Profiles of Users and Items from Text Reviews

no code implementations COLING 2016 Wenliang Chen, Zhenjie Zhang, Zhenghua Li, Min Zhang

In this paper, we propose an approach to learn distributed representations of users and items from text comments for recommendation systems.

Collaborative Filtering Decision Making +3

Cannot find the paper you are looking for? You can Submit a new open access paper.