Search Results for author: Richong Zhang

Found 36 papers, 15 papers with code

An Unsupervised Multiple-Task and Multiple-Teacher Model for Cross-lingual Named Entity Recognition

no code implementations ACL 2022 Zhuoran Li, Chunming Hu, Xiaohui Guo, Junfan Chen, Wenyi Qin, Richong Zhang

In this study, based on the knowledge distillation framework and multi-task learning, we introduce the similarity metric model as an auxiliary task to improve the cross-lingual NER performance on the target domain.

Cross-Lingual NER Knowledge Distillation +4

Learning VAE-LDA Models with Rounded Reparameterization Trick

no code implementations EMNLP 2020 Runzhi Tian, Yongyi Mao, Richong Zhang

The introduction of VAE provides an efficient framework for the learning of generative models, including generative topic models.

Topic Models

Keyphrase Extraction with Incomplete Annotated Training Data

no code implementations WNUT (ACL) 2021 Yanfei Lei, Chunming Hu, Guanghui Ma, Richong Zhang

Extracting keyphrases that summarize the main points of a document is a fundamental task in natural language processing.

Keyphrase Extraction

Improving Composed Image Retrieval via Contrastive Learning with Scaling Positives and Negatives

no code implementations17 Apr 2024 Zhangchi Feng, Richong Zhang, Zhijie Nie

The Composed Image Retrieval (CIR) task aims to retrieve target images using a composed query consisting of a reference image and a modified text.

Towards Better Understanding of Contrastive Sentence Representation Learning: A Unified Paradigm for Gradient

no code implementations28 Feb 2024 Mingxin Li, Richong Zhang, Zhijie Nie

To address these questions, we start from the perspective of gradients and discover that four effective contrastive losses can be integrated into a unified paradigm, which depends on three components: the Gradient Dissipation, the Weight, and the Ratio.

Representation Learning Self-Supervised Learning +1

Anaphor Assisted Document-Level Relation Extraction

1 code implementation28 Oct 2023 Chonggang Lu, Richong Zhang, Kai Sun, Jaein Kim, Cunwang Zhang, Yongyi Mao

Existing methods focus on building a heterogeneous document graph to model the internal structure of an entity and the external interaction between entities.

Document-level Relation Extraction Relation +1

Narrowing the Gap between Supervised and Unsupervised Sentence Representation Learning with Large Language Model

1 code implementation12 Sep 2023 Mingxin Li, Richong Zhang, Zhijie Nie, Yongyi Mao

An intriguing phenomenon in CSE is the significant performance gap between supervised and unsupervised methods, with their only difference lying in the training data.

Attribute Contrastive Learning +5

Code-Style In-Context Learning for Knowledge-Based Question Answering

1 code implementation9 Sep 2023 Zhijie Nie, Richong Zhang, Zhongyuan Wang, Xudong Liu

Current methods for Knowledge-Based Question Answering (KBQA) usually rely on complex training techniques and model frameworks, leading to many limitations in practical applications.

Code Generation In-Context Learning +2

Adversarial Word Dilution as Text Data Augmentation in Low-Resource Regime

1 code implementation16 May 2023 Junfan Chen, Richong Zhang, Zheyan Luo, Chunming Hu, Yongyi Mao

Data augmentation is widely used in text classification, especially in the low-resource regime where a few examples for each class are available during training.

Data Augmentation text-classification +1

ContrastNet: A Contrastive Learning Framework for Few-Shot Text Classification

1 code implementation16 May 2023 Junfan Chen, Richong Zhang, Yongyi Mao, Jie Xu

Few-shot text classification has recently been promoted by the meta-learning paradigm which aims to identify target classes with knowledge transferred from source classes with sets of small tasks named episodes.

Contrastive Learning Few-Shot Text Classification +2

Self-training through Classifier Disagreement for Cross-Domain Opinion Target Extraction

no code implementations28 Feb 2023 Kai Sun, Richong Zhang, Samuel Mensah, Nikolaos Aletras, Yongyi Mao, Xudong Liu

Inspired by the theoretical foundations in domain adaptation [2], we propose a new SSL approach that opts for selecting target samples whose model output from a domain-specific teacher and student network disagree on the unlabelled target data, in an effort to boost the target domain performance.

Aspect Extraction Domain Adaptation +1

Dual Contrastive Learning: Text Classification via Label-Aware Data Augmentation

2 code implementations21 Jan 2022 Qianben Chen, Richong Zhang, Yaowei Zheng, Yongyi Mao

Contrastive learning has achieved remarkable success in representation learning via self-supervision in unsupervised settings.

Contrastive Learning Data Augmentation +5

Robust Regularization with Adversarial Labelling of Perturbed Samples

no code implementations28 May 2021 Xiaohui Guo, Richong Zhang, Yaowei Zheng, Yongyi Mao

The ALPS regularization objective is formulated as a min-max problem, in which the outer problem is minimizing an upper-bound of the VRM loss, and the inner problem is L$_1$-ball constrained adversarial labelling on perturbed sample.

Adversarial Robustness Computational Efficiency

Regularizing Neural Networks via Adversarial Model Perturbation

1 code implementation CVPR 2021 Yaowei Zheng, Richong Zhang, Yongyi Mao

This work proposes a new regularization scheme, based on the understanding that the flat local minima of the empirical risk cause the model to generalize better.

Image Classification

Parallel Interactive Networks for Multi-Domain Dialogue State Generation

1 code implementation EMNLP 2020 Junfan Chen, Richong Zhang, Yongyi Mao, Jie Xu

In this study, we argue that the incorporation of these dependencies is crucial for the design of MDST and propose Parallel Interactive Networks (PIN) to model these dependencies.

Dialogue State Tracking Multi-domain Dialogue State Tracking

Neural Dialogue State Tracking with Temporally Expressive Networks

1 code implementation Findings of the Association for Computational Linguistics 2020 Junfan Chen, Richong Zhang, Yongyi Mao, Jie Xu

Existing DST models either ignore temporal feature dependencies across dialogue turns or fail to explicitly model temporal state dependencies in a dialogue.

Dialogue State Tracking

On SkipGram Word Embedding Models with Negative Sampling: Unified Framework and Impact of Noise Distributions

no code implementations2 Sep 2020 Ziqiao Wang, Yongyi Mao, Hongyu Guo, Richong Zhang

SkipGram word embedding models with negative sampling, or SGN in short, is an elegant family of word embedding models.

Recurrent Interaction Network for Jointly Extracting Entities and Classifying Relations

no code implementations EMNLP 2020 Kai Sun, Richong Zhang, Samuel Mensah, Yongyi Mao, Xudong Liu

The idea of using multi-task learning approaches to address the joint extraction of entity and relation is motivated by the relatedness between the entity recognition task and the relation classification task.

General Classification Multi-Task Learning +3

Aspect-Level Sentiment Analysis Via Convolution over Dependency Tree

no code implementations IJCNLP 2019 Kai Sun, Richong Zhang, Samuel Mensah, Yongyi Mao, Xudong Liu

We propose a method based on neural networks to identify the sentiment polarity of opinion words expressed on a specific aspect of a sentence.

Representation Learning Sentence +2

Uncover the Ground-Truth Relations in Distant Supervision: A Neural Expectation-Maximization Framework

1 code implementation IJCNLP 2019 Junfan Chen, Richong Zhang, Yongyi Mao, Hongyu Guo, Jie Xu

Distant supervision for relation extraction enables one to effectively acquire structured relations out of very large text corpora with less human efforts.

Denoising Relation +1

MixUp as Directional Adversarial Training

no code implementations ICLR 2020 Guillaume P. Archambault, Yongyi Mao, Hongyu Guo, Richong Zhang

We prove that the family of Untied MixUp schemes is equivalent to the entire class of DAT schemes.

Augmenting Data with Mixup for Sentence Classification: An Empirical Study

2 code implementations22 May 2019 Hongyu Guo, Yongyi Mao, Richong Zhang

Mixup, a recent proposed data augmentation method through linearly interpolating inputs and modeling targets of random samples, has demonstrated its capability of significantly improving the predictive accuracy of the state-of-the-art networks for image classification.

Classification Data Augmentation +6

Understanding Feature Selection and Feature Memorization in Recurrent Neural Networks

no code implementations3 Mar 2019 Bokang Zhu, Richong Zhang, Dingkun Long, Yongyi Mao

Gated models resolve this conflict by adaptively adjusting their state-update equations, whereas Vanilla RNN resolves this conflict by assigning different dimensions different tasks.

feature selection Memorization

Syntax Encoding with Application in Authorship Attribution

no code implementations EMNLP 2018 Richong Zhang, Zhiyuan Hu, Hongyu Guo, Yongyi Mao

We propose a novel strategy to encode the syntax parse tree of sentence into a learnable distributed representation.

Authorship Attribution Benchmarking +3

MixUp as Locally Linear Out-Of-Manifold Regularization

2 code implementations7 Sep 2018 Hongyu Guo, Yongyi Mao, Richong Zhang

To address this issue, we propose a novel adaptive version of MixUp, where the mixing policies are automatically learned from the data using an additional network and objective function designed to avoid manifold intrusion.

Data Augmentation

Prototypical Recurrent Unit

no code implementations20 Nov 2016 Dingkun Long, Richong Zhang, Yongyi Mao

For this purpose, we design a simple and controllable task, called ``memorization problem'', where the networks are trained to memorize certain targeted information.

Memorization

On the representation and embedding of knowledge bases beyond binary relations

no code implementations28 Apr 2016 Jianfeng Wen, Jian-Xin Li, Yongyi Mao, Shini Chen, Richong Zhang

The models developed to date for knowledge base embedding are all based on the assumption that the relations contained in knowledge bases are binary.

Cannot find the paper you are looking for? You can Submit a new open access paper.