Search Results for author: Zichao Li

Found 24 papers, 9 papers with code

EditNTS: An Neural Programmer-Interpreter Model for Sentence Simplification through Explicit Editing

1 code implementation ACL 2019 Yue Dong, Zichao Li, Mehdi Rezagholizadeh, Jackie Chi Kit Cheung

We present the first sentence simplification model that learns explicit edit operations (ADD, DELETE, and KEEP) via a neural programmer-interpreter approach.

Machine Translation Sentence +2

Decomposable Neural Paraphrase Generation

no code implementations ACL 2019 Zichao Li, Xin Jiang, Lifeng Shang, Qun Liu

Paraphrasing exists at different granularity levels, such as lexical level, phrasal level and sentential level.

Paraphrase Generation Sentence +1

Biclustering with Alternating K-Means

no code implementations9 Sep 2020 Nicolas Fraiman, Zichao Li

Biclustering is the task of simultaneously clustering the rows and columns of the data matrix into different subgroups such that the rows and columns within a subgroup exhibit similar patterns.

Clustering

Overfitting or Underfitting? Understand Robustness Drop in Adversarial Training

2 code implementations15 Oct 2020 Zichao Li, Liyuan Liu, chengyu dong, Jingbo Shang

Our goal is to understand why the robustness drops after conducting adversarial training for too long.

Classification with Nearest Disjoint Centroids

no code implementations21 Sep 2021 Nicolas Fraiman, Zichao Li

In this paper, we develop a new classification method based on nearest centroid, and it is called the nearest disjoint centroid classifier.

Classification feature selection

BFClass: A Backdoor-free Text Classification Framework

no code implementations Findings (EMNLP) 2021 Zichao Li, Dheeraj Mekala, chengyu dong, Jingbo Shang

To recognize the poisoned subset, we examine the training samples with these identified triggers as the most suspicious token, and check if removing the trigger will change the poisoned model's prediction.

Backdoor Attack Language Modelling +2

Perturbation Deterioration: The Other Side of Catastrophic Overfitting

no code implementations29 Sep 2021 Zichao Li, Liyuan Liu, chengyu dong, Jingbo Shang

While this phenomenon is commonly explained as overfitting, we observe that it is a twin process: not only does the model catastrophic overfits to one type of perturbation, but also the perturbation deteriorates into random noise.

Beam Search for Feature Selection

no code implementations8 Mar 2022 Nicolas Fraiman, Zichao Li

In this paper, we present and prove some consistency results about the performance of classification models using a subset of features.

Classification feature selection

Text Revision by On-the-Fly Representation Optimization

1 code implementation In2Writing (ACL) 2022 Jingjing Li, Zichao Li, Tao Ge, Irwin King, Michael R. Lyu

In this approach, we simply fine-tune a pre-trained Transformer with masked language modeling and attribute classification.

Attribute Language Modelling +3

Bag of Tricks for FGSM Adversarial Training

no code implementations6 Sep 2022 Zichao Li, Li Liu, Zeyu Wang, Yuyin Zhou, Cihang Xie

Adversarial training (AT) with samples generated by Fast Gradient Sign Method (FGSM), also known as FGSM-AT, is a computationally simple method to train robust networks.

On the Adversarial Robustness of Camera-based 3D Object Detection

1 code implementation25 Jan 2023 Shaoyuan Xie, Zichao Li, Zeyu Wang, Cihang Xie

In recent years, camera-based 3D object detection has gained widespread attention for its ability to achieve high performance with low computational cost.

3D Object Detection Adversarial Attack +5

Tied-Augment: Controlling Representation Similarity Improves Data Augmentation

1 code implementation22 May 2023 Emirhan Kurtulus, Zichao Li, Yann Dauphin, Ekin Dogus Cubuk

For example, even the simple flips-and-crops augmentation requires training for more than 5 epochs to improve performance, whereas RandAugment requires more than 90 epochs.

Data Augmentation

f-Divergence Minimization for Sequence-Level Knowledge Distillation

1 code implementation27 Jul 2023 Yuqiao Wen, Zichao Li, Wenyu Du, Lili Mou

Experiments across four datasets show that our methods outperform existing KD approaches, and that our symmetric distilling losses can better force the student to learn from the teacher distribution.

Knowledge Distillation

Scaling (Down) CLIP: A Comprehensive Analysis of Data, Architecture, and Training Strategies

no code implementations12 Apr 2024 Zichao Li, Cihang Xie, Ekin Dogus Cubuk

With regards to data, we demonstrate the significance of high-quality training data and show that a smaller dataset of high-quality data can outperform a larger dataset with lower quality.

Data Augmentation

Towards Adaptive Residual Network Training: A Neural-ODE Perspective

1 code implementation ICML 2020 chengyu dong, Liyuan Liu, Zichao Li, Jingbo Shang

Serving as a crucial factor, the depth of residual networks balances model capacity, performance, and training efficiency.

Cannot find the paper you are looking for? You can Submit a new open access paper.