Search Results for author: chengyu dong

Found 18 papers, 8 papers with code

“Average” Approximates “First Principal Component”? An Empirical Analysis on Representations from Neural Language Models

no code implementations EMNLP 2021 Zihan Wang, chengyu dong, Jingbo Shang

In this paper, we present an empirical property of these representations—”average” approximates “first principal component”.

Towards Adaptive Residual Network Training: A Neural-ODE Perspective

1 code implementation ICML 2020 chengyu dong, Liyuan Liu, Zichao Li, Jingbo Shang

Serving as a crucial factor, the depth of residual networks balances model capacity, performance, and training efficiency.

Fast-ELECTRA for Efficient Pre-training

no code implementations11 Oct 2023 chengyu dong, Liyuan Liu, Hao Cheng, Jingbo Shang, Jianfeng Gao, Xiaodong Liu

Although ELECTRA offers a significant boost in efficiency, its potential is constrained by the training cost brought by the auxiliary model.

Language Modelling

Robust and Interpretable Medical Image Classifiers via Concept Bottleneck Models

no code implementations4 Oct 2023 An Yan, Yu Wang, Yiwu Zhong, Zexue He, Petros Karypis, Zihan Wang, chengyu dong, Amilcare Gentili, Chun-Nan Hsu, Jingbo Shang, Julian McAuley

Medical image classification is a critical problem for healthcare, with the potential to alleviate the workload of doctors and facilitate diagnoses of patients.

Image Classification Language Modelling +1

Learning Concise and Descriptive Attributes for Visual Recognition

1 code implementation ICCV 2023 An Yan, Yu Wang, Yiwu Zhong, chengyu dong, Zexue He, Yujie Lu, William Wang, Jingbo Shang, Julian McAuley

Recent advances in foundation models present new opportunities for interpretable visual recognition -- one can first query Large Language Models (LLMs) to obtain a set of attributes that describe each class, then apply vision-language models to classify images via these attributes.

Descriptive

SELFOOD: Self-Supervised Out-Of-Distribution Detection via Learning to Rank

1 code implementation24 May 2023 Dheeraj Mekala, Adithya Samavedhi, chengyu dong, Jingbo Shang

To address the annotation bottleneck, we introduce SELFOOD, a self-supervised OOD detection method that requires only in-distribution samples as supervision.

Learning-To-Rank Out-of-Distribution Detection +1

Debiasing Made State-of-the-art: Revisiting the Simple Seed-based Weak Supervision for Text Classification

1 code implementation24 May 2023 chengyu dong, Zihan Wang, Jingbo Shang

We show that the limited performance of seed matching is largely due to the label bias injected by the simple seed-match rule, which prevents the classifier from learning reliable confidence for selecting high-quality pseudo-labels.

text-classification Text Classification

SoTeacher: A Student-oriented Teacher Network Training Framework for Knowledge Distillation

no code implementations14 Jun 2022 chengyu dong, Liyuan Liu, Jingbo Shang

To fill this gap, we propose a novel student-oriented teacher network training framework SoTeacher, inspired by recent findings that student performance hinges on teacher's capability to approximate the true label distribution of training samples.

Data Augmentation Knowledge Distillation

LOPS: Learning Order Inspired Pseudo-Label Selection for Weakly Supervised Text Classification

1 code implementation25 May 2022 Dheeraj Mekala, chengyu dong, Jingbo Shang

Weakly supervised text classification methods typically train a deep neural classifier based on pseudo-labels.

Memorization Pseudo Label +2

A Heterogeneous Graph Learning Model for Cyber-Attack Detection

no code implementations16 Dec 2021 Mingqi Lv, chengyu dong, Tieming Chen, Tiantian Zhu, Qijie Song, Yuan Fan

To effective and efficient detect cyber-attacks from a huge number of system events in the provenance data, we firstly model the provenance data by a heterogeneous graph to capture the rich context information of each system entities (e. g., process, file, socket, etc.

Cyber Attack Detection Graph Learning +1

Perturbation Deterioration: The Other Side of Catastrophic Overfitting

no code implementations29 Sep 2021 Zichao Li, Liyuan Liu, chengyu dong, Jingbo Shang

While this phenomenon is commonly explained as overfitting, we observe that it is a twin process: not only does the model catastrophic overfits to one type of perturbation, but also the perturbation deteriorates into random noise.

BFClass: A Backdoor-free Text Classification Framework

no code implementations Findings (EMNLP) 2021 Zichao Li, Dheeraj Mekala, chengyu dong, Jingbo Shang

To recognize the poisoned subset, we examine the training samples with these identified triggers as the most suspicious token, and check if removing the trigger will change the poisoned model's prediction.

Backdoor Attack Language Modelling +2

"Average" Approximates "First Principal Component"? An Empirical Analysis on Representations from Neural Language Models

1 code implementation18 Apr 2021 Zihan Wang, chengyu dong, Jingbo Shang

In this paper, we present an empirical property of these representations -- "average" approximates "first principal component".

Data Quality Matters For Adversarial Training: An Empirical Study

1 code implementation15 Feb 2021 chengyu dong, Liyuan Liu, Jingbo Shang

Specifically, we first propose a strategy to measure the data quality based on the learning behaviors of the data during adversarial training and find that low-quality data may not be useful and even detrimental to the adversarial robustness.

Adversarial Robustness

Overfitting or Underfitting? Understand Robustness Drop in Adversarial Training

2 code implementations15 Oct 2020 Zichao Li, Liyuan Liu, chengyu dong, Jingbo Shang

Our goal is to understand why the robustness drops after conducting adversarial training for too long.

Cannot find the paper you are looking for? You can Submit a new open access paper.