Search Results for author: Yen-Chang Hsu

Found 25 papers, 9 papers with code

Token Fusion: Bridging the Gap between Token Pruning and Token Merging

no code implementations2 Dec 2023 Minchul Kim, Shangqian Gao, Yen-Chang Hsu, Yilin Shen, Hongxia Jin

In this paper, we introduce "Token Fusion" (ToFu), a method that amalgamates the benefits of both token pruning and token merging.

Computational Efficiency Image Generation

Continual Diffusion with STAMINA: STack-And-Mask INcremental Adapters

no code implementations30 Nov 2023 James Seale Smith, Yen-Chang Hsu, Zsolt Kira, Yilin Shen, Hongxia Jin

We show that STAMINA outperforms the prior SOTA for the setting of text-to-image continual customization on a 50-concept benchmark composed of landmarks and human faces, with no stored replay data.

Continual Learning Hard Attention +1

Continual Diffusion: Continual Customization of Text-to-Image Diffusion with C-LoRA

no code implementations12 Apr 2023 James Seale Smith, Yen-Chang Hsu, Lingyu Zhang, Ting Hua, Zsolt Kira, Yilin Shen, Hongxia Jin

We show that C-LoRA not only outperforms several baselines for our proposed setting of text-to-image continual customization, which we refer to as Continual Diffusion, but that we achieve a new state-of-the-art in the well-established rehearsal-free continual learning setting for image classification.

Continual Learning Image Classification

Numerical Optimizations for Weighted Low-rank Estimation on Language Model

no code implementations2 Nov 2022 Ting Hua, Yen-Chang Hsu, Felicity Wang, Qian Lou, Yilin Shen, Hongxia Jin

However, standard SVD treats the parameters within the matrix with equal importance, which is a simple but unrealistic assumption.

Language Modelling

A Closer Look at Rehearsal-Free Continual Learning

no code implementations31 Mar 2022 James Seale Smith, Junjiao Tian, Shaunak Halbe, Yen-Chang Hsu, Zsolt Kira

Next, we explore how to leverage knowledge from a pre-trained model in rehearsal-free continual learning and find that vanilla L2 parameter regularization outperforms EWC parameter regularization and feature distillation.

Continual Learning Knowledge Distillation +2

A Closer Look at Knowledge Distillation with Features, Logits, and Gradients

no code implementations18 Mar 2022 Yen-Chang Hsu, James Smith, Yilin Shen, Zsolt Kira, Hongxia Jin

Knowledge distillation (KD) is a substantial strategy for transferring learned knowledge from one neural network model to another.

Incremental Learning Knowledge Distillation +2

Hyperparameter-free Continuous Learning for Domain Classification in Natural Language Understanding

no code implementations NAACL 2021 Ting Hua, Yilin Shen, Changsheng Zhao, Yen-Chang Hsu, Hongxia Jin

Most existing continual learning approaches suffer from low accuracy and performance fluctuation, especially when the distributions of old and new data are significantly different.

Continual Learning domain classification +1

Lite-MDETR: A Lightweight Multi-Modal Detector

no code implementations CVPR 2022 Qian Lou, Yen-Chang Hsu, Burak Uzkent, Ting Hua, Yilin Shen, Hongxia Jin

The key primitive is that Dictionary-Lookup-Transformormations (DLT) is proposed to replace Linear Transformation (LT) in multi-modal detectors where each weight in Linear Transformation (LT) is approximately factorized into a smaller dictionary, index, and coefficient.

object-detection Object Detection +3

A Geometric Perspective towards Neural Calibration via Sensitivity Decomposition

1 code implementation NeurIPS 2021 Junjiao Tian, Dylan Yung, Yen-Chang Hsu, Zsolt Kira

It is well known that vision classification models suffer from poor calibration in the face of data distribution shifts.

Exploring Covariate and Concept Shift for Detection and Confidence Calibration of Out-of-Distribution Data

no code implementations29 Sep 2021 Junjiao Tian, Yen-Chang Hsu, Yilin Shen, Hongxia Jin, Zsolt Kira

To this end, we theoretically derive two score functions for OOD detection, the covariate shift score and concept shift score, based on the decomposition of KL-divergence for both scores, and propose a geometrically-inspired method (Geometric ODIN) to improve OOD detection under both shifts with only in-distribution data.

Out of Distribution (OOD) Detection

DictFormer: Tiny Transformer with Shared Dictionary

no code implementations ICLR 2022 Qian Lou, Ting Hua, Yen-Chang Hsu, Yilin Shen, Hongxia Jin

DictFormer significantly reduces the redundancy in the transformer's parameters by replacing the prior transformer's parameters with compact, shared dictionary, a few unshared coefficients, and indices.

Abstractive Text Summarization Language Modelling +2

Memory-Efficient Semi-Supervised Continual Learning: The World is its Own Replay Buffer

1 code implementation23 Jan 2021 James Smith, Jonathan Balloch, Yen-Chang Hsu, Zsolt Kira

Our work investigates whether we can significantly reduce this memory budget by leveraging unlabeled data from an agent's environment in a realistic and challenging continual learning paradigm.

Continual Learning Knowledge Distillation +1

Posterior Re-calibration for Imbalanced Datasets

no code implementations NeurIPS 2020 Junjiao Tian, Yen-Cheng Liu, Nathan Glaser, Yen-Chang Hsu, Zsolt Kira

Neural Networks can perform poorly when the training label distribution is heavily imbalanced, as well as when the testing data differs from the training distribution.

Long-tail Learning Semantic Segmentation

Generalized ODIN: Detecting Out-of-distribution Image without Learning from Out-of-distribution Data

2 code implementations CVPR 2020 Yen-Chang Hsu, Yilin Shen, Hongxia Jin, Zsolt Kira

Deep neural networks have attained remarkable performance when applied to data that comes from the same distribution as that of the training set, but can significantly degrade otherwise.

Out-of-Distribution Detection Out of Distribution (OOD) Detection

Multi-class Classification without Multi-class Labels

1 code implementation ICLR 2019 Yen-Chang Hsu, Zhaoyang Lv, Joel Schlosser, Phillip Odom, Zsolt Kira

This work presents a new strategy for multi-class classification that requires no class-specific labels, but instead leverages pairwise similarity between examples, which is a weaker form of annotation.

Classification General Classification +1

A probabilistic constrained clustering for transfer learning and image category discovery

no code implementations28 Jun 2018 Yen-Chang Hsu, Zhaoyang Lv, Joel Schlosser, Phillip Odom, Zsolt Kira

The proposed objective directly minimizes the negative log-likelihood of cluster assignment with respect to the pairwise constraints, has no hyper-parameters, and demonstrates improved scalability and performance on both supervised learning and unsupervised transfer learning.

Constrained Clustering Deep Clustering +2

Learning to Cluster for Proposal-Free Instance Segmentation

1 code implementation17 Mar 2018 Yen-Chang Hsu, Zheng Xu, Zsolt Kira, Jiawei Huang

We utilize the most fundamental property of instance labeling -- the pairwise relationship between pixels -- as the supervision to formulate the learning objective, then apply it to train a fully convolutional network (FCN) for learning to perform pixel-wise clustering.

Autonomous Driving Clustering +6

Learning to cluster in order to transfer across domains and tasks

1 code implementation ICLR 2018 Yen-Chang Hsu, Zhaoyang Lv, Zsolt Kira

The key insight is that, in addition to features, we can transfer similarity information and this is sufficient to learn a similarity function and clustering network to perform both domain adaptation and cross-task transfer learning.

Constrained Clustering Transfer Learning +1

Deep Image Category Discovery using a Transferred Similarity Function

no code implementations5 Dec 2016 Yen-Chang Hsu, Zhaoyang Lv, Zsolt Kira

We propose that this network can be learned with contrastive loss which is only based on weak binary pair-wise constraints.

Clustering Transfer Learning

Neural network-based clustering using pairwise constraints

2 code implementations19 Nov 2015 Yen-Chang Hsu, Zsolt Kira

Robustness analysis also shows that the method is largely insensitive to the number of clusters.

Clustering

Cannot find the paper you are looking for? You can Submit a new open access paper.