no code implementations • 30 Nov 2023 • James Seale Smith, Yen-Chang Hsu, Zsolt Kira, Yilin Shen, Hongxia Jin
We show that STAMINA outperforms the prior SOTA for the setting of text-to-image continual customization on a 50-concept benchmark composed of landmarks and human faces, with no stored replay data.
no code implementations • 12 Apr 2023 • James Seale Smith, Yen-Chang Hsu, Lingyu Zhang, Ting Hua, Zsolt Kira, Yilin Shen, Hongxia Jin
We show that C-LoRA not only outperforms several baselines for our proposed setting of text-to-image continual customization, which we refer to as Continual Diffusion, but that we achieve a new state-of-the-art in the well-established rehearsal-free continual learning setting for image classification.
no code implementations • 6 Apr 2023 • Yashas Malur Saidutta, Rakshith Sharma Srinivasa, Ching-Hua Lee, Chouchang Yang, Yilin Shen, Hongxia Jin
We show that existing deep keyword spotting mechanisms can be improved by Successive Refinement, where the system first classifies whether the input audio is speech or not, followed by whether the input is keyword-like or not, and finally classifies which keyword was uttered.
no code implementations • 30 Jan 2023 • Kaiwen Zhou, Kaizhi Zheng, Connor Pryor, Yilin Shen, Hongxia Jin, Lise Getoor, Xin Eric Wang
Such object navigation tasks usually require large-scale training in visual environments with labeled objects, which generalizes poorly to novel objects in unknown environments.
no code implementations • 13 Jan 2023 • Miao Yin, Burak Uzkent, Yilin Shen, Hongxia Jin, Bo Yuan
We first develop a graph-based ranking for measuring the importance of attention heads, and the extracted importance information is further integrated to an optimization-based procedure to impose the heterogeneous structured sparsity patterns on the ViT models.
no code implementations • 2 Nov 2022 • Ting Hua, Yen-Chang Hsu, Felicity Wang, Qian Lou, Yilin Shen, Hongxia Jin
However, standard SVD treats the parameters within the matrix with equal importance, which is a simple but unrealistic assumption.
no code implementations • ICLR 2022 • Yen-Chang Hsu, Ting Hua, SungEn Chang, Qian Lou, Yilin Shen, Hongxia Jin
In other words, the optimization objective of SVD is not aligned with the trained model's task accuracy.
no code implementations • 18 Mar 2022 • Yen-Chang Hsu, James Smith, Yilin Shen, Zsolt Kira, Hongxia Jin
Knowledge distillation (KD) is a substantial strategy for transferring learned knowledge from one neural network model to another.
no code implementations • 25 Jan 2022 • Peixi Xiong, Yilin Shen, Hongxia Jin
In contrast to previous works, our model splits alignment into different levels to achieve learning better correlations without needing additional data and annotations.
no code implementations • NAACL 2021 • Ting Hua, Yilin Shen, Changsheng Zhao, Yen-Chang Hsu, Hongxia Jin
Most existing continual learning approaches suffer from low accuracy and performance fluctuation, especially when the distributions of old and new data are significantly different.
no code implementations • CVPR 2022 • Qian Lou, Yen-Chang Hsu, Burak Uzkent, Ting Hua, Yilin Shen, Hongxia Jin
The key primitive is that Dictionary-Lookup-Transformormations (DLT) is proposed to replace Linear Transformation (LT) in multi-modal detectors where each weight in Linear Transformation (LT) is approximately factorized into a smaller dictionary, index, and coefficient.
no code implementations • 30 Dec 2021 • Changsheng Zhao, Ting Hua, Yilin Shen, Qian Lou, Hongxia Jin
Knowledge distillation, Weight pruning, and Quantization are known to be the main directions in model compression.
no code implementations • 28 Oct 2021 • Junjiao Tian, Yen-Change Hsu, Yilin Shen, Hongxia Jin, Zsolt Kira
We are the first to propose a method that works well across both OOD detection and calibration and under different types of shifts.
no code implementations • 29 Sep 2021 • Junjiao Tian, Yen-Chang Hsu, Yilin Shen, Hongxia Jin, Zsolt Kira
To this end, we theoretically derive two score functions for OOD detection, the covariate shift score and concept shift score, based on the decomposition of KL-divergence for both scores, and propose a geometrically-inspired method (Geometric ODIN) to improve OOD detection under both shifts with only in-distribution data.
no code implementations • ICLR 2022 • Qian Lou, Ting Hua, Yen-Chang Hsu, Yilin Shen, Hongxia Jin
DictFormer significantly reduces the redundancy in the transformer's parameters by replacing the prior transformer's parameters with compact, shared dictionary, a few unshared coefficients, and indices.
no code implementations • ACL 2021 • Yilin Shen, Yen-Chang Hsu, Avik Ray, Hongxia Jin
This paper proposes to train a model with only IND data while supporting both IND intent classification and OOD detection.
2 code implementations • ICCV 2021 • James Smith, Yen-Chang Hsu, Jonathan Balloch, Yilin Shen, Hongxia Jin, Zsolt Kira
Modern computer vision applications suffer from catastrophic forgetting when incrementally learning new concepts over time.
Ranked #5 on
Class Incremental Learning
on cifar100
no code implementations • 6 Jun 2021 • Yu Wang, Yilin Shen, Hongxia Jin
In this paper, we introduce a novel multi-step spoken language understanding system based on adversarial learning that can leverage the multiround user's feedback to update slot values.
no code implementations • ICLR 2021 • Qian Lou, Yilin Shen, Hongxia Jin, Lei Jiang
A cryptographic neural network inference service is an efficient way to allow two parties to execute neural network inference without revealing either party’s data or model.
no code implementations • 16 Oct 2020 • Yilin Shen, Wenhu Chen, Hongxia Jin
We design a Dirichlet Prior RNN to model high-order uncertainty by degenerating as softmax layer for RNN model training.
no code implementations • EMNLP 2020 • Wei-Jen Ko, Avik Ray, Yilin Shen, Hongxia Jin
Existing open-domain dialogue generation models are usually trained to mimic the gold response in the training set using cross-entropy loss on the vocabulary.
3 code implementations • 4 May 2020 • Yang Cao, Yonghui Xiao, Shun Takagi, Li Xiong, Masatoshi Yoshikawa, Yilin Shen, Jinfei Liu, Hongxia Jin, Xiaofeng Xu
Third, we design a private location trace release framework that pipelines the detection of location exposure, policy graph repair, and private trajectory release with customizable and rigorous location privacy.
Cryptography and Security Computers and Society
no code implementations • 4 May 2020 • Ruiyi Zhang, Tong Yu, Yilin Shen, Hongxia Jin, Changyou Chen, Lawrence Carin
Text-based interactive recommendation provides richer user feedback and has demonstrated advantages over traditional interactive recommender systems.
2 code implementations • CVPR 2020 • Yen-Chang Hsu, Yilin Shen, Hongxia Jin, Zsolt Kira
Deep neural networks have attained remarkable performance when applied to data that comes from the same distribution as that of the training set, but can significantly degrade otherwise.
Out-of-Distribution Detection
Out of Distribution (OOD) Detection
no code implementations • NeurIPS 2019 • Ruiyi Zhang, Tong Yu, Yilin Shen, Hongxia Jin, Changyou Chen
Text-based interactive recommendation provides richer user preferences and has demonstrated advantages over traditional interactive recommender systems.
no code implementations • IJCNLP 2019 • Yilin Shen, Xiangyu Zeng, Hongxia Jin
ProgModel consists of a novel context gate that transfers previously learned knowledge to a small size expanded component; and meanwhile enables this new component to be fast trained to learn from new data.
no code implementations • WS 2019 • Avik Ray, Yilin Shen, Hongxia Jin
However, state-of-the art attention based neural parsers are slow to retrain which inhibits real time domain adaptation.
no code implementations • 15 Oct 2019 • Avik Ray, Yilin Shen, Hongxia Jin
Recurrent neural network (RNN) based joint intent classification and slot tagging models have achieved tremendous success in recent years for building spoken language understanding and dialog systems.
no code implementations • NAACL 2019 • Yilin Shen, Avik Ray, Hongxia Jin, S Nama, eep
We present SkillBot that takes the first step to enable end users to teach new skills in personal assistants (PA).
1 code implementation • NAACL 2019 • Wenhu Chen, Yu Su, Yilin Shen, Zhiyu Chen, Xifeng Yan, William Wang
Under deep neural networks, a pre-defined vocabulary is required to vectorize text inputs.
no code implementations • ICCV 2019 • Ramprasaath R. Selvaraju, Stefan Lee, Yilin Shen, Hongxia Jin, Shalini Ghosh, Larry Heck, Dhruv Batra, Devi Parikh
Many vision and language models suffer from poor visual grounding - often falling back on easy-to-learn language priors rather than basing their decisions on visual concepts in the image.
1 code implementation • NAACL 2018 • Yu Wang, Yilin Shen, Hongxia Jin
The most effective algorithms are based on the structures of sequence to sequence models (or "encoder-decoder" models), and generate the intents and semantic tags either using separate models or a joint model.
Ranked #1 on
Intent Detection
on ATIS
no code implementations • ICLR 2019 • Wenhu Chen, Yilin Shen, Hongxia Jin, William Wang
With the recently rapid development in deep learning, deep neural networks have been widely adopted in many real-life applications.
no code implementations • 18 Sep 2018 • Yilin Shen, Xiangyu Zeng, Yu Wang, Hongxia Jin
The results show that our approach leverages such simple user information to outperform state-of-the-art approaches by 0. 25% for intent detection and 0. 31% for slot filling using standard training data.
no code implementations • 17 Sep 2018 • Avik Ray, Yilin Shen, Hongxia Jin
Learning intents and slot labels from user utterances is a fundamental step in all spoken language understanding (SLU) and dialog systems.
no code implementations • ACL 2018 • Yilin Shen, Avik Ray, Abhishek Patel, Hongxia Jin
We present a system, CRUISE, that guides ordinary software developers to build a high quality natural language understanding (NLU) engine from scratch.
no code implementations • 22 Jun 2018 • Xinlei Pan, Eshed Ohn-Bar, Nicholas Rhinehart, Yan Xu, Yilin Shen, Kris M. Kitani
The learning process is interactive, with a human expert first providing input in the form of full demonstrations along with some subgoal states.