no code implementations • 1 Nov 2023 • Ta-Chung Chi, Ting-Han Fan, Alexander I. Rudnicky
This suggests that a flexible positional embedding design and attention alignment can go a long way toward Transformer length extrapolation.
1 code implementation • 14 Sep 2023 • Ting-Han Fan, Ta-Chung Chi, Alexander I. Rudnicky
In recent studies, linear recurrent neural networks (LRNNs) have achieved Transformer-level performance in natural language and long-range modeling, while offering rapid parallel training and constant inference cost.
1 code implementation • SIGDIAL (ACL) 2022 • Ta-Chung Chi, Alexander I. Rudnicky
In addition, unlike in previous work, we do not rely on hand-crafted features; this improves the model's robustness.
Ranked #1 on Discourse Parsing on STAC
no code implementations • 24 May 2023 • Yau-Shian Wang, Ta-Chung Chi, Ruohong Zhang, Yiming Yang
We present PESCO, a novel contrastive learning framework that substantially improves the performance of zero-shot text classification.
no code implementations • 23 May 2023 • Ta-Chung Chi, Ting-Han Fan, Li-Wei Chen, Alexander I. Rudnicky, Peter J. Ramadge
The use of positional embeddings in transformer language models is widely accepted.
no code implementations • 5 May 2023 • Ta-Chung Chi, Ting-Han Fan, Alexander I. Rudnicky, Peter J. Ramadge
Unlike recurrent models, conventional wisdom has it that Transformers cannot perfectly model regular languages.
no code implementations • 20 Dec 2022 • Ta-Chung Chi, Ting-Han Fan, Alexander I. Rudnicky, Peter J. Ramadge
Length extrapolation permits training a transformer language model on short sequences that preserves perplexities when tested on substantially longer sequences.
no code implementations • 8 Oct 2022 • Tzu-Hsiang Lin, Ta-Chung Chi, Anna Rumshisky
Recent advancements in dialogue response selection (DRS) are based on the \textit{task-adaptive pre-training (TAP)} approach, by first initializing their model with BERT~\cite{devlin-etal-2019-bert}, and adapt to dialogue data with dialogue-specific or fine-grained pre-training tasks.
1 code implementation • 15 Jun 2022 • Ting-Han Fan, Ta-Chung Chi, Alexander I. Rudnicky, Peter J. Ramadge
While deep generative models have succeeded in image processing, natural language processing, and reinforcement learning, training that involves discrete random variables remains challenging due to the high variance of its gradient estimation process.
2 code implementations • 20 May 2022 • Ta-Chung Chi, Ting-Han Fan, Peter J. Ramadge, Alexander I. Rudnicky
Relative positional embeddings (RPE) have received considerable attention since RPEs effectively model the relative distance among tokens and enable length extrapolation.
1 code implementation • EMNLP 2021 • Ta-Chung Chi, Alexander I. Rudnicky
In this paper, we are the first to propose a~\textbf{zero-shot} dialogue disentanglement solution.
no code implementations • 12 Oct 2021 • Ting-Rui Chiang, Yi-Ting Yeh, Ta-Chung Chi, Yau-Shian Wang
ALFRED is a recently proposed benchmark that requires a model to complete tasks in simulated house environments specified by instructions in natural language.
1 code implementation • 15 Dec 2020 • Shentong Mo, Haofan Wang, Pinxu Ren, Ta-Chung Chi
Automatic speech verification (ASV) is the technology to determine the identity of a person based on their voice.
no code implementations • 2 Dec 2019 • Ta-Chung Chi, Mihail Eric, Seokhwan Kim, Minmin Shen, Dilek Hakkani-Tur
We demonstrate the proposed strategy is substantially more realistic and data-efficient compared to previously proposed pre-exploration techniques.
2 code implementations • 21 Oct 2018 • Ta-Chung Chi, Ching-Yen Shih, Yun-Nung Chen
This paper introduces the first dataset for evaluating English-Chinese Bilingual Contextual Word Similarity, namely BCWS (https://github. com/MiuLab/BCWS).
1 code implementation • EMNLP 2018 • Ta-Chung Chi, Yun-Nung Chen
The model is evaluated on the Stanford Contextual Word Similarity (SCWS) dataset to ensure the quality of monolingual sense embeddings.
1 code implementation • 10 Sep 2018 • Ting-Yun Chang, Ta-Chung Chi, Shang-Chi Tsai, Yun-Nung Chen
This paper focuses on interpreting the embeddings for various aspects, including sense separation in the vector dimensions and definition generation.
1 code implementation • IJCNLP 2017 • Ta-Chung Chi, Po-Chun Chen, Shang-Yu Su, Yun-Nung Chen
Language understanding (LU) and dialogue policy learning are two essential components in conversational systems.
1 code implementation • 30 Sep 2017 • Po-Chun Chen, Ta-Chung Chi, Shang-Yu Su, Yun-Nung Chen
However, the previous model only paid attention to the content in history utterances without considering their temporal information and speaker roles.