1 code implementation • 12 Oct 2022 • Ke-Han Lu, Kuan-Yu Chen
Non-autoregressive automatic speech recognition (ASR) modeling has received increasing attention recently because of its fast decoding speed and superior performance.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
1 code implementation • 18 Sep 2023 • Yi-Wei Wang, Ke-Han Lu, Kuan-Yu Chen
In addition, we implement and compare several classic and representative methods, showing the recent research progress in revising speech recognition results.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 14 Aug 2021 • Tim Dunn, Harisankar Sadasivan, Jack Wadden, Kush Goliya, Kuan-Yu Chen, Reetuparna Das, David Blaauw, Satish Narayanasamy
The MinION is a recent-to-market handheld nanopore sequencer.
1 code implementation • 29 May 2023 • Kuan-Yu Chen, Ping-Han Chiang, Hsin-Rung Chou, Ting-Wei Chen, Tien-Hao Chang
However, based on a recently published tabular benchmark, we can see deep neural networks still fall behind tree-based models on tabular datasets.
no code implementations • 7 Apr 2018 • Chih-Wei Lee, Yau-Shian Wang, Tsung-Yuan Hsu, Kuan-Yu Chen, Hung-Yi Lee, Lin-shan Lee
Conventional seq2seq chatbot models only try to find the sentences with the highest probabilities conditioned on the input sequences, without considering the sentiment of the output sentences.
no code implementations • 1 Apr 2018 • Da-Rong Liu, Kuan-Yu Chen, Hung-Yi Lee, Lin-shan Lee
Unsupervised discovery of acoustic tokens from audio corpora without annotation and learning vector representations for these tokens have been widely studied.
no code implementations • COLING 2016 • Kuan-Yu Chen, Shih-Hung Liu, Berlin Chen, Hsin-Min Wang
The D-EV model not only inherits the advantages of the EV model but also can infer a more robust representation for a given spoken paragraph against imperfect speech recognition.
no code implementations • 22 Jul 2016 • Kuan-Yu Chen, Shih-Hung Liu, Berlin Chen, Hsin-Min Wang, Hsin-Hsi Chen
Word embedding methods revolve around learning continuous distributed vector representations of words with neural networks, which can capture semantic and/or syntactic cues, and in turn be used to induce similarity measures among words, sentences and documents in context.
no code implementations • 20 Jan 2016 • Kuan-Yu Chen, Shih-Hung Liu, Berlin Chen, Hsin-Min Wang
In addition to MMR, there is only a dearth of research concentrating on reducing redundancy or increasing diversity for the spoken document summarization task, as far as we are aware.
no code implementations • 14 Jun 2015 • Kuan-Yu Chen, Shih-Hung Liu, Hsin-Min Wang, Berlin Chen, Hsin-Hsi Chen
Owing to the rapidly growing multimedia content available on the Internet, extractive spoken document summarization, with the purpose of automatically selecting a set of representative sentences from a spoken document to concisely express the most important theme of the document, has been an active area of research and experimentation.
no code implementations • 8 Apr 2019 • Kuan-Yu Chen, Che-Ping Tsai, Da-Rong Liu, Hung-Yi Lee, Lin-shan Lee
Producing a large annotated speech corpus for training ASR systems remains difficult for more than 95% of languages all over the world which are low-resourced, but collecting a relatively big unlabeled data set for such languages is more achievable.
no code implementations • 31 Oct 2019 • Li-Phen Yen, Zhen-Yu Wu, Kuan-Yu Chen
Recent developments in deep learning have led to a significant innovation in various classic and practical subjects, including speech recognition, computer vision, question answering, information retrieval and so on.
no code implementations • 25 May 2020 • Chia-Chih Kuo, Shang-Bao Luo, Kuan-Yu Chen
In a spoken multiple-choice question answering (SMCQA) task, given a passage, a question, and multiple choices all in the form of speech, the machine needs to pick the correct choice to answer the question.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
no code implementations • 11 Jul 2020 • Hung-Yi Lee, Cheng-Hao Ho, Chien-Fu Lin, Chiung-Chih Chang, Chih-Wei Lee, Yau-Shian Wang, Tsung-Yuan Hsu, Kuan-Yu Chen
Conventional seq2seq chatbot models attempt only to find sentences with the highest probabilities conditioned on the input sequences, without considering the sentiment of the output sentences.
no code implementations • 30 Jan 2021 • Wen-Chin Huang, Chia-Hua Wu, Shang-Bao Luo, Kuan-Yu Chen, Hsin-Min Wang, Tomoki Toda
We propose a simple method for automatic speech recognition (ASR) by fine-tuning BERT, which is a language model (LM) trained on large-scale unlabeled text data and can generate rich contextual representations.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 10 Apr 2021 • Fu-Hao Yu, Kuan-Yu Chen
Transformer-based models have led to significant innovation in classical and practical subjects as varied as speech processing, natural language processing, and computer vision.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 24 Jun 2021 • Ke-Han Lu, Bo-Han Fang, Kuan-Yu Chen
In this paper, inspired by the successes of visionlanguage pre-trained models and the benefits from training with adversarial attacks, we present a novel transformerbased cross-modal fusion modeling by incorporating the both notions for VQA challenge 2021.
no code implementations • ROCLING 2021 • Cheng-Chung Fan, Chia-Chih Kuo, Shang-Bao Luo, Pei-Jun Liao, Kuang-Yu Chang, Chiao-Wei Hsu, Meng-Tse Wu, Shih-Hong Tsai, Tzu-Man Wu, Aleksandra Smolka, Chao-Chun Liang, Hsin-Min Wang, Kuan-Yu Chen, Yu Tsao, Keh-Yih Su
Only a few of them adopt several answer generation modules for providing different mechanisms; however, they either lack an aggregation mechanism to merge the answers from various modules, or are too complicated to be implemented with neural networks.
no code implementations • ROCLING 2021 • Hung-Yun Chiang, Kuan-Yu Chen
Due to the development of deep learning, the natural language processing tasks have made great progresses by leveraging the bidirectional encoder representations from Transformers (BERT).
no code implementations • ROCLING 2021 • Yi-Wei Wang, Wei-Zhe Chang, Bo-Han Fang, Yi-Chia Chen, Wei-Kai Huang, Kuan-Yu Chen
This technical report aims at the ROCLING 2021 Shared Task: Dimensional Sentiment Analysis for Educational Texts.
no code implementations • ROCLING 2021 • Ke-Han Lu, Kuan-Yu Chen
In this paper, we proposed a BERT-based dimensional semantic analyzer, which is designed by incorporating with word-level information.
no code implementations • ROCLING 2022 • Shang-Bao Luo, Cheng-Chung Fan, Kuan-Yu Chen, Yu Tsao, Hsin-Min Wang, Keh-Yih Su
This paper also provides a baseline system and shows its performance on this dataset.
no code implementations • 18 May 2023 • Chong-En Lin, Kuan-Yu Chen
Non-autoregressive automatic speech recognition (ASR) has become a mainstream of ASR modeling because of its fast decoding speed and satisfactory result.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1