Search Results for author: Ke-Han Lu

Found 7 papers, 3 papers with code

ntust-nlp-2 at ROCLING-2021 Shared Task: BERT-based semantic analyzer with word-level information

no code implementations ROCLING 2021 Ke-Han Lu, Kuan-Yu Chen

In this paper, we proposed a BERT-based dimensional semantic analyzer, which is designed by incorporating with word-level information.

Sentiment Analysis

Investigating Zero-Shot Generalizability on Mandarin-English Code-Switched ASR and Speech-to-text Translation of Recent Foundation Models with Self-Supervision and Weak Supervision

no code implementations30 Dec 2023 Chih-Kai Yang, Kuan-Po Huang, Ke-Han Lu, Chun-Yi Kuan, Chi-Yuan Hsiao, Hung-Yi Lee

This work evaluated several cutting-edge large-scale foundation models based on self-supervision or weak supervision, including SeamlessM4T, SeamlessM4T v2, and Whisper-large-v3, on three code-switched corpora.

Speech-to-Text Translation

Dynamic-SUPERB: Towards A Dynamic, Collaborative, and Comprehensive Instruction-Tuning Benchmark for Speech

1 code implementation18 Sep 2023 Chien-yu Huang, Ke-Han Lu, Shih-Heng Wang, Chi-Yuan Hsiao, Chun-Yi Kuan, Haibin Wu, Siddhant Arora, Kai-Wei Chang, Jiatong Shi, Yifan Peng, Roshan Sharma, Shinji Watanabe, Bhiksha Ramakrishnan, Shady Shehata, Hung-Yi Lee

To achieve comprehensive coverage of diverse speech tasks and harness instruction tuning, we invite the community to collaborate and contribute, facilitating the dynamic growth of the benchmark.

HypR: A comprehensive study for ASR hypothesis revising with a reference corpus

1 code implementation18 Sep 2023 Yi-Wei Wang, Ke-Han Lu, Kuan-Yu Chen

In addition, we implement and compare several classic and representative methods, showing the recent research progress in revising speech recognition results.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

A context-aware knowledge transferring strategy for CTC-based ASR

1 code implementation12 Oct 2022 Ke-Han Lu, Kuan-Yu Chen

Non-autoregressive automatic speech recognition (ASR) modeling has received increasing attention recently because of its fast decoding speed and superior performance.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

A Transformer-based Cross-modal Fusion Model with Adversarial Training for VQA Challenge 2021

no code implementations24 Jun 2021 Ke-Han Lu, Bo-Han Fang, Kuan-Yu Chen

In this paper, inspired by the successes of visionlanguage pre-trained models and the benefits from training with adversarial attacks, we present a novel transformerbased cross-modal fusion modeling by incorporating the both notions for VQA challenge 2021.

Visual Question Answering (VQA)

Cannot find the paper you are looking for? You can Submit a new open access paper.