no code implementations • 19 Oct 2024 • Hanqun Cao, Mutian He, Ning Ma, Chang-Yu Hsieh, Chunbin Gu, Pheng-Ann Heng
DNA-encoded library (DEL) screening has revolutionized the detection of protein-ligand interactions through read counts, enabling rapid exploration of vast chemical spaces.
1 code implementation • 9 Oct 2024 • Mutian He, Philip N. Garner
In a series of empirical studies on language processing, language modeling, and speech processing, we show that CALD can effectively recover the result of the original model, and that the guiding strategy contributes to the result.
no code implementations • 7 Sep 2024 • Chunbin Gu, Mutian He, Hanqun Cao, Guangyong Chen, Chang-Yu Hsieh, Pheng Ann Heng
To mitigate these issues, we propose a Multimodal Pretraining DEL-Fusion model (MPDF) that enhances encoder capabilities through pretraining and integrates compound features across various scales.
no code implementations • 22 May 2023 • Mutian He, Philip N. Garner
Recently, large pretrained language models have demonstrated strong language understanding capabilities.
1 code implementation • 16 May 2023 • Mutian He, Philip N. Garner
Motivated particularly by the task of cross-lingual SLU, we demonstrate that the task of speech translation (ST) is a good means of pretraining speech models for end-to-end SLU on both intra- and cross-lingual scenarios.
1 code implementation • 3 Jun 2022 • Mutian He, Tianqing Fang, Weiqi Wang, Yangqiu Song
Conceptualization, or viewing entities and situations as instances of abstract concepts in mind and making inferences based on that, is a vital component in human intelligence for commonsense reasoning.
1 code implementation • 19 Oct 2021 • Mutian He, Jingzhou Yang, Lei He, Frank K. Soong
End-to-end TTS requires a large amount of speech/text paired data to cover all necessary knowledge, particularly how to pronounce different words in diverse contexts, so that a neural model may learn such knowledge accordingly.
2 code implementations • 5 Mar 2021 • Mutian He, Jingzhou Yang, Lei He, Frank K. Soong
To scale neural speech synthesis to various real-world languages, we present a multilingual end-to-end framework that maps byte inputs to spectrograms, thus allowing arbitrary input scripts.
1 code implementation • 6 Mar 2020 • Mutian He, Yangqiu Song, Kun Xu, Dong Yu
Commonsense knowledge graphs (CKGs) like Atomic and ASER are substantially different from conventional KGs as they consist of much larger number of nodes formed by loosely-structured text, which, though, enables them to handle highly diverse queries in natural language related to commonsense, leads to unique challenges for automatic KG construction methods.
1 code implementation • 25 Dec 2019 • Xin Liu, Haojie Pan, Mutian He, Yangqiu Song, Xin Jiang, Lifeng Shang
In this paper, we study a new graph learning problem: learning to count subgraph isomorphisms.
1 code implementation • 3 Jun 2019 • Mutian He, Yan Deng, Lei He
In this paper, we propose a novel stepwise monotonic attention method in sequence-to-sequence acoustic modeling to improve the robustness on out-of-domain inputs.