1 code implementation • ACL 2022 • Guanhua Chen, Shuming Ma, Yun Chen, Dongdong Zhang, Jia Pan, Wenping Wang, Furu Wei
When applied to zero-shot cross-lingual abstractive summarization, it produces an average performance gain of 12. 3 ROUGE-L over mBART-ft. We conduct detailed analyses to understand the key ingredients of SixT+, including multilinguality of the auxiliary parallel data, positional disentangled encoder, and the cross-lingual transferability of its encoder.
Abstractive Text Summarization Cross-Lingual Abstractive Summarization +6
no code implementations • 3 Dec 2024 • Zeqing Zhang, Guangze Zheng, Xuebo Ji, Guanqi Chen, Ruixing Jia, Wentao Chen, Guanhua Chen, Liangjun Zhang, Jia Pan
The generalization capability has also been evaluated and a real-world application on the beach is also demonstrated.
no code implementations • 15 Nov 2024 • Yutao Hou, Yajing Luo, Zhiwen Ruan, Hongru Wang, Weifeng Ge, Yun Chen, Guanhua Chen
In this paper, we introduce Compound Question Synthesis (CQ-Syn) to create the Compound-QA benchmark, focusing on compound questions with multiple sub-questions.
no code implementations • 14 Nov 2024 • Jifan Gao, Guanhua Chen
We conduct validations using a real-world dataset, demonstrating that MCCE achieves promising performance compared to state-of-the-art explanation methods in causal concept effect estimation.
no code implementations • 7 Nov 2024 • Yanjun Gao, Skatje Myers, Shan Chen, Dmitriy Dligach, Timothy A Miller, Danielle Bitterman, Guanhua Chen, Anoop Mayampurath, Matthew Churpek, Majid Afshar
Large language models (LLMs) are being explored for diagnostic decision support, yet their ability to estimate pre-test probabilities, vital for clinical decision-making, remains limited.
no code implementations • 24 Oct 2024 • Hengxiang Zhang, Hongfu Gao, Qiang Hu, Guanhua Chen, Lili Yang, BingYi Jing, Hongxin Wei, Bing Wang, Haifeng Bai, Lei Yang
While previous works have introduced several benchmarks to evaluate the safety risk of LLMs, the community still has a limited understanding of current LLMs' capability to recognize illegal and unsafe content in Chinese contexts.
no code implementations • 2 Aug 2024 • He Zhu, Junyou Su, Tianle Lun, Yicheng Tao, Wenjia Zhang, Zipei Fan, Guanhua Chen
Instruction fine-tuning stands as a crucial advancement in leveraging large language models (LLMs) for enhanced task performance.
1 code implementation • 2 Jul 2024 • Yan Yang, Zeguan Xiao, Xin Lu, Hongru Wang, Hailiang Huang, Guanhua Chen, Yun Chen
The widespread applications of large language models (LLMs) have brought about concerns regarding their potential misuse.
1 code implementation • 18 Jun 2024 • Yixia Li, Boya Xiong, Guanhua Chen, Yun Chen
In this work, we propose SeTAR, a novel, training-free OOD detection method that leverages selective low-rank approximation of weight matrices in vision-language and vision-only models.
Out-of-Distribution Detection Out of Distribution (OOD) Detection
1 code implementation • 13 Jun 2024 • Hanqing Wang, Yixia Li, Shuo Wang, Guanhua Chen, Yun Chen
It is observed that the minor matrix corresponds to the noisy or long-tail information, while the principal matrix contains important knowledge.
no code implementations • 5 May 2024 • Guanhua Chen, Yutong Yao, Derek F. Wong, Lidia S. Chao
Multi-intent natural language understanding (NLU) presents a formidable challenge due to the model confusion arising from multiple intents within a single utterance.
Ranked #2 on Slot Filling on MixSNIPS
no code implementations • 19 Apr 2024 • Guanhua Chen, Wenhan Yu, Lei Sha
While Retrieval-Augmented Generation (RAG) plays a crucial role in the application of Large Language Models (LLMs), existing retrieval methods in knowledge-dense domains like law and medicine still suffer from a lack of multi-perspective views, which are essential for improving interpretability and reliability.
1 code implementation • 13 Mar 2024 • Zeguan Xiao, Yan Yang, Guanhua Chen, Yun Chen
Extensive efforts have been made before the public release of Large language models (LLMs) to align their behaviors with human values.
no code implementations • 21 Feb 2024 • Hongru Wang, Boyang Xue, Baohang Zhou, Tianhua Zhang, Cunxiang Wang, Guanhua Chen, Huimin Wang, Kam-Fai Wong
Retrieve-then-read and generate-then-read are two typical solutions to handle unknown and known questions in open-domain question-answering, while the former retrieves necessary external knowledge and the later prompt the large language models to generate internal known knowledge encoded in the parameters.
1 code implementation • 26 Oct 2023 • Hanqing Wang, Yajing Luo, Boya Xiong, Guanhua Chen, Yun Chen
Stylistic headline generation is the task to generate a headline that not only summarizes the content of an article, but also reflects a desired style that attracts users.
no code implementations • 10 Oct 2023 • Guanqi Chen, Lei Yang, Guanhua Chen, Jia Pan
The ability to navigate robots with natural language instructions in an unknown environment is a crucial step for achieving embodied artificial intelligence (AI).
1 code implementation • 2 Oct 2023 • Tianci Xue, Ziqi Wang, Yixia Li, Yun Chen, Guanhua Chen
Instruction tuning enhances the instruction following ability of large language models by finetuning with supervised instruction data.
1 code implementation • ACL 2023 • Guanhua Chen, Lu Hou, Yun Chen, Wenliang Dai, Lifeng Shang, Xin Jiang, Qun Liu, Jia Pan, Wenping Wang
Furthermore, to enhance the token- and sentence-level multilingual representation of the MTE, we propose to train it with machine translation and contrastive learning jointly before the TriKD to provide a better initialization.
1 code implementation • 28 Jan 2023 • Weikang Wang, Guanhua Chen, Hanqing Wang, Yue Han, Yun Chen
In this paper, we investigate whether multilingual sentence Transformer LaBSE is a strong multilingual word aligner.
no code implementations • 8 Jan 2023 • Shuai Wang, ChiYung Yam, Shuguang Chen, Lihong Hu, Liping Li, Faan-Fung Hung, Jiaqi Fan, Chi-Ming Che, Guanhua Chen
Here, we develop a general protocol for accurate predictions of emission wavelength, radiative decay rate constant, and PL quantum yield for phosphorescent Pt(II) emitters based on the combination of first-principles quantum mechanical method, machine learning (ML) and experimental calibration.
1 code implementation • 28 Nov 2022 • Jay Jojo Cheng, Jared D. Huling, Guanhua Chen
Medical treatments tailored to a patient's baseline characteristics hold the potential of improving patient outcomes while reducing negative side effects.
no code implementations • 24 Feb 2022 • Guanhua Chen, Xiaomao Li, Menggang Yu
In this paper, we propose a new method to estimate such an optimal dose interval, named probability dose interval (PDI).
1 code implementation • 16 Oct 2021 • Guanhua Chen, Shuming Ma, Yun Chen, Dongdong Zhang, Jia Pan, Wenping Wang, Furu Wei
When applied to zero-shot cross-lingual abstractive summarization, it produces an average performance gain of 12. 3 ROUGE-L over mBART-ft. We conduct detailed analyses to understand the key ingredients of SixT+, including multilinguality of the auxiliary parallel data, positional disentangled encoder, and the cross-lingual transferability of its encoder.
Abstractive Text Summarization Cross-Lingual Abstractive Summarization +6
no code implementations • 3 May 2021 • Rui Chen, Jared D. Huling, Guanhua Chen, Menggang Yu
The aim of this paper is to develop a weighting framework to mitigate the impact of such misspecification and thus facilitate the generalizability of optimal ITRs from a source population to a target population.
1 code implementation • EMNLP 2021 • Guanhua Chen, Shuming Ma, Yun Chen, Li Dong, Dongdong Zhang, Jia Pan, Wenping Wang, Furu Wei
In this paper, we focus on a zero-shot cross-lingual transfer task in NMT.
1 code implementation • EMNLP 2020 • Yun Chen, Yang Liu, Guanhua Chen, Xin Jiang, Qun Liu
Shift-Att is an interpretation method that induces alignments from the attention weights of Transformer and does not require parameter update or architecture change.
no code implementations • 3 Aug 2018 • Ying-Qi Zhao, Ruoqing Zhu, Guanhua Chen, Yingye Zheng
We propose a new method termed stabilized O-learning for deriving stabilized dynamic treatment regimes, which are sequential decision rules for individual patients that not only adapt over the course of the disease progression but also remain consistent over time in format.
Methodology
no code implementations • 11 Jul 2014 • Qian Liu, Guanhua Chen, Michael R. Kosorok, Eric Bair
This framework can be used to identify biclusters that differ with respect to the means of the features, the variance of the features, or more general differences.