no code implementations • 24 Oct 2024 • Kun Li, Tianhua Zhang, Xixin Wu, Hongyin Luo, James Glass, Helen Meng
We argue that this concept can serve as a principle for making faithful and sound reasoning for KGQA.
1 code implementation • 2 Oct 2024 • Zhenting Qi, Hongyin Luo, Xuliang Huang, Zhuokai Zhao, Yibo Jiang, Xiangjun Fan, Himabindu Lakkaraju, James Glass
Scylla disentangles generalization from memorization via assessing model performance on both in-distribution (ID) and out-of-distribution (OOD) data through 20 tasks across 5 levels of complexity.
no code implementations • 1 Oct 2024 • Hongyin Luo, Wei Sun
The new algorithm costs significantly less computation resource than 8-bit floating point multiplication but achieves higher precision.
no code implementations • 7 Jul 2024 • Jiaxin Ge, Xueying Jia, Vijay Viswanathan, Hongyin Luo, Graham Neubig
One of the most reliable ways to create deployable models for specialized tasks is to obtain an adequate amount of high-quality task-specific data.
no code implementations • 17 Jun 2024 • Junmo Kang, Leonid Karlinsky, Hongyin Luo, Zhen Wang, Jacob Hansen, James Glass, David Cox, Rameswar Panda, Rogerio Feris, Alan Ritter
Our findings highlight the critical role of modularity, the applicability of Self-MoE to multiple base LLMs, and the potential of self-improvement in achieving efficient, scalable, and adaptable systems.
no code implementations • 16 Jun 2024 • Tianhua Zhang, Kun Li, Hongyin Luo, Xixin Wu, James Glass, Helen Meng
A novel approach is then proposed to assess retriever's preference for these candidates by the probability of answers conditioned on the conversational query by marginalizing the Top-$K$ passages.
1 code implementation • 27 May 2024 • Philip Schroeder, Nathaniel Morgan, Hongyin Luo, James Glass
We apply THREAD in the settings of LLM task solving and question answering, where the dynamic threading allows the model to recursively decompose the given task or question into progressively simpler sub-problems that can be solved by separate child threads.
no code implementations • 31 Mar 2024 • Chi-Min Chan, Chunpu Xu, Ruibin Yuan, Hongyin Luo, Wei Xue, Yike Guo, Jie Fu
To this end, we propose learning to Refine Query for Retrieval Augmented Generation (RQ-RAG) in this paper, endeavoring to enhance the model by equipping it with capabilities for explicit rewriting, decomposition, and disambiguation.
2 code implementations • 1 Mar 2024 • Zhaorun Chen, Zhuokai Zhao, Hongyin Luo, Huaxiu Yao, Bo Li, Jiawei Zhou
While large vision-language models (LVLMs) have demonstrated impressive capabilities in interpreting multi-modal contexts, they invariably suffer from object hallucinations (OH).
no code implementations • 29 Sep 2023 • Junmo Kang, Hongyin Luo, Yada Zhu, Jacob Hansen, James Glass, David Cox, Alan Ritter, Rogerio Feris, Leonid Karlinsky
Recent works have demonstrated the effectiveness of self-alignment in which a large language model is aligned to follow general instructions using instructional data generated from the model itself starting from a handful of human-written seeds.
1 code implementation • 25 Sep 2023 • Yuan Gong, Alexander H. Liu, Hongyin Luo, Leonid Karlinsky, James Glass
Humans are surrounded by audio signals that include both speech and non-speech sounds.
1 code implementation • 19 Sep 2023 • Tianhua Zhang, Jiaxin Ge, Hongyin Luo, Yung-Sung Chuang, Mingye Gao, Yuan Gong, Xixin Wu, Yoon Kim, Helen Meng, James Glass
How can we perform computations over natural language representations to solve tasks that require symbolic and numeric reasoning?
3 code implementations • 7 Sep 2023 • Yung-Sung Chuang, Yujia Xie, Hongyin Luo, Yoon Kim, James Glass, Pengcheng He
Despite their impressive capabilities, large language models (LLMs) are prone to hallucinations, i. e., generating content that deviates from facts seen during pretraining.
1 code implementation • 26 May 2023 • Jiaxin Ge, Hongyin Luo, Yoon Kim, James Glass
Experiments on binary and multi-class classification tasks show that SimPLE leads to more robust self-training results, indicating that the self-trained entailment models are more efficient and trustworthy than large language models on language understanding tasks.
Multi-class Classification Natural Language Understanding +1
no code implementations • 24 May 2023 • Hongyin Luo, Yung-Sung Chuang, Yuan Gong, Tianhua Zhang, Yoon Kim, Xixin Wu, Danny Fox, Helen Meng, James Glass
Large language models (LLMs) have been significantly improved by instruction fine-tuning, but still lack transparency and the ability to utilize up-to-date knowledge and information.
2 code implementations • 18 May 2023 • Yuan Gong, Hongyin Luo, Alexander H. Liu, Leonid Karlinsky, James Glass
On the other hand, modern large language models (LLMs) exhibit emerging reasoning ability but they lack audio perception capabilities.
Ranked #3 on Music Question Answering on MusicQA (using extra training data)
no code implementations • 16 Apr 2023 • Jiaxin Ge, Hongyin Luo, Siyuan Qian, Yulu Gan, Jie Fu, Shanghang Zhang
Chain of Thought is a simple and effective approximation to human reasoning process and has been proven useful for natural language processing (NLP) tasks.
1 code implementation • 7 Apr 2023 • Tianhua Zhang, Hongyin Luo, Yung-Sung Chuang, Wei Fang, Luc Gaitskell, Thomas Hartvigsen, Xixin Wu, Danny Fox, Helen Meng, James Glass
Despite recent concerns about undesirable behaviors generated by large language models (LLMs), including non-factual, biased, and hateful language, we find LLMs are inherent multi-task language checkers based on their latent representations of natural and social knowledge.
1 code implementation • 10 Mar 2023 • Hongyin Luo, James Glass
Due to their similarity-based learning objectives, pretrained sentence encoders often internalize stereotypical assumptions that reflect the social biases that exist within their training corpora.
1 code implementation • NAACL 2022 • Yung-Sung Chuang, Rumen Dangovski, Hongyin Luo, Yang Zhang, Shiyu Chang, Marin Soljačić, Shang-Wen Li, Wen-tau Yih, Yoon Kim, James Glass
We propose DiffCSE, an unsupervised contrastive learning framework for learning sentence embeddings.
Ranked #13 on Semantic Textual Similarity on STS16
1 code implementation • ACL (WOAH) 2021 • Yung-Sung Chuang, Mingye Gao, Hongyin Luo, James Glass, Hung-Yi Lee, Yun-Nung Chen, Shang-Wen Li
Automatic detection of toxic language plays an essential role in protecting social media users, especially minority groups, from verbal abuse.
no code implementations • 6 Jun 2021 • Hongyin Luo, Shuyan Dong, Yung-Sung Chuang, Shang-Wen Li
Neural network pretraining is gaining attention due to its outstanding performance in natural language processing applications.
1 code implementation • NAACL 2022 • Hongyin Luo, Shang-Wen Li, Mingye Gao, Seunghak Yu, James Glass
Pretrained language models have significantly improved the performance of downstream language understanding tasks, including extractive question answering, by providing high-quality contextualized word embeddings.
Ranked #1 on Question Answering on MRQA out-of-domain
Extractive Question-Answering Machine Reading Comprehension +6
2 code implementations • EMNLP (ClinicalNLP) 2020 • Hongyin Luo, Shang-Wen Li, James Glass
Given a set of explicit symptoms provided by the patient to initiate a dialog for diagnosing, the system is trained to collect implicit symptoms by asking questions, in order to collect more information for making an accurate diagnosis.
no code implementations • 19 May 2020 • Hongyin Luo, Shang-Wen Li, James Glass
Experiments showed that the ProtoQN significantly outperformed the baseline DQN model in both supervised and few-shot learning scenarios, and achieves state-of-the-art few-shot learning performances.
1 code implementation • ACL 2019 • Hongyin Luo, Lan Jiang, Yonatan Belinkov, James Glass
In this work, we propose a method that improves language modeling by learning to align the given context and the following phrase.
Ranked #24 on Language Modelling on WikiText-103
no code implementations • ICLR 2019 • Hongyin Luo, Yichen Li, Jie Fu, James Glass
Recently, there have been some attempts to use non-recurrent neural models for language modeling.
1 code implementation • EMNLP 2018 • Hongyin Luo, Jim Glass
In this work, we present a word embedding model that learns cross-sentence dependency for improving end-to-end co-reference resolution (E2E-CR).
Ranked #22 on Coreference Resolution on OntoNotes
2 code implementations • 23 Feb 2017 • Hongyin Luo, Jie Fu, James Glass
However, it has been argued that this is not biologically plausible because back-propagating error signals with the exact incoming weights are not considered possible in biological neural systems.
1 code implementation • 5 Jan 2016 • Jie Fu, Hongyin Luo, Jiashi Feng, Kian Hsiang Low, Tat-Seng Chua
The performance of deep neural networks is well-known to be sensitive to the setting of their hyperparameters.