no code implementations • 29 Mar 2025 • Yuelyu Ji, Rui Meng, Zhuochun Li, Daqing He
Multi-hop question answering (QA) requires models to retrieve and reason over multiple pieces of evidence.
no code implementations • 7 Oct 2024 • Yuelyu Ji, Wenhe Ma, Sonish Sivarajkumar, Hang Zhang, Eugene Mathew Sadhu, Zhuochun Li, Xizhi Wu, Shyam Visweswaran, Yanshan Wang
Recent advancements in large language models have demonstrated their potential in numerous medical applications, particularly in automating clinical trial matching for translational research and enhancing medical question answering for clinical decision support.
no code implementations • 7 Oct 2024 • Yuelyu Ji, Zhuochun Li, Rui Meng, Daqing He
Reranking documents based on their relevance to a given query is a critical task in information retrieval.
no code implementations • 4 Oct 2024 • Zhuochun Li, Yuelyu Ji, Rui Meng, Daqing He
While reasoning capabilities typically emerge in large language models (LLMs) with tens of billions of parameters, recent research focuses on improving smaller open-source models through knowledge distillation (KD) from commercial LLMs.
1 code implementation • 21 May 2024 • Yuelyu Ji, Zhuochun Li, Rui Meng, Sonish Sivarajkumar, Yanshan Wang, Zeshui Yu, Hui Ji, Yushui Han, Hanyu Zeng, Daqing He
This paper introduces the RAG-RLRC-LaySum framework, designed to make complex biomedical research understandable to laymen through advanced Natural Language Processing (NLP) techniques.
no code implementations • 5 Apr 2024 • Zhuochun Li, Bo Xie, Robin Hilsabeck, Alyssa Aguirre, Ning Zou, Zhimeng Luo, Daqing He
Evidence suggests that different prompts lead large language models (LLMs) to generate responses with varying quality.