1 code implementation • 2 Aug 2024 • Kunlun Zhu, Yifan Luo, Dingling Xu, Ruobing Wang, Shi Yu, Shuo Wang, Yukun Yan, Zhenghao Liu, Xu Han, Zhiyuan Liu, Maosong Sun
By benchmarking RAG models in vertical domains, RAGEval has the ability to better evaluate the knowledge usage ability of LLMs, which avoids the confusion regarding the source of knowledge in answering question in existing QA datasets--whether it comes from parameterized memory or retrieval.
2 code implementations • 12 Feb 2024 • Yifan Zhang, Yifan Luo, Yang Yuan, Andrew Chi-Chih Yao
Our method showcases a 2 times increase in pretraining token efficiency compared to state-of-the-art baselines, underscoring the potential of our approach in enhancing models' mathematical reasoning capabilities.
1 code implementation • 17 Jan 2024 • Haoxiong Liu, Yifan Zhang, Yifan Luo, Andrew Chi-Chih Yao
The MMIQC dataset is available on the HuggingFace hub at https://huggingface. co/datasets/Vivacem/MMIQC.
Ranked #58 on Math Word Problem Solving on MATH (using extra training data)
no code implementations • 22 Oct 2023 • Yifan Luo, Yiming Tang, Chengfeng Shen, Zhennan Zhou, Bin Dong
In this paper, we propose an optimal control framework tailored for multi-round interactions with LLMs.
1 code implementation • NeurIPS 2023 • ZHIXUN LI, Xin Sun, Yifan Luo, Yanqiao Zhu, Dingshuo Chen, Yingtao Luo, Xiangxin Zhou, Qiang Liu, Shu Wu, Liang Wang, Jeffrey Xu Yu
To fill this gap, we systematically analyze the performance of GSL in different scenarios and develop a comprehensive Graph Structure Learning Benchmark (GSLB) curated from 20 diverse graph datasets and 16 distinct GSL algorithms.
1 code implementation • 5 Jul 2023 • Shengding Hu, Yifan Luo, Huadong Wang, Xingyi Cheng, Zhiyuan Liu, Maosong Sun
In this paper, we find that the PLMs already possess the knowledge required to rebut such questions, and the key is how to activate the knowledge.
no code implementations • 25 May 2023 • Yifan Luo, Bin Dong
In this paper, we studied two identically-trained neural networks (i. e. networks with the same architecture, trained on the same dataset using the same algorithm, but with different initialization) and found that their outputs discrepancy on the training dataset exhibits a "double descent" phenomenon.
2 code implementations • 21 Jan 2021 • Yunpeng Gong, Zhiyong Zeng, Liwen Chen, Yifan Luo, Bin Weng, Feng Ye
This method can not only improve the accuracy of the model, but also help the model defend against adversarial examples; 2) Multi-Modal Defense, it integrates three homogeneous modal images of visible, grayscale and sketch, and further strengthens the defense ability of the model.
Ranked #19 on Person Re-Identification on Market-1501-C
no code implementations • 18 Oct 2020 • Yifan Luo, Jindan Xu, Wei Xu, Kezhi Wang
Federated learning (FL) in a bandwidth-limited network with energy-limited user equipments (UEs) is under-explored.