Search Results for author: Feifan Song

Found 10 papers, 8 papers with code

Omni-MATH: A Universal Olympiad Level Mathematic Benchmark For Large Language Models

1 code implementation10 Oct 2024 Bofei Gao, Feifan Song, Zhe Yang, Zefan Cai, Yibo Miao, Qingxiu Dong, Lei LI, Chenghao Ma, Liang Chen, Runxin Xu, Zhengyang Tang, Benyou Wang, Daoguang Zan, Shanghaoran Quan, Ge Zhang, Lei Sha, Yichang Zhang, Xuancheng Ren, Tianyu Liu, Baobao Chang

However, existing benchmarks like GSM8K or MATH are now being solved with high accuracy (e. g., OpenAI o1 achieves 94. 8% on MATH dataset), indicating their inadequacy for truly challenging these models.

GSM8K Math +1

Learning Spatial Similarity Distribution for Few-shot Object Counting

1 code implementation20 May 2024 Yuanwu Xu, Feifan Song, Haofeng Zhang

To address this issue, we propose a network learning Spatial Similarity Distribution (SSD) for few-shot object counting, which preserves the spatial structure of exemplar features and calculates a 4D similarity pyramid point-to-point between the query features and exemplar features, capturing the complete distribution information for each point in the 4D similarity space.

Object Counting

Scaling Data Diversity for Fine-Tuning Language Models in Human Alignment

1 code implementation17 Mar 2024 Feifan Song, Bowen Yu, Hao Lang, Haiyang Yu, Fei Huang, Houfeng Wang, Yongbin Li

Additionally, the concept of diversity for prompts can be more complex than responses that are typically quantified by single digits.

Data Augmentation Diversity

ICDPO: Effectively Borrowing Alignment Capability of Others via In-context Direct Preference Optimization

1 code implementation14 Feb 2024 Feifan Song, Yuxuan Fan, Xin Zhang, Peiyi Wang, Houfeng Wang

Large Language Models (LLMs) rely on Human Preference Alignment (HPA) to ensure the generation of safe content.

In-Context Learning

Making Large Language Models Better Reasoners with Alignment

no code implementations5 Sep 2023 Peiyi Wang, Lei LI, Liang Chen, Feifan Song, Binghuai Lin, Yunbo Cao, Tianyu Liu, Zhifang Sui

To address this problem, we introduce an \textit{Alignment Fine-Tuning (AFT)} paradigm, which involves three steps: 1) fine-tuning LLMs with COT training data; 2) generating multiple COT responses for each question, and categorizing them into positive and negative ones based on whether they achieve the correct answer; 3) calibrating the scores of positive and negative responses given by LLMs with a novel constraint alignment loss.

Preference Ranking Optimization for Human Alignment

1 code implementation30 Jun 2023 Feifan Song, Bowen Yu, Minghao Li, Haiyang Yu, Fei Huang, Yongbin Li, Houfeng Wang

In this manner, PRO effectively transforms human alignment into aligning the probability ranking of n responses generated by LLM with the preference ranking of humans towards these responses.

Interacting with Non-Cooperative User: A New Paradigm for Proactive Dialogue Policy

no code implementations7 Apr 2022 Wenqiang Lei, Yao Zhang, Feifan Song, Hongru Liang, Jiaxin Mao, Jiancheng Lv, Zhenglu Yang, Tat-Seng Chua

To this end, we contribute to advance the study of the proactive dialogue policy to a more natural and challenging setting, i. e., interacting dynamically with users.

Cannot find the paper you are looking for? You can Submit a new open access paper.