Search Results for author: Shichao Song

Found 12 papers, 10 papers with code

TurtleBench: Evaluating Top Language Models via Real-World Yes/No Puzzles

1 code implementation7 Oct 2024 Qingchen Yu, Shichao Song, Ke Fang, Yunfeng Shi, Zifan Zheng, Hanyu Wang, Simin Niu, Zhiyu Li

This approach allows for the relatively dynamic generation of evaluation datasets, mitigating the risk of model cheating while aligning assessments more closely with genuine user needs for reasoning capabilities, thus enhancing the reliability of evaluations.

Logical Reasoning

Attention Heads of Large Language Models: A Survey

1 code implementation5 Sep 2024 Zifan Zheng, Yezhaohui Wang, Yuxin Huang, Shichao Song, MingChuan Yang, Bo Tang, Feiyu Xiong, Zhiyu Li

Our survey aims to shed light on the internal reasoning processes of LLMs by concentrating on the underlying mechanisms of attention heads.

Survey

Controllable Text Generation for Large Language Models: A Survey

1 code implementation22 Aug 2024 Xun Liang, Hanyu Wang, Yezhaohui Wang, Shichao Song, Jiawei Yang, Simin Niu, Jie Hu, Dan Liu, Shunyu Yao, Feiyu Xiong, Zhiyu Li

This paper systematically reviews the latest advancements in CTG for LLMs, offering a comprehensive definition of its core concepts and clarifying the requirements for control conditions and text quality.

Attribute Prompt Engineering +2

Internal Consistency and Self-Feedback in Large Language Models: A Survey

1 code implementation19 Jul 2024 Xun Liang, Shichao Song, Zifan Zheng, Hanyu Wang, Qingchen Yu, Xunkai Li, Rong-Hua Li, Yi Wang, Zhonghao Wang, Feiyu Xiong, Zhiyu Li

In this paper, we use a unified perspective of internal consistency, offering explanations for reasoning deficiencies and hallucinations.

$\text{Memory}^3$: Language Modeling with Explicit Memory

no code implementations1 Jul 2024 Hongkang Yang, Zehao Lin, Wenjin Wang, Hao Wu, Zhiyu Li, Bo Tang, Wenqiang Wei, Jinbo Wang, Zeyun Tang, Shichao Song, Chenyang Xi, Yu Yu, Kai Chen, Feiyu Xiong, Linpeng Tang, Weinan E

The model is named $\text{Memory}^3$, since explicit memory is the third form of memory in LLMs after implicit memory (model parameters) and working memory (context key-values).

Language Modelling RAG +1

Empowering Large Language Models to Set up a Knowledge Retrieval Indexer via Self-Learning

1 code implementation27 May 2024 Xun Liang, Simin Niu, Zhiyu Li, Sensen Zhang, Shichao Song, Hanyu Wang, Jiawei Yang, Feiyu Xiong, Bo Tang, Chenyang Xi

Retrieval-Augmented Generation (RAG) offers a cost-effective approach to injecting real-time knowledge into large language models (LLMs).

Question Answering RAG +2

xFinder: Robust and Pinpoint Answer Extraction for Large Language Models

1 code implementation20 May 2024 Qingchen Yu, Zifan Zheng, Shichao Song, Zhiyu Li, Feiyu Xiong, Bo Tang, Ding Chen

The continuous advancement of large language models (LLMs) has brought increasing attention to the critical issue of developing fair and reliable methods for evaluating their performance.

Fake Artificial Intelligence Generated Contents (FAIGC): A Survey of Theories, Detection Methods, and Opportunities

no code implementations25 Apr 2024 Xiaomin Yu, Yezhaohui Wang, Yanfang Chen, Zhen Tao, Dinghao Xi, Shichao Song, Simin Niu, Zhiyu Li

In recent years, generative artificial intelligence models, represented by Large Language Models (LLMs) and Diffusion Models (DMs), have revolutionized content production methods.

DeepFake Detection Face Swapping +2

Controlled Text Generation for Large Language Model with Dynamic Attribute Graphs

1 code implementation17 Feb 2024 Xun Liang, Hanyu Wang, Shichao Song, Mengting Hu, Xunzhi Wang, Zhiyu Li, Feiyu Xiong, Bo Tang

In this study, we introduce a pluggable CTG framework for Large Language Models (LLMs) named Dynamic Attribute Graphs-based controlled text generation (DATG).

Attribute Language Modelling +2

Grimoire is All You Need for Enhancing Large Language Models

1 code implementation7 Jan 2024 Ding Chen, Shichao Song, Qingchen Yu, Zhiyu Li, Wenjin Wang, Feiyu Xiong, Bo Tang

In this paper, we propose a method SLEICL that involves learning from examples using strong language models and then summarizing and transferring these learned skills to weak language models for inference and application.

In-Context Learning

UHGEval: Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation

1 code implementation26 Nov 2023 Xun Liang, Shichao Song, Simin Niu, Zhiyu Li, Feiyu Xiong, Bo Tang, Yezhaohui Wang, Dawei He, Peng Cheng, Zhonghao Wang, Haiying Deng

These techniques encompass the use of directed hallucination induction and strategies that deliberately alter authentic text to produce hallucinations.

Benchmarking Hallucination +2

Cannot find the paper you are looking for? You can Submit a new open access paper.