Search Results for author: Zhiyuan Peng

Found 22 papers, 10 papers with code

ScopeQA: A Framework for Generating Out-of-Scope Questions for RAG

no code implementations18 Oct 2024 Zhiyuan Peng, Jinming Nian, Alexandre Evfimievski, Yi Fang

Conversational AI agents use Retrieval Augmented Generation (RAG) to provide verifiable document-grounded responses to user inquiries.

Hallucination Natural Questions +4

Evaluating and Enhancing Large Language Models for Novelty Assessment in Scholarly Publications

1 code implementation25 Sep 2024 Ethan Lin, Zhiyuan Peng, Yi Fang

Recent studies have evaluated the creativity/novelty of large language models (LLMs) primarily from a semantic perspective, using benchmarks from cognitive science.

RAG Retrieval

Language-Queried Target Sound Extraction Without Parallel Training Data

no code implementations14 Sep 2024 Hao Ma, Zhiyuan Peng, Xu Li, Yukai Li, Mingjie Shao, Qiuqiang Kong, Ju Liu

In a vanilla parallel-data-free training stage, target audio is encoded using the pre-trained CLAP audio encoder to form a condition embedding, while during testing, user language queries are encoded by CLAP text encoder as the condition embedding.

Language Modelling Large Language Model +2

W-RAG: Weakly Supervised Dense Retrieval in RAG for Open-domain Question Answering

1 code implementation15 Aug 2024 Jinming Nian, Zhiyuan Peng, Qifan Wang, Yi Fang

In knowledge-intensive tasks such as open-domain question answering (OpenQA), Large Language Models (LLMs) often struggle to generate factual answers relying solely on their internal (parametric) knowledge.

Open-Domain Question Answering RAG +1

Dy-mer: An Explainable DNA Sequence Representation Scheme using Sparse Recovery

no code implementations6 Jul 2024 Zhiyuan Peng, Yuanbo Tang, Yang Li

DNA sequences encode vital genetic and biological information, yet these unfixed-length sequences cannot serve as the input of common data mining algorithms.

Digital Twin-Assisted Data-Driven Optimization for Reliable Edge Caching in Wireless Networks

no code implementations29 Jun 2024 Zifan Zhang, Yuchen Liu, Zhiyuan Peng, Mingzhe Chen, Dongkuan Xu, Shuguang Cui

To bridge this gap, we introduce a novel digital twin-assisted optimization framework, called D-REC, which integrates reinforcement learning (RL) with diverse intervention modules to ensure reliable caching in nextG wireless networks.

Reinforcement Learning (RL)

Passage-specific Prompt Tuning for Passage Reranking in Question Answering with Large Language Models

1 code implementation31 May 2024 Xuyang Wu, Zhiyuan Peng, Krishna Sravanthi Rajanala Sai, Hsin-Tai Wu, Yi Fang

In this paper, we propose passage-specific prompt tuning for reranking in open-domain question answering (PSPT): a parameter-efficient method that fine-tunes learnable passage-specific soft prompts, incorporating passage-specific knowledge from a limited set of question-passage relevance pairs.

Open-Domain Question Answering Passage Reranking +1

Q-PEFT: Query-dependent Parameter Efficient Fine-tuning for Text Reranking with Large Language Models

no code implementations6 Apr 2024 Zhiyuan Peng, Xuyang Wu, Qifan Wang, Sravanthi Rajanala, Yi Fang

Parameter Efficient Fine-Tuning (PEFT) methods have been extensively utilized in Large Language Models (LLMs) to improve the down-streaming tasks without the cost of fine-tuing the whole LLMs.

parameter-efficient fine-tuning Text Reranking

ToolNet: Connecting Large Language Models with Massive Tools via Tool Graph

no code implementations29 Feb 2024 Xukun Liu, Zhiyuan Peng, Xiaoyuan Yi, Xing Xie, Lirong Xiang, Yuchen Liu, Dongkuan Xu

While achieving remarkable progress in a broad range of tasks, large language models (LLMs) remain significantly limited in properly using massive external tools.

In-Context Learning

FNSPID: A Comprehensive Financial News Dataset in Time Series

1 code implementation9 Feb 2024 Zihan Dong, Xinyu Fan, Zhiyuan Peng

Financial market predictions utilize historical data to anticipate future stock prices and market trends.

Financial Analysis Time Series

Explainable Trajectory Representation through Dictionary Learning

no code implementations13 Dec 2023 Yuanbo Tang, Zhiyuan Peng, Yang Li

A hierarchical dictionary learning scheme is also proposed to ensure the algorithm's scalability on large networks, leading to a multi-scale trajectory representation.

Data Compression Dictionary Learning +1

Extending Whisper with prompt tuning to target-speaker ASR

1 code implementation13 Dec 2023 Hao Ma, Zhiyuan Peng, Mingjie Shao, Jing Li, Ju Liu

Target-speaker automatic speech recognition (ASR) aims to transcribe the desired speech of a target speaker from multi-talker overlapped utterances.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

CoMFLP: Correlation Measure based Fast Search on ASR Layer Pruning

1 code implementation21 Sep 2023 Wei Liu, Zhiyuan Peng, Tan Lee

The search process is carried out in two steps: (1) coarse search: to determine top $K$ candidates by pruning the most redundant layers based on the correlation matrix; (2) fine search: to select the best pruning proposal among $K$ candidates using a task-specific evaluation metric.

speech-recognition Speech Recognition

Gentopia: A Collaborative Platform for Tool-Augmented LLMs

1 code implementation8 Aug 2023 Binfeng Xu, Xukun Liu, Hua Shen, Zeyu Han, Yuhan Li, Murong Yue, Zhiyuan Peng, Yuchen Liu, Ziyu Yao, Dongkuan Xu

We present gentopia, an ALM framework enabling flexible customization of agents through simple configurations, seamlessly integrating various language models, task formats, prompting modules, and plugins into a unified paradigm.

Soft Prompt Tuning for Augmenting Dense Retrieval with Large Language Models

1 code implementation17 Jul 2023 Zhiyuan Peng, Xuyang Wu, Qifan Wang, Yi Fang

We design a filter to select high-quality example document-query pairs in the prompt to further improve the quality of weak tagged queries.

Retrieval TAG +1

ReWOO: Decoupling Reasoning from Observations for Efficient Augmented Language Models

2 code implementations23 May 2023 Binfeng Xu, Zhiyuan Peng, Bowen Lei, Subhabrata Mukherjee, Yuchen Liu, Dongkuan Xu

Augmented Language Models (ALMs) blend the reasoning capabilities of Large Language Models (LLMs) with tools that allow for knowledge retrieval and action execution.

Retrieval

Mixture factorized auto-encoder for unsupervised hierarchical deep factorization of speech signal

no code implementations30 Oct 2019 Zhiyuan Peng, Siyuan Feng, Tan Lee

The USM experiments on ZeroSpeech 2017 dataset verify that the frame tokenizer is able to capture linguistic content and the utterance embedder can acquire speaker-related information.

Clustering Decoder +1

Combining Adversarial Training and Disentangled Speech Representation for Robust Zero-Resource Subword Modeling

no code implementations17 Jun 2019 Siyuan Feng, Tan Lee, Zhiyuan Peng

Experimental results on ZeroSpeech 2017 show that both approaches are effective while the latter is more prominent, and that their combination brings further marginal improvement in across-speaker condition.

Representation Learning Speech Representation Learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.