Search Results for author: Qinyuan Wu

Found 7 papers, 1 papers with code

Revisiting Privacy, Utility, and Efficiency Trade-offs when Fine-Tuning Large Language Models

no code implementations18 Feb 2025 Soumi Das, Camila Kolling, Mohammad Aflah Khan, Mahsa Amani, Bishwamittra Ghosh, Qinyuan Wu, Till Speicher, Krishna P. Gummadi

A number of recent works in privacy research have attempted to mitigate privacy risks posed by memorizing fine-tuning data by using differentially private training methods (e. g., DP), albeit at a significantly higher computational cost (inefficiency).

Computational Efficiency parameter-efficient fine-tuning

Position: Episodic Memory is the Missing Piece for Long-Term LLM Agents

no code implementations10 Feb 2025 Mathis Pink, Qinyuan Wu, Vy Ai Vo, Javier Turek, Jianing Mu, Alexander Huth, Mariya Toneva

As Large Language Models (LLMs) evolve from text-completion tools into fully fledged agents operating in dynamic environments, they must address the challenge of continually learning and retaining long-term knowledge.

Position

Assessing Episodic Memory in LLMs with Sequence Order Recall Tasks

no code implementations10 Oct 2024 Mathis Pink, Vy A. Vo, Qinyuan Wu, Jianing Mu, Javier S. Turek, Uri Hasson, Kenneth A. Norman, Sebastian Michelmann, Alexander Huth, Mariya Toneva

To address the gap in evaluating memory in LLMs, we introduce Sequence Order Recall Tasks (SORT), which we adapt from tasks used to study episodic memory in cognitive psychology.

Understanding Memorisation in LLMs: Dynamics, Influencing Factors, and Implications

no code implementations27 Jul 2024 Till Speicher, Mohammad Aflah Khan, Qinyuan Wu, Vedant Nanda, Soumi Das, Bishwamittra Ghosh, Krishna P. Gummadi, Evimaria Terzi

Understanding whether and to what extent large language models (LLMs) have memorised training data has important implications for the reliability of their output and the privacy of their training data.

In-Context Learning

Towards Reliable Latent Knowledge Estimation in LLMs: Zero-Prompt Many-Shot Based Factual Knowledge Extraction

1 code implementation19 Apr 2024 Qinyuan Wu, Mohammad Aflah Khan, Soumi Das, Vedant Nanda, Bishwamittra Ghosh, Camila Kolling, Till Speicher, Laurent Bindschaedler, Krishna P. Gummadi, Evimaria Terzi

Our knowledge estimator is both conceptually simpler (i. e., doesn't depend on meta-linguistic judgments of LLMs) and easier to apply (i. e., is not LLM-specific), and we demonstrate that it can surface more of the latent knowledge embedded in LLMs.

In-Context Learning Prompt Engineering

A Negation Quantum Decision Model to Predict the Interference Effect in Categorization

no code implementations19 Apr 2021 Qinyuan Wu, Yong Deng

Categorization is a significant task in decision-making, which is a key part of human behavior.

Decision Making Negation

Exponential Negation of a Probability Distribution

no code implementations22 Oct 2020 Qinyuan Wu, Yong Deng, Neal Xiong

Some basic properties of the proposed negation is investigated, we find that the fix point is the uniform probability distribution.

Negation

Cannot find the paper you are looking for? You can Submit a new open access paper.