Search Results for author: Xianjun Yang

Found 25 papers, 18 papers with code

Unveiling the Misuse Potential of Base Large Language Models via In-Context Learning

no code implementations16 Apr 2024 Xiao Wang, Tianze Chen, Xianjun Yang, Qi Zhang, Xun Zhao, Dahua Lin

The open-sourcing of large language models (LLMs) accelerates application development, innovation, and scientific progress.

In-Context Learning Instruction Following

Test-Time Backdoor Attacks on Multimodal Large Language Models

1 code implementation13 Feb 2024 Dong Lu, Tianyu Pang, Chao Du, Qian Liu, Xianjun Yang, Min Lin

Backdoor attacks are commonly executed by contaminating training data, such that a trigger can activate predetermined harmful effects during the test phase.

Backdoor Attack

TrustAgent: Towards Safe and Trustworthy LLM-based Agents through Agent Constitution

1 code implementation2 Feb 2024 Wenyue Hua, Xianjun Yang, Zelong Li, Wei Cheng, Yongfeng Zhang

This paper presents an Agent-Constitution-based agent framework, TrustAgent, an initial investigation into improving the safety dimension of trustworthiness in LLM-based agents.

Weak-to-Strong Jailbreaking on Large Language Models

1 code implementation30 Jan 2024 Xuandong Zhao, Xianjun Yang, Tianyu Pang, Chao Du, Lei LI, Yu-Xiang Wang, William Yang Wang

In this paper, we propose the weak-to-strong jailbreaking attack, an efficient method to attack aligned LLMs to produce harmful text.

PLLaMa: An Open-source Large Language Model for Plant Science

1 code implementation3 Jan 2024 Xianjun Yang, Junfeng Gao, Wenxin Xue, Erik Alexandersson

Large Language Models (LLMs) have exhibited remarkable capabilities in understanding and interacting with natural language across various sectors.

Language Modelling Large Language Model

Quokka: An Open-source Large Language Model ChatBot for Material Science

1 code implementation2 Jan 2024 Xianjun Yang, Stephen D. Wilson, Linda Petzold

This paper presents the development of a specialized chatbot for materials science, leveraging the Llama-2 language model, and continuing pre-training on the expansive research articles in the materials science domain from the S2ORC dataset.

Chatbot Language Modelling +1

A Survey on Detection of LLMs-Generated Content

1 code implementation24 Oct 2023 Xianjun Yang, Liangming Pan, Xuandong Zhao, Haifeng Chen, Linda Petzold, William Yang Wang, Wei Cheng

The burgeoning capabilities of advanced large language models (LLMs) such as ChatGPT have led to an increase in synthetic content generation with implications across a variety of sectors, including media, cybersecurity, public discourse, and education.

AlpaCare:Instruction-tuned Large Language Models for Medical Application

1 code implementation23 Oct 2023 Xinlu Zhang, Chenxin Tian, Xianjun Yang, Lichang Chen, Zekun Li, Linda Ruth Petzold

Instruction-finetuning (IFT) has become crucial in aligning Large Language Models (LLMs) with diverse human needs and has shown great potential in medical applications.

Instruction Following

Zero-Shot Detection of Machine-Generated Codes

1 code implementation8 Oct 2023 Xianjun Yang, Kexun Zhang, Haifeng Chen, Linda Petzold, William Yang Wang, Wei Cheng

We then modify the previous zero-shot text detection method, DetectGPT (Mitchell et al., 2023) by utilizing a surrogate white-box model to estimate the probability of the rightmost tokens, allowing us to identify code snippets generated by language models.

Language Modelling Text Detection

Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models

no code implementations4 Oct 2023 Xianjun Yang, Xiao Wang, Qi Zhang, Linda Petzold, William Yang Wang, Xun Zhao, Dahua Lin

This study serves as a clarion call for a collective effort to overhaul and fortify the safety of open-source LLMs against malicious attackers.

Large Language Models Can Be Good Privacy Protection Learners

no code implementations3 Oct 2023 Yijia Xiao, Yiqiao Jin, Yushi Bai, Yue Wu, Xianjun Yang, Xiao Luo, Wenchao Yu, Xujiang Zhao, Yanchi Liu, Haifeng Chen, Wei Wang, Wei Cheng

To address this challenge, we introduce Privacy Protection Language Models (PPLM), a novel paradigm for fine-tuning LLMs that effectively injects domain-specific knowledge while safeguarding data privacy.

DNA-GPT: Divergent N-Gram Analysis for Training-Free Detection of GPT-Generated Text

1 code implementation27 May 2023 Xianjun Yang, Wei Cheng, Yue Wu, Linda Petzold, William Yang Wang, Haifeng Chen

However, this progress also presents a significant challenge in detecting the origin of a given text, and current research on detection methods lags behind the rapid evolution of LLMs.

Enhancing Small Medical Learners with Privacy-preserving Contextual Prompting

1 code implementation22 May 2023 Xinlu Zhang, Shiyang Li, Xianjun Yang, Chenxin Tian, Yao Qin, Linda Ruth Petzold

Large language models (LLMs) demonstrate remarkable medical expertise, but data privacy concerns impede their direct use in healthcare environments.

Decision Making Privacy Preserving

LLMScore: Unveiling the Power of Large Language Models in Text-to-Image Synthesis Evaluation

1 code implementation NeurIPS 2023 Yujie Lu, Xianjun Yang, Xiujun Li, Xin Eric Wang, William Yang Wang

Existing automatic evaluation on text-to-image synthesis can only provide an image-text matching score, without considering the object-level compositionality, which results in poor correlation with human judgments.

Attribute Image Generation +2

Dynamic Prompting: A Unified Framework for Prompt Tuning

1 code implementation6 Mar 2023 Xianjun Yang, Wei Cheng, Xujiang Zhao, Wenchao Yu, Linda Petzold, Haifeng Chen

Experimental results underscore the significant performance improvement achieved by dynamic prompt tuning across a wide range of tasks, including NLP tasks, vision recognition tasks, and vision-language tasks.

Position

Exploring the Limits of ChatGPT for Query or Aspect-based Text Summarization

no code implementations16 Feb 2023 Xianjun Yang, Yan Li, Xinlu Zhang, Haifeng Chen, Wei Cheng

Text summarization has been a crucial problem in natural language processing (NLP) for several decades.

Abstractive Text Summarization

MatKB: Semantic Search for Polycrystalline Materials Synthesis Procedures

1 code implementation11 Feb 2023 Xianjun Yang, Stephen Wilson, Linda Petzold

In this paper, we present a novel approach to knowledge extraction and retrieval using Natural Language Processing (NLP) techniques for material science.

Document Classification Retrieval

OASum: Large-Scale Open Domain Aspect-based Summarization

1 code implementation19 Dec 2022 Xianjun Yang, Kaiqiang Song, Sangwoo Cho, Xiaoyang Wang, Xiaoman Pan, Linda Petzold, Dong Yu

Specifically, zero/few-shot and fine-tuning results show that the model pre-trained on our corpus demonstrates a strong aspect or query-focused generation ability compared with the backbone model.

PcMSP: A Dataset for Scientific Action Graphs Extraction from Polycrystalline Materials Synthesis Procedure Text

1 code implementation22 Oct 2022 Xianjun Yang, Ya Zhuo, Julia Zuo, Xinlu Zhang, Stephen Wilson, Linda Petzold

Scientific action graphs extraction from materials synthesis procedures is important for reproducible research, machine automation, and material prediction.

Named Entity Recognition Named Entity Recognition (NER) +3

Few-Shot Document-Level Event Argument Extraction

1 code implementation6 Sep 2022 Xianjun Yang, Yujie Lu, Linda Petzold

To fill this gap, we present FewDocAE, a Few-Shot Document-Level Event Argument Extraction benchmark, based on the existing document-level event extraction dataset.

Document-level Event Extraction Event Argument Extraction +2

Cannot find the paper you are looking for? You can Submit a new open access paper.