Search Results for author: Xudong Han

Found 20 papers, 12 papers with code

Diverse Adversaries for Mitigating Bias in Training

1 code implementation EACL 2021 Xudong Han, Timothy Baldwin, Trevor Cohn

Adversarial learning can learn fairer and less biased models of language than standard methods.

Learning-based Optoelectronically Innervated Tactile Finger for Rigid-Soft Interactive Grasping

no code implementations29 Jan 2021 Linhan Yang, Xudong Han, Weijie Guo, Fang Wan, Jia Pan, Chaoyang Song

This paper presents a novel design of a soft tactile finger with omni-directional adaptation using multi-channel optical fibers for rigid-soft interactive grasping.

Robotics

Balancing out Bias: Achieving Fairness Through Balanced Training

no code implementations16 Sep 2021 Xudong Han, Timothy Baldwin, Trevor Cohn

Group bias in natural language processing tasks manifests as disparities in system error rates across texts authorized by different demographic groups, typically disadvantaging minority groups.

Fairness

Contrastive Learning for Fair Representations

no code implementations22 Sep 2021 Aili Shen, Xudong Han, Trevor Cohn, Timothy Baldwin, Lea Frermann

Trained classification models can unintentionally lead to biased representations and predictions, which can reinforce societal preconceptions and stereotypes.

Attribute Contrastive Learning

Towards Equal Opportunity Fairness through Adversarial Learning

1 code implementation12 Mar 2022 Xudong Han, Timothy Baldwin, Trevor Cohn

Adversarial training is a common approach for bias mitigation in natural language processing.

Fairness

Fair Enough: Standardizing Evaluation and Model Selection for Fairness Research in NLP

1 code implementation11 Feb 2023 Xudong Han, Timothy Baldwin, Trevor Cohn

Modern NLP systems exhibit a range of biases, which a growing literature on model debiasing attempts to correct.

Fairness Model Selection +1

Proprioceptive Learning with Soft Polyhedral Networks

no code implementations16 Aug 2023 Xiaobo Liu, Xudong Han, Wei Hong, Fang Wan, Chaoyang Song

Proprioception is the "sixth sense" that detects limb postures with motor neurons.

Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs

1 code implementation25 Aug 2023 Yuxia Wang, Haonan Li, Xudong Han, Preslav Nakov, Timothy Baldwin

With the rapid evolution of large language models (LLMs), new and hard-to-predict harmful capabilities are emerging.

Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language Models as Agents

1 code implementation18 Feb 2024 Renxi Wang, Haonan Li, Xudong Han, Yixuan Zhang, Timothy Baldwin

However, LLMs are optimized for language generation instead of tool use during training or alignment, limiting their effectiveness as agents.

Mathematical Reasoning Multi-hop Question Answering +2

A Chinese Dataset for Evaluating the Safeguards in Large Language Models

no code implementations19 Feb 2024 Yuxia Wang, Zenan Zhai, Haonan Li, Xudong Han, Lizhi Lin, Zhenxuan Zhang, Jingru Zhao, Preslav Nakov, Timothy Baldwin

Previous studies have proposed comprehensive taxonomies of the risks posed by LLMs, as well as corresponding prompts that can be used to examine the safety mechanisms of LLMs.

Against The Achilles' Heel: A Survey on Red Teaming for Generative Models

no code implementations31 Mar 2024 Lizhi Lin, Honglin Mu, Zenan Zhai, Minghan Wang, Yuxia Wang, Renxi Wang, Junjie Gao, Yixuan Zhang, Wanxiang Che, Timothy Baldwin, Xudong Han, Haonan Li

Generative models are rapidly gaining popularity and being integrated into everyday applications, raising concerns over their safety issues as various vulnerabilities are exposed.

Cannot find the paper you are looking for? You can Submit a new open access paper.