Search Results for author: Yi Qi

Found 9 papers, 4 papers with code

Safeguarding Large Language Models: A Survey

no code implementations3 Jun 2024 Yi Dong, Ronghui Mu, Yanghao Zhang, Siqi Sun, Tianle Zhang, Changshun Wu, Gaojie Jin, Yi Qi, Jinwei Hu, Jie Meng, Saddek Bensalem, Xiaowei Huang

In the burgeoning field of Large Language Models (LLMs), developing a robust safety mechanism, colloquially known as "safeguards" or "guardrails", has become imperative to ensure the ethical use of LLMs within prescribed boundaries.

Fairness

Direct Training Needs Regularisation: Anytime Optimal Inference Spiking Neural Network

1 code implementation15 Apr 2024 Dengyu Wu, Yi Qi, Kaiwen Cai, Gaojie Jin, Xinping Yi, Xiaowei Huang

Notably, with STR and cutoff, SNN achieves 2. 14 to 2. 89 faster in inference compared to the pre-configured timestep with near-zero accuracy drop of 0. 50% to 0. 64% over the event-based datasets.

Building Guardrails for Large Language Models

no code implementations2 Feb 2024 Yi Dong, Ronghui Mu, Gaojie Jin, Yi Qi, Jinwei Hu, Xingyu Zhao, Jie Meng, Wenjie Ruan, Xiaowei Huang

As Large Language Models (LLMs) become more integrated into our daily lives, it is crucial to identify and mitigate their risks, especially when the risks can have profound impacts on human users and societies.

A Survey of Safety and Trustworthiness of Large Language Models through the Lens of Verification and Validation

no code implementations19 May 2023 Xiaowei Huang, Wenjie Ruan, Wei Huang, Gaojie Jin, Yi Dong, Changshun Wu, Saddek Bensalem, Ronghui Mu, Yi Qi, Xingyu Zhao, Kaiwen Cai, Yanghao Zhang, Sihao Wu, Peipei Xu, Dengyu Wu, Andre Freitas, Mustafa A. Mustafa

Large Language Models (LLMs) have exploded a new heatwave of AI for their ability to engage end-users in human-level conversations with detailed and articulate answers across many knowledge domains.

Continual Learning for CTR Prediction: A Hybrid Approach

no code implementations18 Jan 2022 Ke Hu, Yi Qi, Jianqiang Huang, Jia Cheng, Jun Lei

To address this problem, we formulate CTR prediction as a continual learning task and propose COLF, a hybrid COntinual Learning Framework for CTR prediction, which has a memory-based modular architecture that is designed to adapt, learn and give predictions continuously when faced with non-stationary drifting click data streams.

Click-Through Rate Prediction Continual Learning

Deep Position-wise Interaction Network for CTR Prediction

1 code implementation10 Jun 2021 Jianqiang Huang, Ke Hu, Qingtao Tang, Mingjian Chen, Yi Qi, Jia Cheng, Jun Lei

Click-through rate (CTR) prediction plays an important role in online advertising and recommender systems.

Click-Through Rate Prediction Position +1

A Deep Spatio-Temporal Fuzzy Neural Network for Passenger Demand Prediction

no code implementations13 May 2019 Xiaoyuan Liang, Guiling Wang, Martin Renqiang Min, Yi Qi, Zhu Han

In spite of its importance, passenger demand prediction is a highly challenging problem, because the demand is simultaneously influenced by the complex interactions among many spatial and temporal factors and other external factors such as weather.

Bandit Learning with Implicit Feedback

1 code implementation NeurIPS 2018 Yi Qi, Qingyun Wu, Hongning Wang, Jie Tang, Maosong Sun

Implicit feedback, such as user clicks, although abundant in online information service systems, does not provide substantial evidence on users' evaluation of system's output.

Bayesian Inference Thompson Sampling

Cannot find the paper you are looking for? You can Submit a new open access paper.