1 code implementation • 5 Sep 2024 • Zifan Zheng, Yezhaohui Wang, Yuxin Huang, Shichao Song, Bo Tang, Feiyu Xiong, Zhiyu Li
As a result, many researchers have begun exploring the potential internal mechanisms of LLMs, aiming to identify the essence of their reasoning bottlenecks, with most studies focusing on attention heads.
no code implementations • 21 Aug 2024 • Yiquan Wu, Bo Tang, Chenyang Xi, Yu Yu, Pengyu Wang, Yifei Liu, Kun Kuang, Haiying Deng, Zhiyu Li, Feiyu Xiong, Jie Hu, Peng Cheng, Zhonghao Wang, Yi Wang, Yi Luo, MingChuan Yang
To address the advanced requirements, we present an argument ranking model for arguments and establish a comprehensive evidence database that includes up-to-date events and classic books, thereby strengthening the substantiation of the evidence with retrieval augmented generation (RAG) technology.
no code implementations • 1 Jul 2024 • Hongkang Yang, Zehao Lin, Wenjin Wang, Hao Wu, Zhiyu Li, Bo Tang, Wenqiang Wei, Jinbo Wang, Zeyun Tang, Shichao Song, Chenyang Xi, Yu Yu, Kai Chen, Feiyu Xiong, Linpeng Tang, Weinan E
The model is named $\text{Memory}^3$, since explicit memory is the third form of memory in LLMs after implicit memory (model parameters) and working memory (context key-values).
1 code implementation • 23 Jun 2024 • Junyi Zhu, Shuochen Liu, Yu Yu, Bo Tang, Yibo Yan, Zhiyu Li, Feiyu Xiong, Tong Xu, Matthew B. Blaschko
Large language models (LLMs) excel in generating coherent text, but they often struggle with context awareness, leading to inaccuracies in tasks requiring faithful adherence to provided information.
no code implementations • 13 Jun 2024 • Xuemin Hu, Shen Li, Yingfen Xu, Bo Tang, Long Chen
Offline reinforcement learning (RL) can learn optimal policies from pre-collected offline datasets without interacting with the environment, but the sampled actions of the agent cannot often cover the action distribution under a given state, resulting in the extrapolation error issue.
1 code implementation • 27 May 2024 • Xun Liang, Simin Niu, Zhiyu Li, Sensen Zhang, Shichao Song, Hanyu Wang, Jiawei Yang, Feiyu Xiong, Bo Tang, Chenyang Xi
Retrieval-Augmented Generation (RAG) offers a cost-effective approach to injecting real-time knowledge into large language models (LLMs).
1 code implementation • 20 May 2024 • Qingchen Yu, Zifan Zheng, Shichao Song, Zhiyu Li, Feiyu Xiong, Bo Tang, Ding Chen
The continuous advancement of large language models (LLMs) has brought increasing attention to the critical issue of developing fair and reliable methods for evaluating their performance.
no code implementations • 15 May 2024 • Xingyu Li, Bo Tang
Deep neural networks suffer from the catastrophic forgetting problem in the field of continual learning (CL).
no code implementations • 8 May 2024 • Renjie Liu, Yichuan Wang, Xiao Yan, Zhenkun Cai, Minjie Wang, Haitian Jiang, Bo Tang, Jinyang Li
In particular, by conducting graph sampling beforehand, DiskGNN acquires the node features that will be accessed by model computation, and such information is utilized to pack the target node features contiguously on disk to avoid read amplification.
no code implementations • 27 Mar 2024 • Xuemin Hu, Pan Chen, Yijun Wen, Bo Tang, Long Chen
Reinforcement learning (RL) has been widely used in decision-making and control tasks, but the risk is very high for the agent in the training process due to the requirements of interaction with the environment, which seriously limits its industrial applications such as autonomous driving systems.
no code implementations • 7 Mar 2024 • Yu Zhu, Chuxiong Sun, Wenfei Yang, Wenqiang Wei, Bo Tang, Tianzhu Zhang, Zhiyu Li, Shifeng Zhang, Feiyu Xiong, Jie Hu, MingChuan Yang
Reinforcement Learning from Human Feedback (RLHF) is the prevailing approach to ensure Large Language Models (LLMs) align with human values.
1 code implementation • 29 Feb 2024 • Miao Li, Ming-Bin Chen, Bo Tang, Shengbin Hou, Pengyu Wang, Haiying Deng, Zhiyu Li, Feiyu Xiong, Keming Mao, Peng Cheng, Yi Luo
We present NewsBench, a novel evaluation framework to systematically assess the capabilities of Large Language Models (LLMs) for editorial capabilities in Chinese journalism.
1 code implementation • 17 Feb 2024 • Xun Liang, Hanyu Wang, Shichao Song, Mengting Hu, Xunzhi Wang, Zhiyu Li, Feiyu Xiong, Bo Tang
In this study, we introduce a pluggable CTG framework for Large Language Models (LLMs) named Dynamic Attribute Graphs-based controlled text generation (DATG).
1 code implementation • 12 Feb 2024 • Wentao Ning, Reynold Cheng, Xiao Yan, Ben Kao, Nan Huo, Nur AI Hasan Haldar, Bo Tang
Many methods have been proposed to reduce GP bias but they fail to notice the fundamental problem of GP, i. e., it considers popularity from a \textit{global} perspective of \textit{all users} and uses a single set of popular items, and thus cannot capture the interests of individual users.
1 code implementation • 30 Jan 2024 • Yuanjie Lyu, Zhiyu Li, Simin Niu, Feiyu Xiong, Bo Tang, Wenjin Wang, Hao Wu, Huanyong Liu, Tong Xu, Enhong Chen
For each of these CRUD categories, we have developed comprehensive datasets to evaluate the performance of RAG systems.
2 code implementations • 26 Jan 2024 • Zifan Wu, Bo Tang, Qian Lin, Chao Yu, Shangqin Mao, Qianlong Xie, Xingxing Wang, Dong Wang
Results on benchmark tasks show that our method not only achieves an asymptotic performance comparable to state-of-the-art on-policy methods while using much fewer samples, but also significantly reduces constraint violation during training.
1 code implementation • 7 Jan 2024 • Ding Chen, Shichao Song, Qingchen Yu, Zhiyu Li, Wenjin Wang, Feiyu Xiong, Bo Tang
In this paper, we propose a method SLEICL that involves learning from examples using strong language models and then summarizing and transferring these learned skills to weak language models for inference and application.
no code implementations • 29 Dec 2023 • Hao Wang, Bo Tang, Chi Harold Liu, Shangqin Mao, Jiahong Zhou, Zipeng Dai, Yaqi Sun, Qianlong Xie, Xingxing Wang, Dong Wang
Online display advertising platforms service numerous advertisers by providing real-time bidding (RTB) for the scale of billions of ad requests every day.
no code implementations • 27 Dec 2023 • Jiahong Zhou, Shunhui Mao, Guoliang Yang, Bo Tang, Qianlong Xie, Lebin Lin, Xingxing Wang, Dong Wang
The existing studies focus on dynamically allocating CRs in queue truncation scenarios (i. e., allocating the size of candidates), and formulate the CR allocation problem as an optimization problem with constraints.
1 code implementation • 12 Dec 2023 • Bo Tang, Elias B. Khalil
The end-to-end predict-then-optimize framework, also known as decision-focused learning, has gained popularity for its ability to integrate optimization into the training procedure of machine learning models that predict the unknown cost (objective function) coefficients of optimization problems from contextual instance information.
1 code implementation • 26 Nov 2023 • Xun Liang, Shichao Song, Simin Niu, Zhiyu Li, Feiyu Xiong, Bo Tang, Yezhaohui Wang, Dawei He, Peng Cheng, Zhonghao Wang, Haiying Deng
These techniques encompass the use of directed hallucination induction and strategies that deliberately alter authentic text to produce hallucinations.
1 code implementation • 10 Aug 2023 • Wentao Ning, Xiao Yan, Weiwen Liu, Reynold Cheng, Rui Zhang, Bo Tang
We propose a new MDR method named EDDA with two key components, i. e., embedding disentangling recommender and domain alignment, to tackle the two challenges respectively.
no code implementations • 7 Aug 2023 • Xingyu Li, Bo Tang
Deep neural networks (DNNs) have demonstrated promising results in various complex tasks.
no code implementations • 7 Aug 2023 • Xingyu Li, Bo Tang, Haifeng Li
Continual lifelong learning is an machine learning framework inspired by human learning, where learners are trained to continuously acquire new knowledge in a sequential manner.
1 code implementation • 1 Jun 2023 • Qian Lin, Bo Tang, Zifan Wu, Chao Yu, Shangqin Mao, Qianlong Xie, Xingxing Wang, Dong Wang
Aiming at promoting the safe real-world deployment of Reinforcement Learning (RL), research on safe RL has made significant progress in recent years.
no code implementations • 2 May 2023 • Xuemin Hu, Shen Li, Tingyu Huang, Bo Tang, Rouxing Huai, Long Chen
In general, a large scale of testing in simulation environment is conducted and then the learned driving knowledge is transferred to the real world, so how to adapt driving knowledge learned in simulation to reality becomes a critical issue.
no code implementations • 5 Apr 2023 • Da Li, Bo Tang, Lei Xue
This paper focuses on the joint design of transmit waveforms and receive filters for airborne multiple-input-multiple-output (MIMO) radar systems in spectrally crowded environments.
no code implementations • 4 Apr 2023 • Da Li, Bo Tang, Xuyang Wang, Wenjun Wu, Lei Xue
Reconfigurable intelligent surface (RIS) refers to a signal reflection surface containing a large number of low-cost passive reflecting elements.
no code implementations • 28 Feb 2023 • Wenjun Wu, Bo Tang, Xuyang Wang
We investigate the constant-modulus (CM) waveform design for dual-function radar communication systems in the presence of clutter. To minimize the interference power and enhance the target acquisition performance, we use the signal-to-interference-plus-noise-ratio as the design metric. In addition, to ensure the quality of the service for each communication user, we enforce a constraint on the synthesis error of every communication signals. An iterative algorithm, which is based on cyclic optimization, Dinkinbach's transform, and alternating direction of method of multipliers, is proposed to tackle the encountered non-convex optimization problem. Simulations illustrate that the CM waveforms synthesized by the proposed algorithm allow to suppress the clutter efficiently and control the synthesis error of communication signals to a low level.
no code implementations • 13 Dec 2022 • Ruobing Shen, Bo Tang, Andrea Lodi, Ismail Ben Ayed, Thomas Guthier
We address interactive panoptic annotation, where one segment all object and stuff regions in an image.
no code implementations • 28 Nov 2022 • Peiqi Yin, Xiao Yan, Jinjing Zhou, Qiang Fu, Zhenkun Cai, James Cheng, Bo Tang, Minjie Wang
In this paper, we develop Deep Graph Inference (DGI) -- a system for easy and efficient GNN model inference, which automatically translates the training code of a GNN model for layer-wise execution.
no code implementations • 8 Nov 2022 • Bo Tang, Vijay K. Shah, Vuk Marojevic, Jeffrey H. Reed
This article presents a general automated, distributed and AI-enabled testing framework to test AI models deployed in O-RAN in terms of their decision-making performance, vulnerability and security.
no code implementations • 8 Aug 2022 • Arindam Bose, Bo Tang, Wenjie Huang, Mojtaba Soltanalian, Jian Li
The mutual interference between similar radar systems can result in reduced radar sensitivity and increased false alarm rates.
no code implementations • 16 Jul 2022 • Fanglin Chen, Xiao Liu, Bo Tang, Feiyu Xiong, Serim Hwang, Guomian Zhuang
During deployment, we combine the offline RL model with the LP model to generate a robust policy under the budget constraints.
1 code implementation • 28 Jun 2022 • Bo Tang, Elias B. Khalil
PyEPO provides a simple interface for the definition of new optimization problems, the implementation of state-of-the-art predict-then-optimize training algorithms, the use of custom neural network architectures, and the comparison of end-to-end approaches with the two-stage approach.
no code implementations • 6 Jun 2022 • Zhe Qu, Xingyu Li, Rui Duan, Yao Liu, Bo Tang, Zhuo Lu
Therefore, in this paper, we revisit the solutions to the distribution shift problem in FL with a focus on local learning generality.
no code implementations • 9 Apr 2022 • Xuyang Wang, Bo Tang, Ming Zhang
This paper addresses robust waveform design for multiple-input-multiple-output (MIMO) radar detection.
no code implementations • 8 Jan 2022 • Xingyu Li, Zhe Qu, Shangqing Zhao, Bo Tang, Zhuo Lu, Yao Liu
Federated learning (FL) provides a high efficient decentralized machine learning framework, where the training data remains distributed at remote clients in a network.
1 code implementation • CVPR 2022 • Dan Zeng, Zhiyuan Lin, Xiao Yan, YuTing Liu, Fei Wang, Bo Tang
To combat the mismatch between FR and FER data, Meta-Face2Exp uses a circuit feedback mechanism, which improves the base network with the feedback from the adaptation network.
no code implementations • 23 Dec 2021 • Wentao Ning, Reynold Cheng, Jiajun Shen, Nur Al Hasan Haldar, Ben Kao, Xiao Yan, Nan Huo, Wai Kit Lam, Tian Li, Bo Tang
Specifically, we define a vector encoding for meta-paths and design a policy network to extend meta-paths.
no code implementations • 22 Dec 2021 • Xingyu Li, Zhe Qu, Bo Tang, Zhuo Lu
Federated Learning (FL) is a decentralized machine learning architecture, which leverages a large number of remote devices to learn a joint model with distributed training data.
no code implementations • 4 Dec 2021 • Jian Peng, Dingqi Ye, Bo Tang, Yinjie Lei, Yu Liu, Haifeng Li
This work proposes a general framework named Cycled Memory Networks (CMN) to address the anterograde forgetting in neural networks for lifelong learning.
1 code implementation • NeurIPS 2021 • Yang Zhang, Bo Tang, Qingyu Yang, Dou An, Hongyin Tang, Chenyang Xi, Xueying Li, Feiyu Xiong
Further, a novel offline reinforcement learning method and an off-policy evaluation algorithm are proposed for policy learning and policy evaluation, respectively.
no code implementations • 23 Nov 2021 • Yifan Chang, Wenbo Li, Jian Peng, Bo Tang, Yu Kang, Yinjie Lei, Yuanmiao Gui, Qing Zhu, Yu Liu, Haifeng Li
Different from previous reviews that mainly focus on the catastrophic forgetting phenomenon in CL, this paper surveys CL from a more macroscopic perspective based on the Stability Versus Plasticity mechanism.
no code implementations • 21 Nov 2021 • Jian Peng, Xian Sun, Min Deng, Chao Tao, Bo Tang, Wenbo Li, Guohua Wu, QingZhu, Yu Liu, Tao Lin, Haifeng Li
This paper presents a learning model by active forgetting mechanism with artificial neural networks.
no code implementations • Proceedings of the 2021 International Conference on Management of Data 2021 • Yidi Wu, Yuntao Gui, Tatiana Jin, James Cheng, Xiao Yan, Peiqi Yin, Yufei Cai, Bo Tang, Fan Yu
Graph neural networks (GNNs) have achieved remarkable performance in many graph analytics tasks such as node classification, link prediction and graph clustering.
no code implementations • 12 May 2021 • Chenyang Xi, Bo Tang, Jiajun Shen, Xinfu Liu, Feiyu Xiong, Xueying Li
We make it open-source for fair and comprehensive competitions between offline RL algorithms with complete datasets and checkpoints being provided.
no code implementations • 18 Mar 2021 • Bo Tang, Jun Liu, Hai Wang, Yihua Hu
Range profiling refers to the measurement of target response along the radar slant range.
no code implementations • 12 Feb 2021 • Xingyu Li, Zhe Qu, Bo Tang, Zhuo Lu
Federated learning (FL) is a new machine learning framework which trains a joint model across a large amount of decentralized computing devices.
1 code implementation • 19 Dec 2019 • Jian Peng, Bo Tang, Hao Jiang, Zhuo Li, Yinjie Lei, Tao Lin, Haifeng Li
It is due to two facts: first, as the model learns more tasks, the intersection of the low-error parameter subspace satisfying for these tasks becomes smaller or even does not exist; second, when the model learns a new task, the cumulative error keeps increasing as the model tries to protect the parameter configuration of previous tasks from interference.
no code implementations • 25 Sep 2019 • Ruobing Shen, Bo Tang, Ismail Ben Ayed, Andrea Lodi, Thomas Guthier
Large-scale ground truth data sets are of crucial importance for deep learning based segmentation models, but annotating per-pixel masks is prohibitively time consuming.
no code implementations • 13 Dec 2018 • Mingyue Shang, Zhenxin Fu, Hongzhi Yin, Bo Tang, Dongyan Zhao, Rui Yan
In this paper, we incorporate the logic information with the help of the Natural Language Inference (NLI) task to the Story Cloze Test (SCT).
no code implementations • NeurIPS 2016 • Xi Chen, Yu Cheng, Bo Tang
This is the first upper bound for $RTD(C)$ that depends only on $VCD(C)$, independent of the size of the concept class $|C|$ and its~domain size $n$.
no code implementations • 28 Jun 2016 • Bo Tang, Haibo He
A Relative Density-based Outlier Score (RDOS) is introduced to measure the local outlierness of objects, in which the density distribution at the location of an object is estimated with a local KDE method based on extended nearest neighbors of the object.
no code implementations • 21 Jun 2016 • Bo Tang, Paul M. Baggenstoss, Haibo He
The recognition diversity indicates that a hybrid combination of the proposed generative classifier and the discriminative classifier could further improve the classification performance.
no code implementations • 20 Jun 2016 • Bo Tang, Haibo He
In this paper, we present a new wrapper feature selection approach based on Jensen-Shannon (JS) divergence, termed feature selection with maximum JS-divergence (FSMJ), for text categorization.
no code implementations • 11 May 2016 • Bo Tang, Steven Kay, Haibo He, Paul M. Baggenstoss
In this letter, we present a novel exponentially embedded families (EEF) based classification method, in which the probability density function (PDF) on raw data is estimated from the PDF on features.
no code implementations • 9 Feb 2016 • Bo Tang, Steven Kay, Haibo He
Based on the JMH-divergence, we develop two efficient feature selection methods, termed maximum discrimination ($MD$) and $MD-\chi^2$ methods, for text categorization.