1 code implementation • 18 Dec 2024 • Xiaobao Wu, Liangming Pan, Yuxi Xie, Ruiwen Zhou, Shuai Zhao, Yubo Ma, Mingzhe Du, Rui Mao, Anh Tuan Luu, William Yang Wang
Data contamination hinders fair LLM evaluation by introducing test data into newer models' training sets.
no code implementations • 21 Nov 2024 • Junhua Liu, Fanfan Lin, Xinze Li, Kwan Hui Lim, Shuai Zhao
LLM-based autonomous agents have demonstrated outstanding performance in solving complex industrial tasks.
no code implementations • 21 Oct 2024 • Wenyi Xiao, Zechuan Wang, Leilei Gan, Shuai Zhao, Wanggui He, Luu Anh Tuan, Long Chen, Hao Jiang, Zhou Zhao, Fei Wu
With the rapid advancement of large language models (LLMs), aligning policy models with human preferences has become increasingly critical.
1 code implementation • 18 Oct 2024 • Shuai Zhao, Xiaobao Wu, Cong-Duy Nguyen, Meihuizi Jia, Yichao Feng, Luu Anh Tuan
Then, this teacher model guides the large-scale poisoned student model in unlearning the backdoor, leveraging PEFT.
no code implementations • 26 Sep 2024 • Shuai Zhao, Leilei Gan, Zhongliang Guo, Xiaobao Wu, Luwei Xiao, Xiaoyu Xu, Cong-Duy Nguyen, Luu Anh Tuan
Despite being widely applied due to their exceptional capabilities, Large Language Models (LLMs) have been proven to be vulnerable to backdoor attacks.
no code implementations • 20 Aug 2024 • Zhongliang Guo, Lei Fang, Jingyu Lin, Yifei Qian, Shuai Zhao, Zeyu Wang, Junhao Dong, Cunjian Chen, Ognjen Arandjelović, Chun Pong Lau
Recent advancements in generative AI, particularly Latent Diffusion Models (LDMs), have revolutionized image synthesis and manipulation.
1 code implementation • 11 Aug 2024 • Shuai Zhao, Yongkun Du, Zhineng Chen, Yu-Gang Jiang
Extensive experiments across various STR decoders and language recognition tasks underscore the broad applicability and remarkable performance of DPTR, providing a novel insight for STR pre-training.
no code implementations • 10 Jun 2024 • Shuai Zhao, Meihuizi Jia, Zhongliang Guo, Leilei Gan, Xiaoyu Xu, Xiaobao Wu, Jie Fu, Yichao Feng, Fengjun Pan, Luu Anh Tuan
Large Language Models (LLMs), which bridge the gap between human language understanding and complex problem-solving, achieve state-of-the-art performance on several NLP tasks, particularly in few-shot and zero-shot settings.
1 code implementation • 16 Apr 2024 • Fan Liu, Shuai Zhao, Zhiyong Cheng, Liqiang Nie, Mohan Kankanhalli
This model performs high-order graph convolution on cluster-specific graphs, which are constructed by capturing the multiple interests of users and identifying the common interests among them.
1 code implementation • 23 Mar 2024 • Shuai Zhao, Linchao Zhu, Ruijie Quan, Yi Yang
For the last-$k$ words test with OpenLLaMA-3B, 11 out of 16 users, with an average of 24 examples each, successfully identify their data from 1. 8M examples.
no code implementations • 21 Mar 2024 • Fanfan Lin, Junhua Liu, Xinze Li, Shuai Zhao, Bohui Zhao, Hao Ma, Xin Zhang
This paper proposes PE-GPT, a custom-tailored large language model uniquely adapted for power converter modulation design.
no code implementations • 19 Feb 2024 • Shuai Zhao, Leilei Gan, Luu Anh Tuan, Jie Fu, Lingjuan Lyu, Meihuizi Jia, Jinming Wen
Motivated by this insight, we developed a Poisoned Sample Identification Module (PSIM) leveraging PEFT, which identifies poisoned samples through confidence, providing robust defense against weight-poisoning backdoor attacks.
no code implementations • 11 Jan 2024 • Shuai Zhao, Meihuizi Jia, Luu Anh Tuan, Fengjun Pan, Jinming Wen
Our studies demonstrate that an attacker can manipulate the behavior of large language models by poisoning the demonstration context, without the need for fine-tuning the model.
no code implementations • 10 Oct 2023 • Hongbin Xu, Yamei Xia, Shuai Zhao, Bo Cheng
We improve the self-attention by isolating connections between irrelevant objects that makes it focus on local regions but not global regions.
no code implementations • 1 Aug 2023 • Minghao Chen, Zepeng Gao, Shuai Zhao, Qibo Qiu, Wenxiao Wang, Binbin Lin, Xiaofei He
Unsupervised domain adaptation (UDA) methods facilitate the transfer of models to target domains without labels.
1 code implementation • 29 May 2023 • Shuai Zhao, Xiaohan Wang, Linchao Zhu, Yi Yang
Given a single test sample, the VLM is forced to maximize the CLIP reward between the input and sampled results from the VLM output distribution.
1 code implementation • 23 May 2023 • Shuai Zhao, Ruijie Quan, Linchao Zhu, Yi Yang
With such merits, we transform CLIP into a scene text reader and introduce CLIP4STR, a simple yet effective STR method built upon image and text encoders of CLIP.
Ranked #1 on Scene Text Recognition on Uber-Text
no code implementations • 2 May 2023 • Shuai Zhao, Jinming Wen, Luu Anh Tuan, Junbo Zhao, Jie Fu
Our method does not require external triggers and ensures correct labeling of poisoned samples, improving the stealthy nature of the backdoor attack.
1 code implementation • 2 Mar 2023 • Yingting Li, Ambuj Mehrish, Shuai Zhao, Rishabh Bhardwaj, Amir Zadeh, Navonil Majumder, Rada Mihalcea, Soujanya Poria
To mitigate this issue, parameter-efficient transfer learning algorithms, such as adapters and prefix tuning, have been proposed as a way to introduce a few trainable parameters that can be plugged into large pre-trained language models such as BERT, and HuBERT.
1 code implementation • 2 Jan 2023 • Pengfei Wen, Zhi-Sheng Ye, Yong Li, Shaowei Chen, Pu Xie, Shuai Zhao
Physics-Informed Neural Network (PINN) is an efficient tool to fuse empirical or physical dynamic models with data-driven models.
no code implementations • 8 Dec 2022 • Xinle Wu, Dalin Zhang, Miao Zhang, Chenjuan Guo, Shuai Zhao, Yi Zhang, Huai Wang, Bin Yang
We then propose a resource-aware search strategy to explore the search space to find the best PINN model under different resource constraints.
no code implementations • 4 Nov 2022 • Wenting Ye, Hongfei Yang, Shuai Zhao, Haoyang Fang, Xingjian Shi, Naveen Neppalli
The substitute-based recommendation is widely used in E-commerce to provide better alternatives to customers.
1 code implementation • 22 Oct 2022 • Jiale Han, Shuai Zhao, Bo Cheng, Shengkun Ma, Wei Lu
Current prompt tuning methods mostly convert the downstream tasks to masked language modeling problems by adding cloze-style phrases and mapping all labels to verbalizations with fixed length, which has proven effective for tasks with simple label spaces.
Ranked #2 on Relation Extraction on Re-TACRED
1 code implementation • 30 Sep 2022 • Shuai Zhao, Linchao Zhu, Xiaohan Wang, Yi Yang
In this work, we introduce another one-stage solution to obtain pre-trained small models without the need for extra teachers, namely, slimmable networks for contrastive self-supervised learning (SlimCLR).
1 code implementation • Neural Processing Letters 2022 • Shuai Zhao
The summary generation model equipped with gradient penalty avoids overfitting and makes the model more stable.
Ranked #1 on Abstractive Text Summarization on EDUsum
no code implementations • 29 Aug 2022 • Pengfei Zhu, Xinjie Yao, Yu Wang, Meng Cao, Binyuan Hui, Shuai Zhao, QinGhua Hu
Multi-view learning has progressed rapidly in recent years.
no code implementations • 28 Jul 2022 • Hongyu Shen, Jinoh Oh, Shuai Zhao, Guoyin Wang, Tara Taghavi, Sungjin Lee
Then we propose a graph convolutional network(GCN) based model, namely Personalized Dynamic Routing Feature Encoder(PDRFE), that generates personalized customer representations learned from the built graph.
1 code implementation • NAACL 2022 • Devamanyu Hazarika, Yingting Li, Bo Cheng, Shuai Zhao, Roger Zimmermann, Soujanya Poria
In this work, we hope to address that by (i) Proposing simple diagnostic checks for modality robustness in a trained multimodal model.
no code implementations • 4 May 2022 • Yi Liang, Shuai Zhao, Bo Cheng, Yuwei Yin, Hao Yang
Few-shot relation learning refers to infer facts for relations with a limited number of observed triples.
1 code implementation • 2 May 2022 • Shuai Zhao, Linchao Zhu, Xiaohan Wang, Yi Yang
In this paper, to reduce the number of redundant video tokens, we design a multi-segment token clustering algorithm to find the most representative tokens and drop the non-essential ones.
Ranked #11 on Video Retrieval on MSVD (using extra training data)
no code implementations • 26 Mar 2022 • Sha Yuan, Hanyu Zhao, Shuai Zhao, Jiahong Leng, Yangxiao Liang, Xiaozhi Wang, Jifan Yu, Xin Lv, Zhou Shao, Jiaao He, Yankai Lin, Xu Han, Zhenghao Liu, Ning Ding, Yongming Rao, Yizhao Gao, Liang Zhang, Ming Ding, Cong Fang, Yisen Wang, Mingsheng Long, Jing Zhang, Yinpeng Dong, Tianyu Pang, Peng Cui, Lingxiao Huang, Zheng Liang, HuaWei Shen, HUI ZHANG, Quanshi Zhang, Qingxiu Dong, Zhixing Tan, Mingxuan Wang, Shuo Wang, Long Zhou, Haoran Li, Junwei Bao, Yingwei Pan, Weinan Zhang, Zhou Yu, Rui Yan, Chence Shi, Minghao Xu, Zuobai Zhang, Guoqiang Wang, Xiang Pan, Mengjie Li, Xiaoyu Chu, Zijun Yao, Fangwei Zhu, Shulin Cao, Weicheng Xue, Zixuan Ma, Zhengyan Zhang, Shengding Hu, Yujia Qin, Chaojun Xiao, Zheni Zeng, Ganqu Cui, Weize Chen, Weilin Zhao, Yuan YAO, Peng Li, Wenzhao Zheng, Wenliang Zhao, Ziyi Wang, Borui Zhang, Nanyi Fei, Anwen Hu, Zenan Ling, Haoyang Li, Boxi Cao, Xianpei Han, Weidong Zhan, Baobao Chang, Hao Sun, Jiawen Deng, Chujie Zheng, Juanzi Li, Lei Hou, Xigang Cao, Jidong Zhai, Zhiyuan Liu, Maosong Sun, Jiwen Lu, Zhiwu Lu, Qin Jin, Ruihua Song, Ji-Rong Wen, Zhouchen Lin, LiWei Wang, Hang Su, Jun Zhu, Zhifang Sui, Jiajun Zhang, Yang Liu, Xiaodong He, Minlie Huang, Jian Tang, Jie Tang
With the rapid development of deep learning, training Big Models (BMs) for multiple downstream tasks becomes a popular paradigm.
no code implementations • 22 Mar 2022 • Sha Yuan, Shuai Zhao, Jiahong Leng, Zhao Xue, Hanyu Zhao, Peiyu Liu, Zheng Gong, Wayne Xin Zhao, Junyi Li, Jie Tang
The results show that WuDaoMM can be applied as an efficient dataset for VLPMs, especially for the model in text-to-image generation task.
no code implementations • 23 Nov 2021 • Pengfei Zhu, Hongtao Yu, Kaihua Zhang, Yu Wang, Shuai Zhao, Lei Wang, Tianzhu Zhang, QinGhua Hu
To address this issue, segmentation-based trackers have been proposed that employ per-pixel matching to improve the tracking performance of deformable objects effectively.
no code implementations • Findings (EMNLP) 2021 • Xu Wang, Hainan Zhang, Shuai Zhao, Yanyan Zou, Hongshen Chen, Zhuoye Ding, Bo Cheng, Yanyan Lan
Furthermore, the consistency signals between each candidate and the speaker's own history are considered to drive a model to prefer a candidate that is logically consistent with the speaker's history logic.
no code implementations • 13 Jun 2021 • Xiaopeng Jiang, Shuai Zhao, Guy Jacobson, Rittwik Jana, Wen-Ling Hsu, Manoop Talasila, Syed Anwar Aftab, Yi Chen, Cristian Borcea
The framework runs on the phones of the users and also on a server that coordinates learning from all users in the system.
no code implementations • 9 Jun 2021 • Boxi Wu, Heng Pan, Li Shen, Jindong Gu, Shuai Zhao, Zhifeng Li, Deng Cai, Xiaofei He, Wei Liu
In this work, we find that the adversarial attacks can also be vulnerable to small perturbations.
1 code implementation • 1 Apr 2021 • Tu Zheng, Shuai Zhao, Yang Liu, Zili Liu, Deng Cai
In this paper, we propose Side Overlap~(SO) loss by maximizing the side overlap of two bounding boxes, which puts more penalty for low overlapping bounding box cases.
no code implementations • 1 Apr 2021 • Xu Wang, Shuai Zhao, Bo Cheng, Jiale Han, Yingting Li, Hao Yang, Ivan Sekulic, Guoshun Nan
Question Answering (QA) models over Knowledge Bases (KBs) are capable of providing more precise answers by utilizing relation information among entities.
no code implementations • 10 Mar 2021 • Dong Shen, Shuai Zhao, Jinming Hu, Hao Feng, Deng Cai, Xiaofei He
In this paper, we propose a novel network, Erasing-Salient Net (ES-Net), to learn comprehensive features by erasing the salient areas in an image.
no code implementations • COLING 2020 • Xu Wang, Shuai Zhao, Jiale Han, Bo Cheng, Hao Yang, Jianchang Ao, Zhenzi Li
The structural information of Knowledge Bases (KBs) has proven effective to Question Answering (QA).
2 code implementations • 30 Nov 2020 • Shuai Zhao, Liguang Zhou, Wenxiao Wang, Deng Cai, Tin Lun Lam, Yangsheng Xu
Each of these small networks has a fraction of the original one's parameters.
Ranked #31 on Image Classification on CIFAR-100 (using extra training data)
no code implementations • 10 Oct 2020 • Wenxiao Wang, Minghao Chen, Shuai Zhao, Long Chen, Jinming Hu, Haifeng Liu, Deng Cai, Xiaofei He, Wei Liu
Specifically, it first casts the relationships between a certain model's accuracy and depth/width/resolution into a polynomial regression and then maximizes the polynomial to acquire the optimal values for the three dimensions.
no code implementations • 10 Sep 2020 • Michal Witold Przewozniczek, Piotr Dziurzanski, Shuai Zhao, Leandro Soares Indrusiak
Evolutionary methods are effective tools for obtaining high-quality results when solving hard practical problems.
1 code implementation • 4 Jan 2020 • Minghao Chen, Shuai Zhao, Haifeng Liu, Deng Cai
In order to combine the strengths of these two methods, we propose a novel method called Adversarial-Learned Loss for Domain Adaptation (ALDA).
no code implementations • 21 Dec 2019 • Wenxiao Wang, Shuai Zhao, Minghao Chen, Jinming Hu, Deng Cai, Haifeng Liu
The dominant pruning methods, filter-level pruning methods, evaluate their performance through the reduction ratio of computations and deem that a higher reduction ratio of computations is equivalent to a higher acceleration ratio in terms of inference time.
2 code implementations • NeurIPS 2019 • Shuai Zhao, Yang Wang, Zheng Yang, Deng Cai
In this paper, we develop a region mutual information (RMI) loss to model the dependencies among pixels more simply and efficiently.
no code implementations • 19 Oct 2019 • Shuai Zhao, Boxi Wu, Wenqing Chu, Yao Hu, Deng Cai
Inspired by the widely-used structural similarity (SSIM) index in image quality assessment, we use the linear correlation between two images to quantify their structural similarity.
no code implementations • 8 Oct 2019 • Shuai Zhao
We generalize the celebrated heavy quark expansion to nonlocal QCD operators.
High Energy Physics - Phenomenology
no code implementations • 23 Aug 2019 • Yanhao Zhu, Zhineng Chen, Shuai Zhao, Hongtao Xie, Wenming Guo, Yongdong Zhang
Nowadays U-net-like FCNs predominate various biomedical image segmentation applications and attain promising performance, largely due to their elegant architectures, e. g., symmetric contracting and expansive paths as well as lateral skip-connections.
no code implementations • 16 Jul 2019 • Boxi Wu, Shuai Zhao, Wenqing Chu, Zheng Yang, Deng Cai
To be specific, our method explicitly requires the network to predict semantic segmentation as well as dilated affinity, which is a sparse version of pair-wise pixel affinity.
no code implementations • WS 2018 • Sizhen Li, Shuai Zhao, Bo Cheng, Hao Yang
With huge amount of information generated every day on the web, fact checking is an important and challenging task which can help people identify the authenticity of most claims as well as providing evidences selected from knowledge source like Wikipedia.
no code implementations • 16 Oct 2018 • Shuai Zhao, Manoop Talasila, Guy Jacobson, Cristian Borcea, Syed Anwar Aftab, John F Murray
Applying Machine Learning (ML) to business applications for automation usually faces difficulties when integrating diverse ML dependencies and services, mainly because of the lack of a common ML framework.