no code implementations • ECCV 2020 • Liang Chen, Faming Fang, Jiawei Zhang, Jun Liu, Guixu Zhang
Even a small amount of outliers can dramatically degrade the quality of the estimated blur kernel, because the outliers are not conforming to the linear formation of the blurring process.
2 code implementations • ACL 2022 • Liang Chen, Runxin Xu, Baobao Chang
Label smoothing and vocabulary sharing are two widely used techniques in neural machine translation models.
no code implementations • ECCV 2020 • Liang Chen, Faming Fang, Shen Lei, Fang Li, Guixu Zhang
Specifically, we use a weighted combination of a dense function (i. e. l2) and a newly designed enhanced sparse model termed as le, which is developed from two sparse models (i. e. l1 and l0), to fulfill the task.
1 code implementation • 14 Oct 2024 • Jintang Li, Ruofan Wu, Yuchang Zhu, Huizhe Zhang, Xinzhou Jin, Guibin Zhang, Zulun Zhu, Zibin Zheng, Liang Chen
Graph autoencoders (GAEs) are self-supervised learning models that can learn meaningful representations of graph-structured data by reconstructing the input graph from a low-dimensional latent space.
no code implementations • 12 Oct 2024 • Lei LI, Zhihui Xie, Mukai Li, Shunian Chen, Peiyi Wang, Liang Chen, Yazheng Yang, Benyou Wang, Lingpeng Kong, Qi Liu
As large vision-language models (LVLMs) evolve rapidly, the demand for high-quality and diverse data to align these models becomes increasingly crucial.
1 code implementation • 10 Oct 2024 • Bofei Gao, Feifan Song, Zhe Yang, Zefan Cai, Yibo Miao, Qingxiu Dong, Lei LI, Chenghao Ma, Liang Chen, Runxin Xu, Zhengyang Tang, Benyou Wang, Daoguang Zan, Shanghaoran Quan, Ge Zhang, Lei Sha, Yichang Zhang, Xuancheng Ren, Tianyu Liu, Baobao Chang
However, existing benchmarks like GSM8K or MATH are now being solved with high accuracy (e. g., OpenAI o1 achieves 94. 8% on MATH dataset), indicating their inadequacy for truly challenging these models.
1 code implementation • 2 Oct 2024 • Liang Chen, Sinan Tan, Zefan Cai, Weichu Xie, Haozhe Zhao, Yichi Zhang, Junyang Lin, Jinze Bai, Tianyu Liu, Baobao Chang
This work tackles the information loss bottleneck of vector-quantization (VQ) autoregressive image generation by introducing a novel model architecture called the 2-Dimensional Autoregression (DnD) Transformer.
no code implementations • 26 Sep 2024 • Zhongxin Yu, Liang Chen, Zhiyun Zeng, Kunping Yang, Shaofei Luo, Shaorui Chen, Cheng Zhong
Specifically owing to neighboring regions of the same pixel position in different sub-aperture images exhibit similar structural relationships we design a lightweight CNN-based feature extraction module (namely DGCE) to extract local features better through feature modulation.
no code implementations • 9 Sep 2024 • Junkun Chen, Jilin Mei, Liang Chen, Fangzhou Zhao, Yu Hu
The limited training samples for object detectors commonly result in low accuracy out-of-distribution (OOD) object detection.
1 code implementation • 4 Sep 2024 • Bofei Gao, Feifan Song, Yibo Miao, Zefan Cai, Zhe Yang, Liang Chen, Helan Hu, Runxin Xu, Qingxiu Dong, Ce Zheng, Wen Xiao, Ge Zhang, Daoguang Zan, Keming Lu, Bowen Yu, Dayiheng Liu, Zeyu Cui, Jian Yang, Lei Sha, Houfeng Wang, Zhifang Sui, Peiyi Wang, Tianyu Liu, Baobao Chang
Finally, based on our unified perspective, we explore the challenges and future research directions for aligning large language models with human preferences.
1 code implementation • 28 Aug 2024 • Junbao Zhou, Jilin Mei, Pengze Wu, Liang Chen, Fangzhou Zhao, Xijun Zhao, Yu Hu
However, this approach introduces a data imbalance biased to novel data that presents a new challenge of catastrophic forgetting.
1 code implementation • 19 Jul 2024 • Xinzhou Jin, Jintang Li, Liang Chen, Chenyun Yu, Yuanzhen Xie, Tao Xie, Chengxiang Zhuo, Zang Li, Zibin Zheng
Surprisingly, we find that L2CL, using only one-hop contrastive learning paradigm, is able to capture intrinsic semantic structures and improve the quality of node representation, leading to a simple yet effective architecture.
1 code implementation • 12 Jul 2024 • Fangyuan Mao, Jilin Mei, Shun Lu, Fuyang Liu, Liang Chen, Fangzhou Zhao, Yu Hu
Infrared imaging technology has gained significant attention for its reliable sensing ability in low visibility conditions, prompting many studies to convert the abundant RGB images to infrared images.
no code implementations • 7 Jul 2024 • Haozhe Zhao, Xiaojian Ma, Liang Chen, Shuzheng Si, Rujie Wu, Kaikai An, Peiyu Yu, Minjia Zhang, Qing Li, Baobao Chang
This paper presents UltraEdit, a large-scale (approximately 4 million editing samples), automatically generated dataset for instruction-based image editing.
1 code implementation • 29 Jun 2024 • Jinsheng Huang, Liang Chen, Taian Guo, Fu Zeng, Yusheng Zhao, Bohan Wu, Ye Yuan, Haozhe Zhao, Zhihui Guo, Yichi Zhang, Jingyang Yuan, Wei Ju, Luchen Liu, Tianyu Liu, Baobao Chang, Ming Zhang
Large Multimodal Models (LMMs) exhibit impressive cross-modal understanding and reasoning abilities, often assessed through multiple-choice questions (MCQs) that include an image, a question, and several options.
1 code implementation • 20 Jun 2024 • Yunfei Liu, Jintang Li, Yuehe Chen, Ruofan Wu, Ericbk Wang, Jing Zhou, Sheng Tian, Shuheng Shen, Xing Fu, Changhua Meng, Weiqiang Wang, Liang Chen
Another promising line of research involves the adoption of modularity maximization, a popular and effective measure for community detection, as the guiding principle for clustering tasks.
1 code implementation • 19 Jun 2024 • Yuchang Zhu, Jintang Li, Yatao Bian, Zibin Zheng, Liang Chen
Accordingly, we propose a graph fairness framework based on invariant learning, namely FairINV, which enables the training of fair GNNs to accommodate various sensitive attributes within a single training session.
1 code implementation • 3 Jun 2024 • Jintang Li, Ruofan Wu, Xinzhou Jin, Boqun Ma, Liang Chen, Zibin Zheng
Recently, state space models (SSMs), which are framed as discretized representations of an underlying continuous-time linear dynamical system, have garnered substantial attention and achieved breakthrough advancements in independent sequence modeling.
1 code implementation • 24 May 2024 • Bowei He, Yunpeng Weng, Xing Tang, Ziqiang Cui, Zexu Sun, Liang Chen, Xiuqiang He, Chen Ma
Uplift modeling has been widely employed in online marketing by predicting the response difference between the treatment and control groups, so as to identify the sensitive individuals toward interventions like coupons or discounts.
1 code implementation • 20 May 2024 • Jianhong Han, Liang Chen, Yupei Wang
Firstly, we propose the Class-wise Prototypes Alignment (CPA) module, which effectively aligns cross-domain features in a class-aware manner by bridging the gap between object detection task and domain adaptation task.
1 code implementation • 11 May 2024 • Yuchang Zhu, Jintang Li, Zibin Zheng, Liang Chen
In particular, the objective of group fairness is to ensure that the decisions made by GNNs are independent of the sensitive attribute.
1 code implementation • 5 May 2024 • Ziqi Gao, Qichao Wang, Aochuan Chen, Zijing Liu, Bingzhe Wu, Liang Chen, Jia Li
Low-rank adaptation~(LoRA) has recently gained much interest in fine-tuning foundation models.
1 code implementation • 12 Apr 2024 • Haozhe Zhao, Zefan Cai, Shuzheng Si, Liang Chen, Yufeng He, Kaikai An, Baobao Chang
Therefore, we introduce ALSACE to leverage the learned knowledge from the well-performing languages to guide under-performing ones within the same mPLM, eliminating the need for additional labeled multilingual data.
1 code implementation • 26 Mar 2024 • Huizhe Zhang, Jintang Li, Liang Chen, Zibin Zheng
However, the costs behind outstanding performances of GTs are higher energy consumption and computational overhead.
1 code implementation • 13 Mar 2024 • Liang Chen, Yong Zhang, Yibing Song, Zhen Zhang, Lingqiao Liu
By d-separation, we observe that the causal feature can be further characterized by being independent of the domain conditioned on the object, and we propose the following two strategies as complements for the basic framework.
1 code implementation • 11 Mar 2024 • Liang Chen, Haozhe Zhao, Tianyu Liu, Shuai Bai, Junyang Lin, Chang Zhou, Baobao Chang
To this end, we introduce FastV, a versatile plug-and-play method designed to optimize computational efficiency by learning adaptive attention patterns in early layers and pruning visual tokens in subsequent ones.
1 code implementation • 8 Mar 2024 • Shuaiyi Li, Yang Deng, Deng Cai, Hongyuan Lu, Liang Chen, Wai Lam
As the typical retraining paradigm is unacceptably time- and resource-consuming, researchers are turning to model editing to find an effective way that supports both consecutive and batch scenarios to edit the model behavior directly.
no code implementations • 25 Feb 2024 • Xiangdi Meng, Damai Dai, Weiyao Luo, Zhe Yang, Shaoxiang Wu, Xiaochen Wang, Peiyi Wang, Qingxiu Dong, Liang Chen, Zhifang Sui
Although LoRA fine-tuning is effective, there is still a performance gap compared to full fine-tuning, since its weight update is limited to low-rank matrices.
1 code implementation • 21 Feb 2024 • Liang Chen, Yichi Zhang, Shuhuai Ren, Haozhe Zhao, Zefan Cai, Yuchi Wang, Peiyi Wang, Xiangdi Meng, Tianyu Liu, Baobao Chang
To address this, we introduce Embodied-Instruction-Evolution (EIE), an automatic framework for synthesizing instruction tuning examples in multimodal embodied environments.
1 code implementation • 16 Feb 2024 • Yuanzhen Xie, Xinzhou Jin, Tao Xie, Mingxiong Lin, Liang Chen, Chenyun Yu, Lei Cheng, Chengxiang Zhuo, Bo Hu, Zang Li
To improve the contextual learning capabilities of LLMs in text-to-SQL, a workflow paradigm method is proposed, aiming to enhance the attention and problem-solving scope of LLMs through decomposition.
1 code implementation • 12 Jan 2024 • Ziqiang Cui, Xing Tang, Yang Qiao, Bowei He, Liang Chen, Xiuqiang He, Chen Ma
Firstly, TAHyper employs the hyperbolic space to encode the social networks, thereby effectively reducing the distortion of confounder representation caused by Euclidean embeddings.
no code implementations • 3 Jan 2024 • Yunpeng Weng, Xing Tang, Liang Chen, Dugang Liu, Xiuqiang He
In addition to predicting the click-through rate (CTR) or the conversion rate (CVR) as in traditional recommendations, it is essential for FinTech platforms to estimate the customers' purchase amount for each delivered fund and achieve an effective allocation of impressions based on the predicted results to optimize the total expected transaction value (ETV).
no code implementations • 17 Dec 2023 • Lei LI, Zhihui Xie, Mukai Li, Shunian Chen, Peiyi Wang, Liang Chen, Yazheng Yang, Benyou Wang, Lingpeng Kong
This paper explores preference distillation for large vision language models (LVLMs), improving their ability to generate helpful and faithful responses anchoring the visual context.
Ranked #38 on Visual Question Answering on MM-Vet
1 code implementation • 14 Dec 2023 • Wenbin Zou, Hongxia Gao, Tian Ye, Liang Chen, Weipeng Yang, Shasha Huang, Hongsheng Chen, Sixiang Chen
In this paper, we propose Clearer Night Image Restoration with Vector-Quantized Codebook (VQCNIR) to achieve remarkable and consistent restoration outcomes on real-world and synthetic benchmarks.
1 code implementation • 5 Dec 2023 • Wangbin Sun, Jintang Li, Liang Chen, Bingzhe Wu, Yatao Bian, Zibin Zheng
Graph contrastive learning (GCL) has emerged as a representative paradigm in graph self-supervised learning, where negative samples are commonly regarded as the key to preventing model collapse and producing distinguishable representations.
1 code implementation • 29 Nov 2023 • Yuchang Zhu, Jintang Li, Liang Chen, Zibin Zheng
Experiments on several benchmark datasets demonstrate that FairGKD, which does not require access to demographic information, significantly improves the fairness of GNNs by a large margin while maintaining their utility.
no code implementations • 28 Nov 2023 • Hongru Wang, Lingzhi Wang, Yiming Du, Liang Chen, Jingyan Zhou, YuFei Wang, Kam-Fai Wong
This survey delves into the historical trajectory of dialogue systems, elucidating their intricate relationship with advancements in language models by categorizing this evolution into four distinct stages, each marked by pivotal LM breakthroughs: 1) Early_Stage: characterized by statistical LMs, resulting in rule-based or machine-learning-driven dialogue_systems; 2) Independent development of TOD and ODD based on neural_language_models (NLM; e. g., LSTM and GRU), since NLMs lack intrinsic knowledge in their parameters; 3) fusion between different types of dialogue systems with the advert of pre-trained_language_models (PLMs), starting from the fusion between four_sub-tasks_within_TOD, and then TOD_with_ODD; and 4) current LLM-based_dialogue_system, wherein LLMs can be used to conduct TOD and ODD seamlessly.
no code implementations • 28 Nov 2023 • Jintang Li, Jiawang Dan, Ruofan Wu, Jing Zhou, Sheng Tian, Yunfei Liu, Baokun Wang, Changhua Meng, Weiqiang Wang, Yuchang Zhu, Liang Chen, Zibin Zheng
Over the past few years, graph neural networks (GNNs) have become powerful and practical tools for learning on (static) graph-structure data.
no code implementations • 16 Nov 2023 • Liang Chen, Yatao Bian, Yang Deng, Deng Cai, Shuaiyi Li, Peilin Zhao, Kam-Fai Wong
Text watermarking has emerged as a pivotal technique for identifying machine-generated text.
1 code implementation • 16 Nov 2023 • Xiangru Tang, Yuliang Liu, Zefan Cai, Yanjun Shao, Junjie Lu, Yichi Zhang, Zexuan Deng, Helan Hu, Kaikai An, Ruijun Huang, Shuzheng Si, Sheng Chen, Haozhe Zhao, Liang Chen, Yan Wang, Tianyu Liu, Zhiwei Jiang, Baobao Chang, Yin Fang, Yujia Qin, Wangchunshu Zhou, Yilun Zhao, Arman Cohan, Mark Gerstein
Despite Large Language Models (LLMs) like GPT-4 achieving impressive results in function-level code generation, they struggle with repository-scale code understanding (e. g., coming up with the right arguments for calling routines), requiring a deeper comprehension of complex file interactions.
no code implementations • 24 Oct 2023 • Zezhong Wang, Fangkai Yang, Lu Wang, Pu Zhao, Hongru Wang, Liang Chen, QIngwei Lin, Kam-Fai Wong
Currently, there are two main approaches to address jailbreak attacks: safety training and safeguards.
1 code implementation • NeurIPS 2023 • Fuyuan Lyu, Xing Tang, Dugang Liu, Chen Ma, Weihong Luo, Liang Chen, Xiuqiang He, Xue Liu
In this work, we introduce a hybrid-grained feature interaction selection approach that targets both feature field and feature value for deep sparse networks.
no code implementations • 18 Oct 2023 • Jintang Li, Zheng Wei, Jiawang Dan, Jing Zhou, Yuchang Zhu, Ruofan Wu, Baokun Wang, Zhang Zhen, Changhua Meng, Hong Jin, Zibin Zheng, Liang Chen
Through in-depth investigations on several real-world heterogeneous graphs exhibiting varying levels of heterophily, we have observed that heterogeneous graph neural networks (HGNNs), which inherit many mechanisms from GNNs designed for homogeneous graphs, fail to generalize to heterogeneous graphs with heterophily or low level of homophily.
no code implementations • 18 Oct 2023 • Qichao Wang, Tian Bian, Yian Yin, Tingyang Xu, Hong Cheng, Helen M. Meng, Zibin Zheng, Liang Chen, Bingzhe Wu
The recent surge in the research of diffusion models has accelerated the adoption of text-to-image models in various Artificial Intelligence Generated Content (AIGC) commercial products.
1 code implementation • 13 Oct 2023 • Bofei Gao, Liang Chen, Peiyi Wang, Zhifang Sui, Baobao Chang
Abstract Meaning Representation (AMR) parsing aims to extract an abstract semantic graph from a given sentence.
1 code implementation • 11 Oct 2023 • Liang Chen, Yang Deng, Yatao Bian, Zeyu Qin, Bingzhe Wu, Tat-Seng Chua, Kam-Fai Wong
Large language models (LLMs) outperform information retrieval techniques for downstream knowledge-intensive tasks when being prompted to generate world knowledge.
no code implementations • 10 Oct 2023 • Peng Di, Jianguo Li, Hang Yu, Wei Jiang, Wenting Cai, Yang Cao, Chaoyu Chen, Dajun Chen, Hongwei Chen, Liang Chen, Gang Fan, Jie Gong, Zi Gong, Wen Hu, Tingting Guo, Zhichao Lei, Ting Li, Zheng Li, Ming Liang, Cong Liao, Bingchang Liu, Jiachen Liu, Zhiwei Liu, Shaojun Lu, Min Shen, Guangpei Wang, Huan Wang, Zhi Wang, Zhaogui Xu, Jiawei Yang, Qing Ye, Gehao Zhang, Yu Zhang, Zelin Zhao, Xunjin Zheng, Hailian Zhou, Lifu Zhu, Xianying Zhu
It is specifically designed for code-related tasks with both English and Chinese prompts and supports over 40 programming languages.
no code implementations • 6 Oct 2023 • Liang Chen, Sue Ann Campbell
We investigate the impact of the heterogeneity of the quenched input current, the SFA mechanism and the synaptic delay on macroscopic dynamics.
1 code implementation • 3 Oct 2023 • Liang Chen, Yichi Zhang, Shuhuai Ren, Haozhe Zhao, Zefan Cai, Yuchi Wang, Peiyi Wang, Tianyu Liu, Baobao Chang
In this study, we explore the potential of Multimodal Large Language Models (MLLMs) in improving embodied decision-making processes for agents.
no code implementations • 1 Oct 2023 • Yachuan Liu, Liang Chen, Jindong Wang, Qiaozhu Mei, Xing Xie
We hope this initial work can shed light on future research of LLMs evaluation.
no code implementations • 20 Sep 2023 • Liang Chen, Eugene Choo, Alfred Galichon, Simon Weber
We propose new results for the existence and uniqueness of a general nonparametric and nonseparable competitive equilibrium with substitutes.
2 code implementations • 14 Sep 2023 • Haozhe Zhao, Zefan Cai, Shuzheng Si, Xiaojian Ma, Kaikai An, Liang Chen, Zixuan Liu, Sheng Wang, Wenjuan Han, Baobao Chang
In this paper, we address the limitation above by 1) introducing vision-language Model with Multi-Modal In-Context Learning(MMICL), a new approach to allow the VLM to deal with multi-modal inputs efficiently; 2) proposing a novel context scheme to augment the in-context learning ability of the VLM; 3) constructing the Multi-modal In-Context Learning (MIC) dataset, designed to enhance the VLM's ability to understand complex multi-modal prompts.
Ranked #16 on Visual Reasoning on Winoground
no code implementations • 5 Sep 2023 • Peiyi Wang, Lei LI, Liang Chen, Feifan Song, Binghuai Lin, Yunbo Cao, Tianyu Liu, Zhifang Sui
To address this problem, we introduce an \textit{Alignment Fine-Tuning (AFT)} paradigm, which involves three steps: 1) fine-tuning LLMs with COT training data; 2) generating multiple COT responses for each question, and categorizing them into positive and negative ones based on whether they achieve the correct answer; 3) calibrating the scores of positive and negative responses given by LLMs with a novel constraint alignment loss.
1 code implementation • ICCV 2023 • Liang Chen, Yong Zhang, Yibing Song, Anton Van Den Hengel, Lingqiao Liu
Specifically, we propose treating the element-wise contributions to the final results as the rationale for making a decision and representing the rationale for each sample as a matrix.
1 code implementation • 13 Aug 2023 • Jie Liao, Jintang Li, Liang Chen, Bingzhe Wu, Yatao Bian, Zibin Zheng
In the pursuit of promoting the expressiveness of GNNs for tail nodes, we explore how the deficiency of structural information deteriorates the performance of tail nodes and propose a general Structural Augmentation based taIL nOde Representation learning framework, dubbed as SAILOR, which can jointly learn to augment the graph structure and extract more informative representations for tail nodes.
no code implementations • 11 Aug 2023 • Liang Chen, Yifei Yin, Hao Shi, Qingqing Sheng, Wei Li
The training image pairs are generated by the sub-sampler from real-word SAR image to estimate the noise distribution.
no code implementations • 10 Aug 2023 • Liang Chen, Jiawei Zhang, Zhenhua Li, Yunxuan Wei, Faming Fang, Jimmy Ren, Jinshan Pan
In this paper, we develop a data-driven approach to model the saturated pixels by a learned latent map.
no code implementations • 7 Jun 2023 • Lei LI, Yuwei Yin, Shicheng Li, Liang Chen, Peiyi Wang, Shuhuai Ren, Mukai Li, Yazheng Yang, Jingjing Xu, Xu sun, Lingpeng Kong, Qi Liu
To tackle this challenge and promote research in the vision-language field, we introduce the Multi-Modal, Multilingual Instruction Tuning (M$^3$IT) dataset, designed to optimize VLM alignment with human instructions.
1 code implementation • 3 Jun 2023 • Jintang Li, Wangbin Sun, Ruofan Wu, Yuchang Zhu, Liang Chen, Zibin Zheng
Oversmoothing is a common phenomenon observed in graph neural networks (GNNs), in which an increase in the network depth leads to a deterioration in their performance.
1 code implementation • 30 May 2023 • Jintang Li, Huizhe Zhang, Ruofan Wu, Zulun Zhu, Baokun Wang, Changhua Meng, Zibin Zheng, Liang Chen
While contrastive self-supervised learning has become the de-facto learning paradigm for graph neural networks, the pursuit of higher task accuracy requires a larger hidden dimensionality to learn informative and discriminative full-precision representations, raising concerns about computation, memory footprint, and energy consumption burden (largely overlooked) for real-world applications.
1 code implementation • 29 May 2023 • Peiyi Wang, Lei LI, Liang Chen, Zefan Cai, Dawei Zhu, Binghuai Lin, Yunbo Cao, Qi Liu, Tianyu Liu, Zhifang Sui
In this paper, we uncover a systematic bias in the evaluation paradigm of adopting large language models~(LLMs), e. g., GPT-4, as a referee to score and compare the quality of responses generated by candidate models.
no code implementations • 26 May 2023 • Qichao Wang, Huan Ma, WenTao Wei, Hangyu Li, Liang Chen, Peilin Zhao, Binwen Zhao, Bo Hu, Shu Zhang, Zibin Zheng, Bingzhe Wu
The rapid development of digital economy has led to the emergence of various black and shadow internet industries, which pose potential risks that can be identified and managed through digital risk management (DRM) that uses different techniques such as machine learning and deep learning.
1 code implementation • 25 May 2023 • Ding Wang, Xuhong Wang, Liang Chen, Shengyue Yao, Ming Jing, Honghai Li, Li Li, Shiqiang Bao, Fei-Yue Wang, Yilun Lin
To the best of our knowledge, this is the first traffic simulator that can automatically learn traffic patterns from real-world data and efficiently generate accurate and realistic traffic environments.
no code implementations • 24 May 2023 • Xuhong Wang, Ding Wang, Liang Chen, Yilun Lin
This data-driven and model-free simulation method addresses the challenges faced by traditional systems in terms of structural complexity and model accuracy and provides a foundation for solving complex transportation problems with real data.
1 code implementation • 23 May 2023 • Yang Deng, Lizi Liao, Liang Chen, Hongru Wang, Wenqiang Lei, Tat-Seng Chua
Conversational systems based on Large Language Models (LLMs), such as ChatGPT, show exceptional proficiency in context understanding and response generation.
no code implementations • 23 May 2023 • Ying Huang, Liang Chen
In this note, we consider the problem of robust learning mixtures of linear regressions.
1 code implementation • 23 May 2023 • Sheng Tian, Jihai Dong, Jintang Li, Wenlong Zhao, Xiaolong Xu, Baokun Wang, Bowen Song, Changhua Meng, Tianyi Zhang, Liang Chen
Anomaly detection aims to distinguish abnormal instances that deviate significantly from the majority of benign ones.
1 code implementation • 22 May 2023 • Liang Chen, Hongru Wang, Yang Deng, Wai-Chung Kwan, Zezhong Wang, Kam-Fai Wong
Generating persona consistent dialogue response is important for developing an intelligent conversational agent.
1 code implementation • 18 May 2023 • Jintang Li, Sheng Tian, Ruofan Wu, Liang Zhu, Welong Zhao, Changhua Meng, Liang Chen, Zibin Zheng, Hongzhi Yin
We approach the problem by our proposed STEP, a self-supervised temporal pruning framework that learns to remove potentially redundant edges from input dynamic graphs.
2 code implementations • 18 May 2023 • Liang Chen, Shuming Ma, Dongdong Zhang, Furu Wei, Baobao Chang
We conduct experiments on a multilingual machine translation benchmark in 11 languages.
no code implementations • 26 Apr 2023 • Liang Chen, Juan Jose Dolado, Jesus Gonzalo, Haozi Pan
This paper studies the estimation of characteristic-based quantile factor models where the factor loadings are unknown functions of observed individual characteristics while the idiosyncratic error terms are subject to conditional quantile restrictions.
no code implementations • 25 Apr 2023 • Liang Chen, Minyuan Zhang
This paper focuses on estimating the coefficients and average partial effects of observed regressors in nonlinear panel data models with interactive fixed effects, using the common correlated effects (CCE) framework.
no code implementations • 25 Apr 2023 • Yunpeng Weng, Xing Tang, Liang Chen, Xiuqiang He
For example, in online marketing, the cascade behavior pattern of $impression \rightarrow click \rightarrow conversion$ is usually modeled as multiple tasks in a multi-task manner, where the sequential dependence between tasks is simply connected with an explicitly defined function or implicitly transferred information in current works.
no code implementations • 24 Apr 2023 • Haitian Jiang, Dongliang Xiong, Xiaowen Jiang, Li Ding, Liang Chen, Kai Huang
In this paper, we propose a fast and structure-aware halftoning method via a data-driven approach.
1 code implementation • 13 Apr 2023 • Wenbin Zou, Hongxia Gao, Liang Chen, Yunchen Zhang, Mingchao Jiang, Zhongxin Yu, Ming Tan
Stereo image super-resolution aims to improve the quality of high-resolution stereo image pairs by exploiting complementary information across views.
1 code implementation • CVPR 2023 • Liang Chen, Yong Zhang, Yibing Song, Ying Shan, Lingqiao Liu
Generally, a TTT strategy hinges its performance on two main factors: selecting an appropriate auxiliary TTT task for updating and identifying reliable parameters to update during the test phase.
Ranked #4 on Image to sketch recognition on PACS
Image to sketch recognition Single-Source Domain Generalization +1
no code implementations • 6 Apr 2023 • Wei Yuan, Quoc Viet Hung Nguyen, Tieke He, Liang Chen, Hongzhi Yin
To reveal the real vulnerability of FedRecs, in this paper, we present a new poisoning attack method to manipulate target items' ranks and exposure rates effectively in the top-$K$ recommendation without relying on any prior knowledge.
1 code implementation • 7 Feb 2023 • Dugang Liu, Yang Qiao, Xing Tang, Liang Chen, Xiuqiang He, Weike Pan, Zhong Ming
Specifically, SSTE uses a self-sampling module to generate some subsets with different degrees of bias from the original training and validation data.
1 code implementation • 26 Jan 2023 • Fuyuan Lyu, Xing Tang, Dugang Liu, Liang Chen, Xiuqiang He, Xue Liu
Because of the large-scale search space, we develop a learning-by-continuation training scheme to learn such gates.
Ranked #4 on Click-Through Rate Prediction on KDD12
1 code implementation • 11 Jan 2023 • Ye Huang, Di Kang, Liang Chen, Wenjing Jia, Xiangjian He, Lixin Duan, Xuefei Zhe, Linchao Bao
Extensive experiments and ablation studies conducted on multiple benchmark datasets demonstrate that the proposed CAR can boost the accuracy of all baseline models by up to 2. 23% mIOU with superior generalization ability.
no code implementations • 6 Jan 2023 • Shengyuan Ma, Runke Wang, Suhao Qiu, Ruokun Li, Qi Yue, Qingfang Sun, Liang Chen, Fuhua Yan, Guang-Zhong Yang, Yuan Feng
Here we propose a pipeline for processing MRE images using optimization-based displacement extraction and Traveling Wave Expansion-based Neural Network (TWENN) modulus estimation.
1 code implementation • 20 Nov 2022 • Jintang Li, Jiaying Peng, Liang Chen, Zibin Zheng, TingTing Liang, Qing Ling
In this work, we seek to address these challenges and propose Spectral Adversarial Training (SAT), a simple yet effective adversarial training approach for GNNs.
no code implementations • 9 Nov 2022 • Zishan Gu, Jintang Li, Liang Chen
Graph Neural Networks (GNNs) as deep learning models working on graph-structure data have achieved advanced performance in many works.
1 code implementation • 20 Oct 2022 • Zeyue Xue, Jianming Liang, Guanglu Song, Zhuofan Zong, Liang Chen, Yu Liu, Ping Luo
To address this challenge, we propose a simple yet effective algorithm, named Adaptive Gradient Variance Modulator (AGVM), which can train dense visual predictors with very large batch size, enabling several benefits more appealing than prior arts.
1 code implementation • 29 Sep 2022 • Liang Chen, Bofei Gao, Baobao Chang
In this paper, we provide a detailed description of our system at CAMRP-2022 evaluation.
1 code implementation • 30 Aug 2022 • Jiangbo Pei, Zhuqing Jiang, Aidong Men, Liang Chen, Yang Liu, Qingchao Chen
Secondly, based on the UTR, we propose a novel Calibrated Adaption Framework (CAF) for SFUDA, including i)the source knowledge calibration module that guides the target model to learn the transferable source knowledge and discard the non-transferable one, and ii)the target semantics calibration module that calibrates the unreliable semantics.
1 code implementation • 15 Aug 2022 • Jintang Li, Zhouxin Yu, Zulun Zhu, Liang Chen, Qi Yu, Zibin Zheng, Sheng Tian, Ruofan Wu, Changhua Meng
We explore a new direction in that we can capture the evolving dynamics of temporal graphs with spiking neural networks (SNNs) instead of RNNs.
no code implementations • 28 Jul 2022 • Hanxiao Zhang, Xiao Gu, Minghui Zhang, Weihao Yu, Liang Chen, Zhexin Wang, Feng Yao, Yun Gu, Guang-Zhong Yang
The LIDC-IDRI database is the most popular benchmark for lung cancer prediction.
no code implementations • 20 May 2022 • Bingzhe Wu, Jintang Li, Junchi Yu, Yatao Bian, Hengtong Zhang, Chaochao Chen, Chengbin Hou, Guoji Fu, Liang Chen, Tingyang Xu, Yu Rong, Xiaolin Zheng, Junzhou Huang, Ran He, Baoyuan Wu, Guangyu Sun, Peng Cui, Zibin Zheng, Zhe Liu, Peilin Zhao
Deep graph learning has achieved remarkable progresses in both business and scientific areas ranging from finance and e-commerce, to drug and advanced material discovery.
2 code implementations • 20 May 2022 • Jintang Li, Ruofan Wu, Wangbin Sun, Liang Chen, Sheng Tian, Liang Zhu, Changhua Meng, Zibin Zheng, Weiqiang Wang
The last years have witnessed the emergence of a promising self-supervised learning strategy, referred to as masked autoencoding.
1 code implementation • 5 May 2022 • Zulun Zhu, Jiaying Peng, Jintang Li, Liang Chen, Qi Yu, Siqiang Luo
Graph Convolutional Networks (GCNs) achieve an impressive performance due to the remarkable representation ability in learning the graph information.
1 code implementation • 3 May 2022 • Ruoting Wu, Yuxin Zhang, Qibiao Peng, Liang Chen, Zibin Zheng
In recent years, the rise of deep learning and automation requirements in the software industry has elevated Intelligent Software Engineering to new heights.
no code implementations • 2 May 2022 • Yuansheng Wang, Wangbin Sun, Kun Xu, Zulun Zhu, Liang Chen, Zibin Zheng
Graph contrastive learning (GCL), as a popular approach to graph self-supervised learning, has recently achieved a non-negligible effect.
1 code implementation • 20 Apr 2022 • Jintang Li, Jie Liao, Ruofan Wu, Liang Chen, Zibin Zheng, Jiawang Dan, Changhua Meng, Weiqiang Wang
To mitigate such a threat, considerable research efforts have been devoted to increasing the robustness of GCNs against adversarial attacks.
2 code implementations • Findings (NAACL) 2022 • Liang Chen, Peiyi Wang, Runxin Xu, Tianyu Liu, Zhifang Sui, Baobao Chang
As Abstract Meaning Representation (AMR) implicitly involves compound semantic annotations, we hypothesize auxiliary tasks which are semantically or formally related can better enhance AMR parsing.
Ranked #7 on AMR Parsing on LDC2020T02 (using extra training data)
1 code implementation • 19 Apr 2022 • Wenbin Zou, Tian Ye, Weixin Zheng, Yunchen Zhang, Liang Chen, Yi Wu
Recently, deep learning has been successfully applied to the single-image super-resolution (SISR) with remarkable performance.
1 code implementation • CVPR 2022 • Liang Chen, Yong Zhang, Yibing Song, Lingqiao Liu, Jue Wang
Following this principle, we propose to enrich the "diversity" of forgeries by synthesizing augmented forgeries with a pool of forgery configurations and strengthen the "sensitivity" to the forgeries by enforcing the model to predict the forgery configurations.
no code implementations • 16 Mar 2022 • Liang Chen, Sue Ann Campbell
Networks of spiking neurons with adaption have been shown to be able to reproduce a wide range of neural activities, including the emergent population bursting and spike synchrony that underpin brain disorders and normal function.
1 code implementation • arXiv:2203.07160 2022 • Ye Huang, Di Kang, Liang Chen, Xuefei Zhe, Wenjing Jia, Xiangjian He, Linchao Bao
Recent segmentation methods, such as OCR and CPNet, utilizing "class level" information in addition to pixel features, have achieved notable success for boosting the accuracy of existing network modules.
Ranked #8 on Semantic Segmentation on PASCAL Context
2 code implementations • 6 Mar 2022 • Liang Chen, Runxin Xu, Baobao Chang
Label smoothing and vocabulary sharing are two widely used techniques in neural machine translation models.
no code implementations • 25 Feb 2022 • Hanxiao Zhang, Liang Chen, Xiao Gu, Minghui Zhang, Yulei Qin, Feng Yao, Zhexin Wang, Yun Gu, Guang-Zhong Yang
In this study, we construct a sure dataset with pathologically-confirmed labels and propose a collaborative learning framework to facilitate sure nodule classification by integrating unsure data knowledge through nodule segmentation and malignancy score regression.
no code implementations • 17 Feb 2022 • Lesley Tan, Liang Chen
We propose new Enhanced DeepONet or EDeepONet high-level neural network structure, in which two input functions are represented by two branch DNN sub-networks, which are then connected with output truck network via inner product to generate the output of the whole neural network.
no code implementations • 15 Feb 2022 • Jintang Li, Bingzhe Wu, Chengbin Hou, Guoji Fu, Yatao Bian, Liang Chen, Junzhou Huang, Zibin Zheng
Despite the progress, applying DGL to real-world applications faces a series of reliability threats including inherent noise, distribution shift, and adversarial attacks.
no code implementations • 4 Feb 2022 • Liang Chen, Yang Li, Manyun Huang, Xinxin Hui, Songlin Gu
A novel robust dynamic state estimation methodology for integrated natural gas and electric power systems is proposed based on Kalman filter.
no code implementations • 17 Jan 2022 • Liang Chen, Qibiao Peng, Jintang Li, Yang Liu, Jiawei Chen, Yong Li, Zibin Zheng
To address such a challenge, we set the trigger as a single node, and the backdoor is activated when the trigger node is connected to the target node.
no code implementations • CVPR 2022 • Liang Chen, Yihang Lou, Jianzhong He, Tao Bai, Minghua Deng
Therefore, in this paper, we propose a Geometric anchor-guided Adversarial and conTrastive learning framework with uncErtainty modeling called GATE to alleviate these issues.
Ranked #5 on Universal Domain Adaptation on Office-Home
no code implementations • 22 Dec 2021 • Liang Chen, Xin Zhou, Feifei Chen, Lie-Liang Yang, Ruizhi Chen
Indoor positioning is one of the core technologies of Internet of Things (IoT) and artificial intelligence (AI), and is expected to play a significant role in the upcoming era of AI.
1 code implementation • 18 Nov 2021 • Tian Ye, Mingchao Jiang, Yunchen Zhang, Liang Chen, ErKang Chen, Pen Chen, Zhiyong Lu
However, due to the paradox caused by the variation of real captured haze and the fixed degradation parameters of the current networks, the generalization ability of recent dehazing methods on real-world hazy images is not ideal. To address the problem of modeling real-world haze degradation, we propose to solve this problem by perceiving and modeling density for uneven haze distribution.
Ranked #7 on Image Dehazing on Haze4k
1 code implementation • ACL 2022 • Peiyi Wang, Liang Chen, Tianyu Liu, Damai Dai, Yunbo Cao, Baobao Chang, Zhifang Sui
Abstract Meaning Representation (AMR) parsing aims to translate sentences to semantic representation with a hierarchical structure, and is recently empowered by pretrained sequence-to-sequence models.
1 code implementation • 12 Oct 2021 • Wenbin Zou, Mingchao Jiang, Yunchen Zhang, Liang Chen, Zhiyong Lu, Yi Wu
On this basis, we reduce the number of up-sampling and down-sampling and design a simple network structure.
Ranked #1 on Image Deblurring on RealBlur-R(trained on GoPro)
1 code implementation • 25 Aug 2021 • Xuan Wu, Jizong Han, Di Wang, Pengyue Gao, Quanlong Cui, Liang Chen, Yanchun Liang, Han Huang, Heow Pueh Lee, Chunyan Miao, You Zhou, Chunguo Wu
While many Particle Swarm Optimization (PSO) algorithms only use fitness to assess the performance of particles, in this work, we adopt Surprisingly Popular Algorithm (SPA) as a complementary metric in addition to fitness.
2 code implementations • 16 Aug 2021 • Yu Zheng, Chen Gao, Liang Chen, Depeng Jin, Yong Li
These years much effort has been devoted to improving the accuracy or relevance of the recommendation system.
1 code implementation • 13 Aug 2021 • Liang Chen, Jintang Li, Qibiao Peng, Yang Liu, Zibin Zheng, Carl Yang
In this work, we theoretically and empirically demonstrate that structural adversarial examples can be attributed to the non-robust aggregation scheme (i. e., the weighted mean) of GCNs.
1 code implementation • 7 Aug 2021 • Chen Chen, Chen Qin, Cheng Ouyang, Zeju Li, Shuo Wang, Huaqi Qiu, Liang Chen, Giacomo Tarroni, Wenjia Bai, Daniel Rueckert
The success of neural networks on medical image segmentation tasks typically relies on large labeled datasets for model training.
no code implementations • 7 Aug 2021 • Shengsen Wu, Liang Chen, Yihang Lou, Yan Bai, Tao Bai, Minghua Deng, Lingyu Duan
Therefore, backward-compatible representation is proposed to enable "new" features to be compared with "old" features directly, which means that the database is active when there are both "new" and "old" features in it.
no code implementations • 27 Jul 2021 • Liang Chen, Lesley Tan
In the second method, a third-order polynomial expression is defined first, which ensures positiveness, to approximate the insertion loss, then DeepONet neural network structure, which was proposed recently for function and system modeling, was employed to model the coefficients of polynomials.
no code implementations • 13 Jul 2021 • Liang Chen, Xinxin Hui, Songlin Gu, Manyun Huang, Yang Li
Boundary conditions of pipeline networks are used as supplementary constraints in the system model.
no code implementations • CVPR 2021 • Liang Chen, Jiawei Zhang, Songnan Lin, Faming Fang, Jimmy S. Ren
To address this problem, we introduce a new blur model to fit both saturated and unsaturated pixels, and all informative pixels can be considered during deblurring process.
no code implementations • CVPR 2021 • Liang Chen, Jiawei Zhang, Jinshan Pan, Songnan Lin, Faming Fang, Jimmy S. Ren
Deblurring night blurry images is difficult, because the common-used blur model based on the linear convolution operation does not hold in this situation due to the influence of saturated pixels.
1 code implementation • 2 Jun 2021 • Yuhang Guo, Xiao Luo, Liang Chen, Minghua Deng
Predicting DNA-protein binding is an important and classic problem in bioinformatics.
no code implementations • 31 May 2021 • Liang Chen, Xiangchen Lu, Nan Shen, Lei Wang, Yuan Zhuang, Ye Su, Deren Li, Ruizhi Chen
The performance of those integration algorithms on expanding the successful acquisition time range is verified by the real data collected from the Luojia-1A satellite.
1 code implementation • Findings (ACL) 2021 • Yulong Chen, Yang Liu, Liang Chen, Yue Zhang
Proposal of large-scale datasets has facilitated research on deep neural models for news summarization.
Abstractive Dialogue Summarization Common Sense Reasoning +3
1 code implementation • 10 May 2021 • Jiawei Chen, Hande Dong, Yang Qiu, Xiangnan He, Xin Xin, Liang Chen, Guli Lin, Keping Yang
This provides a valuable opportunity to develop a universal solution for debiasing, e. g., by learning the debiasing parameters from data.
no code implementations • 12 Apr 2021 • Qilin Deng, Kai Wang, Minghao Zhao, Zhene Zou, Runze Wu, Jianrong Tao, Changjie Fan, Liang Chen
In business domains, \textit{bundling} is one of the most important marketing strategies to conduct product promotions, which is commonly used in online e-commerce and offline retailers.
no code implementations • 7 Apr 2021 • Kai Wang, Zhene Zou, Qilin Deng, Runze Wu, Jianrong Tao, Changjie Fan, Liang Chen, Peng Cui
As a part of the value function, free from the sparse and high-variance reward signals, a high-capacity reward-independent world model is trained to simulate complex environmental dynamics under a certain goal.
Model-based Reinforcement Learning Recommendation Systems +3
no code implementations • 12 Mar 2021 • Xuan Wu, Linhan Jia, Xiuyi Zhang, Liang Chen, Yanchun Liang, You Zhou, Chunguo Wu
To evolve the architectures under the framework of CGP, the operations such as convolution are identified as the types of function nodes of CGP, and the evolutionary operations are designed based on Evolutionary Strategy.
no code implementations • 26 Feb 2021 • Liang Chen, Peng Jin, Jing Yang, Yang Li, Yi Song
To obtain the accurate transient states of the big scale natural gas pipeline networks under the bad data and non-zero mean noises conditions, a robust Kalman filter-based dynamic state estimation method is proposed using the linearized gas pipeline transient flow equations in this paper.
no code implementations • 22 Feb 2021 • Liang Chen
In this paper, we theoretically propose a new hashing scheme to establish the sparse Fourier transform in high-dimensional space.
Data Structures and Algorithms Information Theory Numerical Analysis Information Theory Numerical Analysis 41A63 G.1.2; F.2.1
1 code implementation • 16 Feb 2021 • Jintang Li, Kun Xu, Liang Chen, Zibin Zheng, Xiao Liu
Graph Neural Networks (GNNs) have recently shown to be powerful tools for representing and analyzing graph data.
no code implementations • 12 Feb 2021 • Liang Chen, Hakan Bagci
In this work, an optical absorption-based model is proposed to accurately calculate the generation rate in time-domain simulations.
Optics Computational Engineering, Finance, and Science Computational Physics
no code implementations • 11 Feb 2021 • Jianeng Zhou, Zhongxiang Wang, Feng Huang, Liang Chen
We investigate this possibility by fitting the spectrum with the photon-ALP oscillation model, and find that the parameter space of ALP mass $m_a\leq 10^{-8}$\, eV and the coupling constant (between photons and ALPs) $g_{a\gamma}$=1. 16--1. 48$\times 10^{-10}$\, GeV$^{-1}$ can provide a fit to the line-like feature, while the magnetic field at the emission site of $\gamma$-rays is fixed at 0. 7\, G.
High Energy Astrophysical Phenomena
no code implementations • 4 Feb 2021 • Ting-Yi Wu, Yunghsiang S. Han, Zhengrui Li, Bo Bai, Gong Zhang, Liang Chen, Xiang Wu
Accessing the data in the failed disk (degraded read) with low latency is crucial for an erasure-coded storage system.
Information Theory Information Theory
no code implementations • 3 Feb 2021 • Liang Chen, Eugene Choo, Alfred Galichon, Simon Weber
We argue that models coming from a variety of fields, such as matching models and discrete choice models among others, share a common structure that we call matching function equilibria with partial assignment.
no code implementations • 21 Jan 2021 • Qiong Wu, Xu Chen, Zhi Zhou, Liang Chen, Junshan Zhang
To meet the ever increasing mobile traffic demand in 5G era, base stations (BSs) have been densely deployed in radio access networks (RANs) to increase the network coverage and capacity.
no code implementations • 27 Nov 2020 • Xiangchen Lu, Liang Chen, Nan Shen, Lei Wang, Zhenhang Jiao, Ruizhi Chen
With the rapid development of China's BeiDou Navigation Satellite System(BDS), the application of real-time precise point positioning (RTPPP) based on BDS has become an active research area in the field of Global Navigation Satellite System (GNSS).
no code implementations • 24 Nov 2020 • Shuyang Sun, Liang Chen, Gregory Slabaugh, Philip Torr
Some image restoration tasks like demosaicing require difficult training samples to learn effective models.
no code implementations • 8 Nov 2020 • Jieming Zhu, Jinyang Liu, Weiqi Li, Jincai Lai, Xiuqiang He, Liang Chen, Zibin Zheng
Recently, deep learning-based models have been widely studied for click-through rate (CTR) prediction and lead to improved prediction accuracy in many industrial applications.
no code implementations • 3 Nov 2020 • Yunpeng Weng, Xu Chen, Liang Chen, Wei Liu
Most existing GNN models exploit a single type of aggregator (e. g., mean-pooling) to aggregate neighboring nodes information, and then add or concatenate the output of aggregator to the current representation vector of the center node.
3 code implementations • 21 Oct 2020 • Jiancan Wu, Xiang Wang, Fuli Feng, Xiangnan He, Liang Chen, Jianxun Lian, Xing Xie
In this work, we explore self-supervised learning on user-item graph, so as to improve the accuracy and robustness of GCNs for recommendation.
Ranked #6 on Collaborative Filtering on Yelp2018
no code implementations • 21 Sep 2020 • Liang Chen
In this paper, we study a class of time-periodic stochastic Tonelli Lagrangians on compact manifolds.
Dynamical Systems
3 code implementations • 15 Sep 2020 • Kai Zhang, Martin Danelljan, Yawei Li, Radu Timofte, Jie Liu, Jie Tang, Gangshan Wu, Yu Zhu, Xiangyu He, Wenjie Xu, Chenghua Li, Cong Leng, Jian Cheng, Guangyang Wu, Wenyi Wang, Xiaohong Liu, Hengyuan Zhao, Xiangtao Kong, Jingwen He, Yu Qiao, Chao Dong, Maitreya Suin, Kuldeep Purohit, A. N. Rajagopalan, Xiaochuan Li, Zhiqiang Lang, Jiangtao Nie, Wei Wei, Lei Zhang, Abdul Muqeet, Jiwon Hwang, Subin Yang, JungHeum Kang, Sung-Ho Bae, Yongwoo Kim, Geun-Woo Jeon, Jun-Ho Choi, Jun-Hyuk Kim, Jong-Seok Lee, Steven Marty, Eric Marty, Dongliang Xiong, Siang Chen, Lin Zha, Jiande Jiang, Xinbo Gao, Wen Lu, Haicheng Wang, Vineeth Bhaskara, Alex Levinshtein, Stavros Tsogkas, Allan Jepson, Xiangzhen Kong, Tongtong Zhao, Shanshan Zhao, Hrishikesh P. S, Densen Puthussery, Jiji C. V, Nan Nan, Shuai Liu, Jie Cai, Zibo Meng, Jiaming Ding, Chiu Man Ho, Xuehui Wang, Qiong Yan, Yuzhi Zhao, Long Chen, Jiangtao Zhang, Xiaotong Luo, Liang Chen, Yanyun Qu, Long Sun, Wenhao Wang, Zhenbing Liu, Rushi Lan, Rao Muhammad Umer, Christian Micheloni
This paper reviews the AIM 2020 challenge on efficient single image super-resolution with focus on the proposed solutions and results.
1 code implementation • 8 Sep 2020 • Jintang Li, Tao Xie, Liang Chen, Fenfang Xie, Xiangnan He, Zibin Zheng
Currently, most works on attacking GNNs are mainly using gradient information to guide the attack and achieve outstanding performance.
no code implementations • 12 Aug 2020 • Liang Chen
In particular, we first propose a principled gradient variance decomposition theorem, which shows that the variance of the stochastic gradient of the language pretraining can be naturally decomposed into two terms: the variance that arises from the sample of data in a batch, and the variance that arises from the sampling of the mask.
no code implementations • 1 Jul 2020 • Wenqiang Lei, Gangyi Zhang, Xiangnan He, Yisong Miao, Xiang Wang, Liang Chen, Tat-Seng Chua
Traditional recommendation systems estimate user preference on items from past interaction history, thus suffering from the limitations of obtaining fine-grained and dynamic user preference.
1 code implementation • 23 Jun 2020 • Chen Chen, Chen Qin, Huaqi Qiu, Cheng Ouyang, Shuo Wang, Liang Chen, Giacomo Tarroni, Wenjia Bai, Daniel Rueckert
In this work, we propose an adversarial data augmentation method for training neural networks for medical image segmentation.
2 code implementations • 31 May 2020 • Liang Chen, Yanchun Liang, Xiaohu Shi, You Zhou, Chunguo Wu
Time Delay Neural Network (TDNN) is a well-performing structure for DNN-based speaker recognition systems.