no code implementations • 14 Aug 2024 • Vibhor Agarwal, Yulong Pei, Salwa Alamir, Xiaomo Liu
We propose the first benchmark CodeMirage dataset for code hallucinations.
no code implementations • 14 Aug 2024 • Ricky Maulana Fajri, Yulong Pei, Lu Yin, Mykola Pechenizkiy
Despite significant advancements in active learning and adversarial attacks, the intersection of these two fields remains underexplored, particularly in developing robust active learning frameworks against dynamic adversarial threats.
no code implementations • 24 Jul 2024 • Tianjin Huang, Fang Meng, Li Shen, Fan Liu, Yulong Pei, Mykola Pechenizkiy, Shiwei Liu, Tianlong Chen
In this paper, we investigate a charming possibility - \textit{leveraging visual prompts to capture the channel importance and derive high-quality structural sparsity}.
no code implementations • 5 Apr 2024 • Ran Zmigrod, Dongsheng Wang, Mathieu Sibue, Yulong Pei, Petr Babkin, Ivan Brugere, Xiaomo Liu, Nacho Navarro, Antony Papadimitriou, William Watson, Zhiqiang Ma, Armineh Nourbakhsh, Sameena Shah
Several datasets exist for research on specific tasks of VRDU such as document classification (DC), key entity extraction (KEE), entity linking, visual question answering (VQA), inter alia.
no code implementations • 31 Dec 2023 • Dongsheng Wang, Natraj Raman, Mathieu Sibue, Zhiqiang Ma, Petr Babkin, Simerjot Kaur, Yulong Pei, Armineh Nourbakhsh, Xiaomo Liu
Enterprise documents such as forms, invoices, receipts, reports, contracts, and other similar records, often carry rich semantics at the intersection of textual and spatial modalities.
1 code implementation • 7 Dec 2023 • Ricky Maulana Fajri, Yulong Pei, Lu Yin, Mykola Pechenizkiy
To address this problem, we propose the Structural-Clustering PageRank method for improved Active learning (SPA) specifically designed for graph-structured data.
no code implementations • 12 Oct 2023 • Zirui Liang, Yuntao Li, Tianjin Huang, Akrati Saxena, Yulong Pei, Mykola Pechenizkiy
This leads to suboptimal performance of standard GNNs on imbalanced graphs.
no code implementations • 12 Oct 2023 • Ethan Callanan, Amarachi Mbakwe, Antony Papadimitriou, Yulong Pei, Mathieu Sibue, Xiaodan Zhu, Zhiqiang Ma, Xiaomo Liu, Sameena Shah
Large Language Models (LLMs) have demonstrated remarkable performance on a wide range of Natural Language Processing (NLP) tasks, often matching or even beating state-of-the-art task-specific models.
no code implementations • 15 Aug 2023 • Tianchao Li, Yulong Pei
This paper is to introduce an asynchronous and local learning framework for neural networks, named Modular Learning Framework (MOLE).
1 code implementation • 25 Jun 2023 • Tianjin Huang, Shiwei Liu, Tianlong Chen, Meng Fang, Li Shen, Vlaod Menkovski, Lu Yin, Yulong Pei, Mykola Pechenizkiy
Despite the fact that adversarial training has become the de facto method for improving the robustness of deep neural networks, it is well-known that vanilla adversarial training suffers from daunting robust overfitting, resulting in unsatisfactory robust generalization.
1 code implementation • 23 May 2023 • Yirui Liu, Xinghao Qiao, Yulong Pei, Liying Wang
This paper introduces the Deep Functional Factor Model (DF2M), a Bayesian nonparametric model designed for analysis of high-dimensional functional time series.
no code implementations • 10 May 2023 • Xianzhi Li, Samuel Chan, Xiaodan Zhu, Yulong Pei, Zhiqiang Ma, Xiaomo Liu, Sameena Shah
The most recent large language models(LLMs) such as ChatGPT and GPT-4 have shown exceptional capabilities of generalist models, achieving state-of-the-art performance on a wide range of NLP tasks with little or no adaptation.
Ranked #1 on
Question Answering
on ConvFinQA
1 code implementation • 28 Nov 2022 • Tianjin Huang, Tianlong Chen, Meng Fang, Vlado Menkovski, Jiaxu Zhao, Lu Yin, Yulong Pei, Decebal Constantin Mocanu, Zhangyang Wang, Mykola Pechenizkiy, Shiwei Liu
Recent works have impressively demonstrated that there exists a subnetwork in randomly initialized convolutional neural networks (CNNs) that can match the performance of the fully trained dense networks at initialization, without any optimization of the weights of the network (i. e., untrained networks).
1 code implementation • 21 Sep 2022 • Ricky Fajri, Akrati Saxena, Yulong Pei, Mykola Pechenizkiy
Active Learning (AL) techniques have proven to be highly effective in reducing data labeling costs across a range of machine learning tasks.
no code implementations • 30 May 2022 • Lu Yin, Vlado Menkovski, Meng Fang, Tianjin Huang, Yulong Pei, Mykola Pechenizkiy, Decebal Constantin Mocanu, Shiwei Liu
Recent works on sparse neural network training (sparse training) have shown that a compelling trade-off between performance and efficiency can be achieved by training intrinsically sparse neural networks from scratch.
no code implementations • 16 Dec 2021 • Lu Yin, Vlado Menkovski, Yulong Pei, Mykola Pechenizkiy
In this work, we advance the few-shot learning towards this more challenging scenario, the semantic-based few-shot learning, and propose a method to address the paradigm by capturing the inner semantic relationships using interactive psychometric learning.
1 code implementation • 11 Dec 2021 • Zeyu Zhang, Yulong Pei
Although a series of robust GNNs have been proposed, they are evaluated with different structural noises, and it lacks a systematic comparison with consistent settings.
1 code implementation • 1 Oct 2021 • Tianjin Huang, Vlado Menkovski, Yulong Pei, Mykola Pechenizkiy
In this paper, we present the Calibrated Adversarial Training, a method that reduces the adverse effects of semantic perturbations in adversarial training.
1 code implementation • 18 Jul 2021 • Pengfei Jiao, Xuan Guo, Ting Pan, Wang Zhang, Yulong Pei
A wide variety of NE methods focus on the proximity of networks.
1 code implementation • 6 Jul 2021 • Tianjin Huang, Yulong Pei, Vlado Menkovski, Mykola Pechenizkiy
Adversarial training is an approach for increasing model's resilience against adversarial perturbations.
1 code implementation • 19 Apr 2021 • Tianjin Huang, Vlado Menkovski, Yulong Pei, Yuhao Wang, Mykola Pechenizkiy
Deep neural networks are vulnerable to adversarial examples that are crafted by imposing imperceptible changes to the inputs.
1 code implementation • 16 Apr 2021 • Tianjin Huang, Yulong Pei, Vlado Menkovski, Mykola Pechenizkiy
Although various approaches have been proposed to solve this problem, two major limitations exist: (1) unsupervised approaches usually work much less efficiently due to the lack of supervisory signal, and (2) existing anomaly detection methods only use local contextual information to detect anomalous nodes, e. g., one- or two-hop information, but ignore the global contextual information.
Self-Supervised Anomaly Detection
Supervised Anomaly Detection
1 code implementation • 22 Jan 2021 • Shiwei Liu, Decebal Constantin Mocanu, Yulong Pei, Mykola Pechenizkiy
Sparse neural networks have been widely applied to reduce the computational demands of training and deploying over-parameterized deep neural networks.
1 code implementation • 7 Nov 2020 • Tianjin Huang, Vlado Menkovski, Yulong Pei, Mykola Pechenizkiy
In addition, it achieves comparable performance of adversarial robustness on MNIST dataset under white-box attack, and it achieves better performance than adv. PGD under white-box attack and effectively defends the transferable adversarial attack on CIFAR-10 dataset.
1 code implementation • 30 Sep 2020 • Yulong Pei, Tianjin Huang, Werner van Ipenburg, Mykola Pechenizkiy
Effectively detecting anomalous nodes in attributed networks is crucial for the success of many real-world applications such as fraud and intrusion detection.
no code implementations • 24 Dec 2019 • Rahul Radhakrishnan Iyer, Yulong Pei, Katia Sycara
Tweet classification has attracted considerable attention recently.
4 code implementations • 26 Jan 2019 • Shiwei Liu, Decebal Constantin Mocanu, Amarsagar Reddy Ramapuram Matavalam, Yulong Pei, Mykola Pechenizkiy
Despite the success of ANNs, it is challenging to train and deploy modern ANNs on commodity hardware due to the ever-increasing model size and the unprecedented growth in the data volumes.
no code implementations • 25 May 2018 • Yulong Pei, Xin Du, Jianpeng Zhang, George Fletcher, Mykola Pechenizkiy
Almost all previous methods represent a node into a point in space and focus on local structural information, i. e., neighborhood information.