no code implementations • 28 May 2018 • Yang Yang, Haoyan Liu, Xia Hu, Jiawei Zhang, Xiao-Ming Zhang, Zhoujun Li, Philip S. Yu
The number of missing people (i. e., people who get lost) greatly increases in recent years.
no code implementations • 23 Mar 2018 • Bokai Cao, Lei Zheng, Chenwei Zhang, Philip S. Yu, Andrea Piscitello, John Zulueta, Olu Ajilore, Kelly Ryan, Alex D. Leow
The increasing use of electronic forms of communication presents new opportunities in the study of mental health, including the ability to investigate the manifestations of psychiatric diseases unobtrusively and in the setting of patients' daily lives.
no code implementations • 12 Mar 2018 • He Huang, Philip S. Yu, Changhu Wang
There has been a drastic growth of research in Generative Adversarial Nets (GANs) in the past few years.
no code implementations • 19 Jan 2018 • Shaika Chowdhury, Chenwei Zhang, Philip S. Yu
Social media has grown to be a crucial information source for pharmacovigilance studies where an increasing number of people post adverse reactions to medical drugs that are previously unreported.
no code implementations • 10 Apr 2017 • Chun-Ta Lu, Lifang He, Hao Ding, Bokai Cao, Philip S. Yu
Real-world relations among entities can often be observed and determined by different perspectives/views.
no code implementations • 1 Jan 2018 • Mehrnaz Najafi, Lifang He, Philip S. Yu
Various types of errors behave differently and inconsistently in each view.
no code implementations • 25 Dec 2017 • Jindong Wang, Yiqiang Chen, Lisha Hu, Xiaohui Peng, Philip S. Yu
The proposed framework, referred to as Stratified Transfer Learning (STL), can dramatically improve the classification accuracy for cross-domain activity recognition.
no code implementations • 6 Dec 2017 • Hu Xu, Sihong Xie, Lei Shu, Philip S. Yu
Functionality is of utmost importance to customers when they purchase products.
no code implementations • 6 Dec 2017 • Hu Xu, Sihong Xie, Lei Shu, Philip S. Yu
Product compatibility and their functionality are of utmost importance to customers when they purchase products, and to sellers and manufacturers when they sell products.
no code implementations • NeurIPS 2017 • Mingsheng Long, Zhangjie Cao, Jian-Min Wang, Philip S. Yu
Deep networks trained on large-scale data can learn transferable features to promote learning multiple tasks.
no code implementations • 22 Oct 2017 • Chenwei Zhang, Nan Du, Wei Fan, Yaliang Li, Chun-Ta Lu, Philip S. Yu
The healthcare status, complex medical information needs of patients are expressed diversely and implicitly in their medical text queries.
no code implementations • 13 Sep 2017 • Bokai Cao, Mia Mao, Siim Viidu, Philip S. Yu
On electronic game platforms, different payment transactions have different levels of risk.
no code implementations • 12 Sep 2017 • Guixiang Ma, Chun-Ta Lu, Lifang He, Philip S. Yu, Ann B. Ragin
Specifically, we propose an auto-weighted framework of Multi-view Graph Embedding with Hub Detection (MVGE-HD) for brain network analysis.
no code implementations • 12 Jun 2017 • Vahid Noroozi, Lei Zheng, Sara Bahaadini, Sihong Xie, Philip S. Yu
The model consists of two complementary components.
no code implementations • 29 May 2017 • Hu Xu, Lei Shu, Philip S. Yu
Extracting opinion targets is an important task in sentiment analysis on product reviews and complementary entities (products) are one important type of opinion targets that may work together with the reviewed product.
no code implementations • 2 May 2017 • Xiaokai Wei, Bokai Cao, Philip S. Yu
In this paper, we study unsupervised feature selection for multi-view data, as class labels are usually expensive to obtain.
no code implementations • 22 Feb 2016 • Yue Cao, Mingsheng Long, Jian-Min Wang, Philip S. Yu
This paper presents a Correlation Hashing Network (CHN) approach to cross-modal hashing, which jointly learns good data representation tailored to hash coding and formally controls the quantization error.
no code implementations • 14 Dec 2016 • Hu Xu, Lei Shu, Jingyuan Zhang, Philip S. Yu
In this paper, we address the problem of extracting compatible and incompatible products from yes/no questions in PCQA.
no code implementations • 4 Dec 2016 • Hu Xu, Sihong Xie, Lei Shu, Philip S. Yu
One important product feature is the complementary entity (products) that may potentially work together with the reviewed product.
no code implementations • 2 Nov 2016 • Weixiang Shao, Lifang He, Chun-Ta Lu, Philip S. Yu
We model the multi-view clustering problem as a joint weighted nonnegative matrix factorization problem and process the multi-view data chunk by chunk to reduce the memory requirement.
no code implementations • 27 Sep 2016 • Weixiang Shao, Lifang He, Chun-Ta Lu, Xiaokai Wei, Philip S. Yu
Third, how to leverage the consistent and complementary information from different views to improve the feature selection in the situation when the data are too big or come in as streams?
no code implementations • 11 Aug 2016 • Chenwei Zhang, Sihong Xie, Yaliang Li, Jing Gao, Wei Fan, Philip S. Yu
We propose a novel multi-source hierarchical prediction consolidation method to effectively exploits the complicated hierarchical label structures to resolve the noisy and conflicting information that inherently originates from multiple imperfect sources.
no code implementations • 4 Oct 2013 • Weixiang Shao, Xiaoxiao Shi, Philip S. Yu
The idea is to collectively completes the kernel matrices of incomplete datasets by optimizing the alignment of the shared instances of the datasets.
no code implementations • 14 Apr 2016 • Weixiang Shao, Jiawei Zhang, Lifang He, Philip S. Yu
In many real-world applications, information can be gathered from multiple sources, while each source can contain multiple views, which are more cohesive for learning.
no code implementations • 3 Apr 2016 • Jiawei Zhang, Xiao Pan, Moyin Li, Philip S. Yu
In bicycle-sharing systems, people can borrow and return bikes at any stations in the service region very conveniently.
no code implementations • 19 Aug 2015 • Bokai Cao, Xiangnan Kong, Jingyuan Zhang, Philip S. Yu, Ann B. Ragin
In this paper, we study the problem of discriminative subgraph selection using multiple side views and propose a novel solution to find an optimal set of subgraph features for graph classification by exploring a plurality of side views.
no code implementations • 5 Aug 2015 • Bokai Cao, Xiangnan Kong, Philip S. Yu
Brain disorder data poses many unique challenges for data mining research.
no code implementations • 31 Jul 2014 • Lifang He, Xiangnan Kong, Philip S. Yu, Ann B. Ragin, Zhifeng Hao, Xiaowei Yang
The dual-tensorial mapping function can map each tensor instance in the input space to another tensor in the feature space while preserving the tensorial structure.
no code implementations • 6 Jul 2014 • Xiangnan Kong, Zhaoming Wu, Li-Jia Li, Ruofei Zhang, Philip S. Yu, Hang Wu, Wei Fan
Unlike prior works, our method can effectively and efficiently consider missing labels and label correlations simultaneously, and is very scalable, that has linear time complexities over the size of the data.
no code implementations • 16 Oct 2013 • Sihong Xie, Xiangnan Kong, Jing Gao, Wei Fan, Philip S. Yu
Nonetheless, data nowadays are usually multilabeled, such that more than one label have to be predicted at the same time.
no code implementations • 13 Oct 2013 • Jiawei Zhang, Xiangnan Kong, Philip S. Yu
We propose a link prediction method called SCAN-PS (Supervised Cross Aligned Networks link prediction with Personalized Sampling), to solve the link prediction problem for new users with information transferred from both the existing active users in the target network and other source networks through aligned accounts.
no code implementations • 28 Sep 2013 • Chuan Shi, Xiangnan Kong, Yue Huang, Philip S. Yu, Bin Wu
Similarity search is an important function in many applications, which usually focuses on measuring the similarity between objects with the same type.
no code implementations • 26 Sep 2013 • Zhung-Xun Liao, Shou-Chung Li, Wen-Chih Peng, Philip S. Yu
By analyzing real App usage log data, we discover two kinds of features: The Explicit Feature (EF) from sensing readings of built-in sensors, and the Implicit Feature (IF) from App usage relations.
no code implementations • 20 May 2013 • Xiangnan Kong, Bokai Cao, Philip S. Yu, Ying Ding, David J. Wild
Moreover, by considering different linkage paths in the network, one can capture the subtlety of different types of dependencies among objects.
no code implementations • 19 Jun 2018 • Ye Liu, Lifang He, Bokai Cao, Philip S. Yu, Ann B. Ragin, Alex D. Leow
Network analysis of human brain connectivity is critically important for understanding brain function and disease states.
no code implementations • 26 Nov 2017 • Jiawei Zhang, Congying Xia, Chenwei Zhang, Limeng Cui, Yanjie Fu, Philip S. Yu
The closeness among users in the networks are defined as the meta proximity scores, which will be fed into DIME to learn the embedding vectors of users in the emerging network.
Social and Information Networks Databases
no code implementations • 2 Sep 2018 • Xi Zhang, Yixuan Li, Senzhang Wang, Binxing Fang, Philip S. Yu
In this work, we study how to explore multiple data sources to improve the performance of the stock prediction.
no code implementations • 10 Sep 2018 • Ji Wang, Jian-Guo Zhang, Weidong Bao, Xiaomin Zhu, Bokai Cao, Philip S. Yu
To benefit from the cloud data center without the privacy risk, we design, evaluate, and implement a cloud-based framework ARDEN which partitions the DNN across mobile devices and cloud data centers.
no code implementations • 10 Sep 2018 • Ji Wang, Bokai Cao, Philip S. Yu, Lichao Sun, Weidong Bao, Xiaomin Zhu
In this paper, we provide an overview of the current challenges and representative achievements about pushing deep learning on mobile devices from three aspects: training with mobile data, efficient inference on mobile devices, and applications of mobile deep learning.
no code implementations • 11 Sep 2018 • Jian-Guo Zhang, Ji Wang, Lifang He, Zhao Li, Philip S. Yu
Then, it is possible to utilize unlabeled data that have a potential of failure to further improve the performance of the model.
no code implementations • 11 Sep 2018 • Lichao Sun, Lifang He, Zhipeng Huang, Bokai Cao, Congying Xia, Xiaokai Wei, Philip S. Yu
Meta-graph is currently the most powerful tool for similarity search on heterogeneous information networks, where a meta-graph is a composition of meta-paths that captures the complex structural information.
no code implementations • 18 Oct 2018 • Lifang He, Chun-Ta Lu, Yong Chen, Jiawei Zhang, Linlin Shen, Philip S. Yu, Fei Wang
In many real-world applications, data are often unlabeled and comprised of different representations/views which often provide information complementary to each other.
no code implementations • 17 Oct 2018 • Jianguo Chen, Kenli Li, Huigui Rong, Kashif Bilal, Keqin Li, Philip S. Yu
In this paper, a Periodicity-based Parallel Time Series Prediction (PPTSP) algorithm for large-scale time-series data is proposed and implemented in the Apache Spark cloud computing environment.
no code implementations • 17 Oct 2018 • Jianguo Chen, Kenli Li, Kashif Bilal, Xu Zhou, Keqin Li, Philip S. Yu
In this paper, we focus on the time-consuming training process of large-scale CNNs and propose a Bi-layered Parallel Training (BPT-CNN) architecture in distributed computing environments.
no code implementations • 24 Oct 2018 • Ye Liu, Jiawei Zhang, Chenwei Zhang, Philip S. Yu
After a thorough investigation of an online movie knowledge library, a novel movie planning framework "Blockbuster Planning with Maximized Movie Configuration Acquaintance" (BigMovie) is introduced in this paper.
no code implementations • 2 Nov 2018 • Guixiang Ma, Nesreen K. Ahmed, Ted Willke, Dipanjan Sengupta, Michael W. Cole, Nicholas B. Turk-Browne, Philip S. Yu
We propose an end-to-end similarity learning framework called Higher-order Siamese GCN for multi-subject fMRI data analysis.
no code implementations • 9 Nov 2018 • Shuaijun Ge, Guixiang Ma, Sihong Xie, Philip S. Yu
In terms of security, DETER is versatile enough to be vaccinated against diverse and unexpected evasions, is agnostic about evasion strategy and can be released without privacy concern.
no code implementations • 11 Nov 2018 • Jian-Guo Zhang, Pengcheng Zou, Zhao Li, Yao Wan, Ye Liu, Xiuming Pan, Yu Gong, Philip S. Yu
Nowadays, an increasing number of customers are in favor of using E-commerce Apps to browse and purchase products.
no code implementations • 11 Nov 2018 • Vahid Noroozi, Sara Bahaadini, Lei Zheng, Sihong Xie, Weixiang Shao, Philip S. Yu
While neural networks for learning representation of multi-view data have been previously proposed as one of the state-of-the-art multi-view dimension reduction techniques, how to make the representation discriminative with only a small amount of labeled data is not well-studied.
Dimensionality Reduction Learning Representation Of Multi-View Data
no code implementations • 13 Nov 2018 • Ji Wang, Weidong Bao, Lichao Sun, Xiaomin Zhu, Bokai Cao, Philip S. Yu
To benefit from the on-device deep learning without the capacity and privacy concerns, we design a private model compression framework RONA.
no code implementations • 12 Nov 2018 • Yao Wan, Wenqiang Yan, Jianwei Gao, Zhou Zhao, Jian Wu, Philip S. Yu
Dialogue Act (DA) classification is a challenging problem in dialogue interpretation, which aims to attach semantic labels to utterances and characterize the speaker's intention.
Ranked #5 on Dialogue Act Classification on Switchboard corpus
no code implementations • 20 Nov 2018 • Zhiyu Yao, Yunbo Wang, Mingsheng Long, Jian-Min Wang, Philip S. Yu, Jiaguang Sun
Rev2Net is shown to be effective on the classic action recognition task.
no code implementations • 10 Dec 2018 • Shen Wang, Zhengzhang Chen, Ding Li, Lu-An Tang, Jingchao Ni, Zhichun Li, Junghwan Rhee, Haifeng Chen, Philip S. Yu
The key idea is to leverage the representation learning of the heterogeneous program behavior graph to guide the reidentification process.
no code implementations • NeurIPS 2017 • Yunbo Wang, Mingsheng Long, Jian-Min Wang, Zhifeng Gao, Philip S. Yu
The core of this network is a new Spatiotemporal LSTM (ST-LSTM) unit that extracts and memorizes spatial and temporal representations simultaneously.
Ranked #6 on Video Prediction on Human3.6M
no code implementations • ICML 2017 • Lifang He, Chun-Ta Lu, Guixiang Ma, Shen Wang, Linlin Shen, Philip S. Yu, Ann B. Ragin
In the context of supervised tensor learning, preserving the structural information and exploiting the discriminative nonlinear relationships of tensor data are crucial for improving the performance of learning tasks.
no code implementations • ICLR 2018 • Chenwei Zhang, Yaliang Li, Nan Du, Wei Fan, Philip S. Yu
Online healthcare services can provide the general public with ubiquitous access to medical knowledge and reduce the information access cost for both individuals and societies.
no code implementations • ICLR 2018 • Hu Xu, Bing Liu, Lei Shu, Philip S. Yu
We observe that domains are not isolated and a small domain corpus can leverage the learned knowledge from many past domains to augment that corpus in order to generate high-quality embeddings.
no code implementations • CVPR 2013 • Mingsheng Long, Guiguang Ding, Jian-Min Wang, Jiaguang Sun, Yuchen Guo, Philip S. Yu
In this paper, we propose a Transfer Sparse Coding (TSC) approach to construct robust sparse representations for classifying cross-distribution images accurately.
no code implementations • CVPR 2014 • Mingsheng Long, Jian-Min Wang, Guiguang Ding, Jiaguang Sun, Philip S. Yu
Visual domain adaptation, which learns an accurate classifier for a new domain using labeled images from an old domain, has shown promising value in computer vision yet still been a challenging problem.
no code implementations • CVPR 2017 • Lifang He, Chun-Ta Lu, Hao Ding, Shen Wang, Linlin Shen, Philip S. Yu, Ann B. Ragin
Owing to prominence as a diagnostic tool for probing the neural correlates of cognition, neuroimaging tensor data has been the focus of intense investigation.
no code implementations • 11 Nov 2018 • Hao Peng, Jian-Xin Li, Qiran Gong, Senzhang Wang, Yuanxing Ning, Philip S. Yu
Different from previous convolutional neural networks on graphs, we first design a motif-matching guided subgraph normalization method to capture neighborhood information.
no code implementations • 26 Feb 2019 • Lu Bai, Lixin Cui, Yue Wang, Philip S. Yu, Edwin R. Hancock
To overcome these issues, we propose a new feature selection method using structural correlation between pairwise samples.
no code implementations • CVPR 2017 • Yunbo Wang, Mingsheng Long, Jian-Min Wang, Philip S. Yu
From the technical perspective, we introduce the spatiotemporal compact bilinear operator into video analysis tasks.
no code implementations • 7 Mar 2019 • Chaozhuo Li, Senzhang Wang, Philip S. Yu, Zhoujun Li
Specifically, we propose a MCNE model to learn compact embeddings from pre-learned node features.
no code implementations • 9 Mar 2019 • Jianping Cao, Senzhang Wang, Danyan Wen, Zhaohui Peng, Philip S. Yu, Fei-Yue Wang
HINT first models multi-sourced texts (e. g. news and tweets) as heterogeneous information networks by introducing the shared ``anchor texts'' to connect the comparative texts.
no code implementations • NAACL 2019 • Jian-Guo Zhang, Pengcheng Zou, Zhao Li, Yao Wan, Xiuming Pan, Yu Gong, Philip S. Yu
To address this discrepancy, previous studies mainly consider textual information of long product titles and lacks of human-like view during training and evaluation process.
no code implementations • 12 Apr 2019 • Jianguo Chen, Kenli Li, Qingying Deng, Keqin Li, Philip S. Yu
We implement the proposed DIVS system and address the problems of parallel training, model synchronization, and workload balancing.
no code implementations • 9 May 2019 • Shen Wang, Zhengzhang Chen, Jingchao Ni, Xiao Yu, Zhichun Li, Haifeng Chen, Philip S. Yu
How to address the vulnerabilities and defense GNN against the adversarial attacks?
no code implementations • 15 May 2019 • Bowen Dong, Jiawei Zhang, Chenwei Zhang, Yang Yang, Philip S. Yu
Online knowledge libraries refer to the online data warehouses that systematically organize and categorize the knowledge-based information about different kinds of concepts and entities.
no code implementations • 5 Jun 2019 • Lichao Sun, Yingbo Zhou, Ji Wang, Jia Li, Richard Sochar, Philip S. Yu, Caiming Xiong
Privacy-preserving deep learning is crucial for deploying deep neural network based solutions, especially when the model works on data that contains sensitive information.
no code implementations • 11 Jun 2019 • Senzhang Wang, Jiannong Cao, Philip S. Yu
Next we classify existing literatures based on the types of ST data, the data mining tasks, and the deep learning models, followed by the applications of deep learning for STDM in different domains including transportation, climate science, human mobility, location based social network, crime analysis, and neuroscience.
no code implementations • 5 Jun 2019 • Lichao Sun, Albert Chen, Philip S. Yu, Wei Chen
We incorporate self activation into influence propagation and propose the self-activation independent cascade (SAIC) model: nodes may be self activated besides being selected as seeds, and influence propagates from both selected seeds and self activated nodes.
Social and Information Networks
no code implementations • 5 Jul 2019 • Yingtong Dou, Weijian Li, Zhirong Liu, Zhenhua Dong, Jiebo Luo, Philip S. Yu
To the best of our knowledge, this is the first work that investigates the download fraud problem in mobile App markets.
no code implementations • 13 Aug 2019 • Yue Wang, Yao Wan, Chenwei Zhang, Lixin Cui, Lu Bai, Philip S. Yu
During the iterations, our model updates the parallel policies and the corresponding scenario-based regrets for agents simultaneously.
no code implementations • 14 Sep 2019 • Chuan Shi, Xiaotian Han, Li Song, Xiao Wang, Senzhang Wang, Junping Du, Philip S. Yu
However, the characteristics of users and the properties of items may stem from different aspects, e. g., the brand-aspect and category-aspect of items.
no code implementations • 15 Oct 2019 • Shaika Chowdhury, Chenwei Zhang, Philip S. Yu, Yuan Luo
Predicting patient mortality is an important and challenging problem in the healthcare domain, especially for intensive care unit (ICU) patients.
no code implementations • 14 Oct 2019 • Shaika Chowdhury, Chenwei Zhang, Philip S. Yu, Yuan Luo
Distributed representations have been used to support downstream tasks in healthcare recently.
no code implementations • 17 Oct 2019 • Shen Wang, Zhengzhang Chen, Xiao Yu, Ding Li, Jingchao Ni, Lu-An Tang, Jiaping Gui, Zhichun Li, Haifeng Chen, Philip S. Yu
Information systems have widely been the target of malware attacks.
no code implementations • 8 Nov 2019 • Jiahao Liu, Guixiang Ma, Fei Jiang, Chun-Ta Lu, Philip S. Yu, Ann B. Ragin
Specifically, we use graph convolutions to learn the structural and functional joint embedding, where the graph structure is defined with structural connectivity and node features are from the functional connectivity.
no code implementations • 26 Nov 2019 • Yue Wang, Chenwei Zhang, Shen Wang, Philip S. Yu, Lu Bai, Lixin Cui, Guandong Xu
We formalize networks with evolving structures as temporal networks and propose a generative link prediction model, Generative Link Sequence Modeling (GLSM), to predict future links for temporal networks.
no code implementations • 6 Dec 2019 • Shaika Chowdhury, Chenwei Zhang, Philip S. Yu, Yuan Luo
Distributed representations of medical concepts have been used to support downstream clinical tasks recently.
no code implementations • 25 Dec 2019 • Guixiang Ma, Nesreen K. Ahmed, Theodore L. Willke, Philip S. Yu
In many domains where data are represented as graphs, learning a similarity metric among graphs is considered a key problem, which can further facilitate various learning tasks, such as classification, clustering, and similarity search.
no code implementations • 31 Dec 2019 • Vahid Noroozi, Sara Bahaadini, Samira Sheikhi, Nooshin Mojab, Philip S. Yu
There has been a growing concern about the fairness of decision-making systems based on machine learning.
no code implementations • 18 Jan 2020 • Yuhui Zhao, Ning Yang, Tao Lin, Philip S. Yu
First, the existing works often assume an underlying information diffusion model, which is impractical in real world due to the complexity of information diffusion.
1 code implementation • 11 Feb 2020 • Shao-Heng Ko, Hsu-Chao Lai, Hong-Han Shuai, De-Nian Yang, Wang-Chien Lee, Philip S. Yu
Shopping in VR malls has been regarded as a paradigm shift for E-commerce, but most of the conventional VR shopping platforms are designed for a single user.
Data Structures and Algorithms
no code implementations • 1 Mar 2020 • Lichao Sun, Yingbo Zhou, Philip S. Yu, Caiming Xiong
Ensuring the privacy of sensitive data used to train modern machine learning models is of paramount importance in many areas of practice.
no code implementations • 7 Nov 2017 • Lichao Sun, Xiaokai Wei, Jiawei Zhang, Lifang He, Philip S. Yu, Witawas Srisa-an
The results indicate that once we remove contaminants from the datasets, we can significantly improve both malware detection rate and detection accuracy
Cryptography and Security
no code implementations • 22 Apr 2020 • Shoujin Wang, Liang Hu, Yan Wang, Xiangnan He, Quan Z. Sheng, Mehmet Orgun, Longbing Cao, Nan Wang, Francesco Ricci, Philip S. Yu
Recent years have witnessed the fast development of the emerging topic of Graph Learning based Recommender Systems (GLRS).
no code implementations • SIGDIAL (ACL) 2020 • Ye Liu, Tao Yang, Zeyu You, Wei Fan, Philip S. Yu
Human tackle reading comprehension not only based on the given context itself but often rely on the commonsense beyond.
no code implementations • 23 May 2020 • Ting-Ting Liang, Congying Xia, Yuyu Yin, Philip S. Yu
This paper proposes a novel neural network, joint training capsule network (JTCN), for the cold start recommendation task.
no code implementations • COLING 2020 • Hu Xu, Seungwhan Moon, Honglei Liu, Pararth Shah, Bing Liu, Philip S. Yu
We study a conversational recommendation model which dynamically manages users' past (offline) preferences and current (online) requests through a structured and cumulative user memory knowledge graph, to allow for natural interactions and accurate recommendations.
no code implementations • 6 Jul 2020 • Di Jin, Zhizhi Yu, Dongxiao He, Carl Yang, Philip S. Yu, Jiawei Han
Graph neural networks for HIN embeddings typically adopt a hierarchical attention (including node-level and meta-path-level attentions) to capture the information from meta-path-based neighbors.
no code implementations • 4 Jul 2020 • Jianguo Chen, Kenli Li, Zhaolei Zhang, Keqin Li, Philip S. Yu
The COVID-19 pandemic caused by the SARS-CoV-2 virus has spread rapidly worldwide, leading to a global outbreak.
no code implementations • 10 Jul 2020 • Longbing Cao, Qiang Yang, Philip S. Yu
Financial technology (FinTech) has been playing an increasingly critical role in driving modern economies, society, technology, and many other areas.
no code implementations • 5 Aug 2020 • Tianqing Zhu, Dayong Ye, Wei Wang, Wanlei Zhou, Philip S. Yu
Artificial Intelligence (AI) has attracted a great deal of attention in recent years.
no code implementations • 6 Aug 2020 • Ye Liu, Shaika Chowdhury, Chenwei Zhang, Cornelia Caragea, Philip S. Yu
Unlike most other QA tasks that focus on linguistic understanding, HeadQA requires deeper reasoning involving not only knowledge extraction, but also complex reasoning with healthcare knowledge.
no code implementations • 16 Aug 2020 • Dayong Ye, Tianqing Zhu, Sheng Shen, Wanlei Zhou, Philip S. Yu
To the best of our knowledge, this paper is the first to apply differential privacy to the field of multi-agent planning as a means of preserving the privacy of agents for logistic-like problems.
no code implementations • 30 Aug 2020 • Qingyun Sun, Hao Peng, Jian-Xin Li, Senzhang Wang, Xiangyu Dong, Liangxuan Zhao, Philip S. Yu, Lifang He
Although these attributes may change, an author's co-authors and research topics do not change frequently with time, which means that papers within a period have similar text and relation information in the academic network.
no code implementations • 14 Sep 2020 • Tao Zhang, Tianqing Zhu, Mengde Han, Jing Li, Wanlei Zhou, Philip S. Yu
Extensive experiments show that our method is able to achieve fair semi-supervised learning, and reach a better trade-off between accuracy and fairness than fair supervised learning.
no code implementations • 25 Sep 2020 • Tao Zhang, Tianqing Zhu, Jing Li, Mengde Han, Wanlei Zhou, Philip S. Yu
A set of experiments on real-world and synthetic datasets show that our method is able to use unlabeled data to achieve a better trade-off between accuracy and discrimination.
no code implementations • COLING 2020 • Lichao Sun, Congying Xia, Wenpeng Yin, TingTing Liang, Philip S. Yu, Lifang He
Our studies show that mixup is a domain-independent data augmentation technique to pre-trained language models, resulting in significant performance improvement for transformer-based models.
no code implementations • 13 Oct 2020 • Yue Wang, Zhuo Xu, Lu Bai, Yao Wan, Lixin Cui, Qian Zhao, Edwin R. Hancock, Philip S. Yu
To verify the effectiveness of our proposed method, we conduct extensive experiments on four real-world datasets as well as compare our method with state-of-the-art methods.
no code implementations • EMNLP 2020 • Shen Wang, Xiaokai Wei, Cicero Nogueira dos santos, Zhiguo Wang, Ramesh Nallapati, Andrew Arnold, Bing Xiang, Philip S. Yu
Existing knowledge graph embedding approaches concentrate on modeling symmetry/asymmetry, inversion, and composition typed relations but overlook the hierarchical nature of relations.
no code implementations • 30 Nov 2020 • Xiao Wang, Deyu Bo, Chuan Shi, Shaohua Fan, Yanfang Ye, Philip S. Yu
Heterogeneous graphs (HGs) also known as heterogeneous information networks have become ubiquitous in real-world scenarios; therefore, HG embedding, which aims to learn representations in a lower-dimension space while preserving the heterogeneous structures and semantics for downstream tasks (e. g., node/graph classification, node clustering, link prediction), has drawn considerable attentions in recent years.
no code implementations • 7 Dec 2020 • Lingjuan Lyu, Han Yu, Xingjun Ma, Chen Chen, Lichao Sun, Jun Zhao, Qiang Yang, Philip S. Yu
Besides training powerful global models, it is of paramount importance to design FL systems that have privacy guarantees and are resistant to different types of adversaries.
no code implementations • 3 Jan 2021 • Di Jin, Zhizhi Yu, Pengfei Jiao, Shirui Pan, Dongxiao He, Jia Wu, Philip S. Yu, Weixiong Zhang
We conclude with discussions of the challenges of the field and suggestions of possible directions for future research.
no code implementations • 17 Jan 2021 • Zheng Liu, Xiaohan Li, Hao Peng, Lifang He, Philip S. Yu
EHRs contain multiple entities and relations and can be viewed as a heterogeneous graph.
no code implementations • 19 Jan 2021 • Jianguo Chen, Kenli Li, Keqin Li, Philip S. Yu, Zeng Zeng
The BSDP system contains four modules: bicycle drop-off location clustering, bicycle-station graph modeling, bicycle-station location prediction, and bicycle-station layout recommendation.
no code implementations • 19 Jan 2021 • Jianguo Chen, Kenli Li, Keqin Li, Philip S. Yu, Zeng Zeng
We model the DL-PBS system from the perspective of CPS and use deep learning to predict the layout of bicycle parking spots and the dynamic demand of bicycle dispatching.
no code implementations • 22 Jan 2021 • Ye Liu, Yao Wan, Jian-Guo Zhang, Wenting Zhao, Philip S. Yu
In this paper, we claim that the syntactic and semantic structures among natural language are critical for non-autoregressive machine translation and can further improve the performance.
no code implementations • 27 Mar 2021 • Mehrnaz Najafi, Philip S. Yu
In this paper, we propose a novel Robust Graph Convolutional Neural Networks for possible erroneous single-view or multi-view data where data may come from multiple sources.
no code implementations • 6 Apr 2021 • Li Sun, Zhongbao Zhang, Jiawei Zhang, Feiyang Wang, Hao Peng, Sen Su, Philip S. Yu
To model the uncertainty, we devise a hyperbolic graph variational autoencoder built upon the proposed TGNN to generate stochastic node representations of hyperbolic normal distributions.
no code implementations • 30 Mar 2021 • Nooshin Mojab, Vahid Noroozi, Abdullah Aleem, Manoj P. Nallabothula, Joseph Baker, Dimitri T. Azar, Mark Rosenblatt, RV Paul Chan, Darvin Yi, Philip S. Yu, Joelle A. Hallak
In this paper, we present a new multi-modal longitudinal ophthalmic imaging dataset, the Illinois Ophthalmic Database Atlas (I-ODA), with the goal of advancing state-of-the-art computer vision applications in ophthalmology, and improving upon the translatable capacity of AI based applications across different clinical settings.
no code implementations • 16 Apr 2021 • Yu Wang, Lifu Huang, Philip S. Yu, Lichao Sun
Membership inference attacks (MIAs) infer whether a specific data record is used for target model training.
no code implementations • 7 May 2021 • Mehrnaz Najafi, Lifang He, Philip S. Yu
Due to inevitable sensor failures, data in each view may contain error.
no code implementations • 26 May 2021 • Xing Su, Shan Xue, Fanzhen Liu, Jia Wu, Jian Yang, Chuan Zhou, Wenbin Hu, Cecile Paris, Surya Nepal, Di Jin, Quan Z. Sheng, Philip S. Yu
A community reveals the features and connections of its members that are different from those in other communities in a network.
no code implementations • 30 Jun 2021 • Qiaomin Yi, Ning Yang, Philip S. Yu
First, the noise injection based methods often draw the noise from a fixed noise distribution given in advance, while in real world, the noise distributions of different users and items may differ from each other due to personal behaviors and item usage patterns.
no code implementations • 5 Jul 2021 • Qian Li, JianXin Li, Jiawei Sheng, Shiyao Cui, Jia Wu, Yiming Hei, Hao Peng, Shu Guo, Lihong Wang, Amin Beheshti, Philip S. Yu
Numerous methods, datasets, and evaluation metrics have been proposed in the literature, raising the need for a comprehensive and updated survey.
no code implementations • 7 Sep 2021 • Haoran Yang, Hongxu Chen, Lin Li, Philip S. Yu, Guandong Xu
They utilize simple and fixed schemes, like neighborhood information aggregation or mathematical calculation of vectors, to fuse the embeddings of different user behaviors to obtain a unified embedding to represent a user's behavioral patterns which will be used in downstream recommendation tasks.
no code implementations • 4 Oct 2021 • Chen Wang, Yingtong Dou, Min Chen, Jia Chen, Zhiwei Liu, Philip S. Yu
The successes of most previous methods heavily rely on rich node features and high-fidelity labels.
no code implementations • 29 Sep 2021 • Hengrui Zhang, Qitian Wu, Shaofeng Zhang, Junchi Yan, David Wipf, Philip S. Yu
In this paper, we propose ESCo (Effective and Scalable Contrastive), a new contrastive framework which is essentially an instantiation of the Information Bottleneck principle under self-supervised learning settings.
no code implementations • 29 Sep 2021 • Zhiwei Liu, Yongjun Chen, Jia Li, Man Luo, Philip S. Yu, Caiming Xiong
However, existing methods all construct views by adopting augmentation from data perspectives, while we argue that 1) optimal data augmentation methods are hard to devise, 2) data augmentation methods destroy sequential correlations, and 3) data augmentation fails to incorporate comprehensive self-supervised signals.
no code implementations • 14 Oct 2021 • Yang Shu, Zhangjie Cao, Jinghan Gao, Jianmin Wang, Philip S. Yu, Mingsheng Long
While pre-training and meta-training can create deep models powerful for few-shot generalization, we find that pre-training and meta-training focuses respectively on cross-domain transferability and cross-task transferability, which restricts their data efficiency in the entangled settings of domain shift and task shift.
no code implementations • 29 Oct 2021 • Nooshin Mojab, Philip S. Yu, Joelle A. Hallak, Darvin Yi
The success of deep learning methods relies heavily on the availability of a large amount of data.
no code implementations • EMNLP 2021 • Tao Zhang, Congying Xia, Philip S. Yu, Zhiwei Liu, Shu Zhao
Cross-domain Named Entity Recognition (NER) transfers the NER knowledge from high-resource domains to the low-resource target domain.
no code implementations • 16 Nov 2021 • Chen Wang, Yueqing Liang, Zhiwei Liu, Tao Zhang, Philip S. Yu
Then, we transfer the pre-trained graph encoder to initialize the node embeddings on the target domain, which benefits the fine-tuning of the single domain recommender system on the target domain.
no code implementations • 21 Nov 2021 • Zhiwei Liu, Liangwei Yang, Ziwei Fan, Hao Peng, Philip S. Yu
However, they all require centralized storage of the social links and item interactions of users, which leads to privacy concerns.
no code implementations • 24 Nov 2021 • Yicong Li, Hongxu Chen, Yile Li, Lin Li, Philip S. Yu, Guandong Xu
Recent advances in path-based explainable recommendation systems have attracted increasing attention thanks to the rich information provided by knowledge graphs.
no code implementations • 27 Sep 2018 • Congying Xia, Chenwei Zhang, Tao Yang, Yaliang Li, Nan Du, Xian Wu, Wei Fan, Fenglong Ma, Philip S. Yu
In this paper, we focus on a new Named Entity Recognition (NER) task, i. e., the Multi-grained NER task.
no code implementations • 27 Sep 2018 • Chenwei Zhang, Yaliang Li, Nan Du, Wei Fan, Philip S. Yu
Being able to automatically discover synonymous entities from a large free-text corpus has transformative effects on structured knowledge discovery.
no code implementations • 25 Sep 2019 • Lichao Sun, Yingbo Zhou, Jia Li, Richard Socher, Philip S. Yu, Caiming Xiong
Ensuring the privacy of sensitive data used to train modern machine learning models is of paramount importance in many areas of practice.
no code implementations • 25 Nov 2021 • Chuanpan Zheng, Xiaoliang Fan, Shirui Pan, Haibing Jin, Zhaopeng Peng, Zonghan Wu, Cheng Wang, Philip S. Yu
However, this approach failed to explicitly reflect the correlations between different nodes at different time steps, thus limiting the learning capability of graph neural networks.
no code implementations • 28 Nov 2021 • Xiaohan Li, Zhiwei Liu, Stephen Guo, Zheng Liu, Hao Peng, Philip S. Yu, Kannan Achan
In this paper, we propose a novel Reinforced Attentive Multi-relational Graph Neural Network (RAM-GNN) to the pre-train user and item embeddings on the user and item graph prior to the recommendation step.
no code implementations • 29 Nov 2021 • Gengsen Huang, Wensheng Gan, Jian Weng, Philip S. Yu
High utility sequential pattern mining (HUSPM) is one kind of utility-driven mining.
no code implementations • 10 Dec 2021 • Li Sun, Zhongbao Zhang, Junda Ye, Hao Peng, Jiawei Zhang, Sen Su, Philip S. Yu
Instead of working on one single constant-curvature space, we construct a mixed-curvature space via the Cartesian product of multiple Riemannian component spaces and design hierarchical attention mechanisms for learning and fusing the representations across these component spaces.
no code implementations • 14 Dec 2021 • Yiqi Wang, Chaozhuo Li, Zheng Liu, Mingzheng Li, Jiliang Tang, Xing Xie, Lei Chen, Philip S. Yu
Thus, graph pre-training has the great potential to alleviate data sparsity in GNN-based recommendations.
no code implementations • 15 Jan 2022 • Yuefei Lyu, Xiaoyu Yang, Jiaxin Liu, Philip S. Yu, Sihong Xie, Xi Zhang
To discover subtle vulnerabilities, we design a powerful attacking algorithm to camouflage rumors in social networks based on reinforcement learning that can interact with and attack any black-box detectors.
no code implementations • 16 Jan 2022 • Xiaoyun Zhao, Ning Yang, Philip S. Yu
Meanwhile, we propose a Multi-Domain Adaptation Network (MDAN) for MSDCR to capture a user's domain-invariant aspect preference.
no code implementations • 16 Jan 2022 • Ziwen Du, Ning Yang, Zhonghua Yu, Philip S. Yu
To address this challenges, we propose a novel model called Temporary Interest Aware Recommendation (TIARec), which can distinguish atypical interactions from normal ones without supervision and capture the temporary interest as well as the general preference of users.
no code implementations • 19 Jan 2022 • Haoran Yang, Hongxu Chen, Shirui Pan, Lin Li, Philip S. Yu, Guandong Xu
In addition, we conduct extensive experiments to analyze the impact of different graph encoders on DSGC, giving insights about how to better leverage the advantages of contrastive learning between different spaces.
no code implementations • 25 Jan 2022 • Daokun Zhang, Jie Yin, Philip S. Yu
To generate informative node embeddings for link prediction, structural context prediction is leveraged as a self-supervised learning task to boost the link prediction performance.
no code implementations • 8 Feb 2022 • Xiaoqin Pan, Xuan Lin, Dongsheng Cao, Xiangxiang Zeng, Philip S. Yu, Lifang He, Ruth Nussinov, Feixiong Cheng
Drug development is time-consuming and expensive.
no code implementations • 14 Feb 2022 • Xin Zheng, Yi Wang, Yixin Liu, Ming Li, Miao Zhang, Di Jin, Philip S. Yu, Shirui Pan
In the end, we point out the potential directions to advance and stimulate more future research and applications on heterophilic graph learning with GNNs.
no code implementations • 26 Feb 2022 • Gengsen Huang, Wensheng Gan, Philip S. Yu
What's more, to improve the efficiency of TaSPM on large-scale datasets and multiple-items-based sequence datasets, we propose several pruning strategies to reduce meaningless operations in mining processes.
no code implementations • 26 Feb 2022 • Wensheng Gan, Guoting Chen, Hongzhi Yin, Philippe Fournier-Viger, Chien-Ming Chen, Philip S. Yu
To fulfill this gap, in this paper, we first propose a general profit-oriented framework to address the problem of revenue maximization based on economic behavior, and compute the 0n-shelf Popular and most Profitable Products (OPPPs) for the targeted marketing.
no code implementations • 12 Mar 2022 • Zhi-Hong Deng, Chang-Dong Wang, Ling Huang, Jian-Huang Lai, Philip S. Yu
G$^3$SR decomposes the session-based recommendation workflow into two steps.
no code implementations • 18 Mar 2022 • Xusheng Zhao, Jia Wu, Hao Peng, Amin Beheshti, Jessica J. M. Monaghan, David Mcalpine, Heivet Hernandez-Perez, Mark Dras, Qiong Dai, Yangyang Li, Philip S. Yu, Lifang He
Modern neuroimaging techniques, such as diffusion tensor imaging (DTI) and functional magnetic resonance imaging (fMRI), enable us to model the human brain as a brain network or connectome.
no code implementations • 1 Apr 2022 • TingTing Liang, Yixuan Jiang, Congying Xia, Ziqiang Zhao, Yuyu Yin, Philip S. Yu
Recently, conversational OpenQA is proposed to address these issues with the abundant contextual information in the conversation.
1 code implementation • Findings (NAACL) 2022 • Yuwei Cao, William Groves, Tanay Kumar Saha, Joel R. Tetreault, Alex Jaimes, Hao Peng, Philip S. Yu
To date, work in this area has mostly focused on English as there is a scarcity of labeled data for other languages.
no code implementations • 24 May 2022 • Jiaqian Ren, Lei Jiang, Hao Peng, Zhiwei Liu, Jia Wu, Philip S. Yu
To incorporate temporal information into the message passing scheme, we introduce a novel temporal-aware aggregator which assigns weights to neighbours according to an adaptive time exponential decay formula.
no code implementations • 31 May 2022 • Shu'ang Li, Xuming Hu, Li Lin, Aiwei Liu, Lijie Wen, Philip S. Yu
Natural Language Inference (NLI) is a growingly essential task in natural language understanding, which requires inferring the relationship between the sentence pairs (premise and hypothesis).
no code implementations • 9 Jun 2022 • Wensheng Gan, Gengsen Huang, Jian Weng, Tianlong Gu, Philip S. Yu
In this paper, we provide the relevant definitions of target sequential rule and formulate the problem of targeted sequential rule mining.
no code implementations • 15 Jun 2022 • Yue Wang, Yao Wan, Lu Bai, Lixin Cui, Zhuo Xu, Ming Li, Philip S. Yu, Edwin R Hancock
To alleviate the challenges of building Knowledge Graphs (KG) from scratch, a more general task is to enrich a KG using triples from an open corpus, where the obtained triples contain noisy entities and relations.
no code implementations • 27 Aug 2022 • Jiahui Chen, Xu Guo, Wensheng Gan, Shichen Wan, Philip S. Yu
Compared with traditional utility mining, OSUM can find more practical and meaningful patterns in real-life applications.
no code implementations • 30 Aug 2022 • Li Sun, Junda Ye, Hao Peng, Philip S. Yu
To bridge this gap, we make the first attempt to study the problem of self-supervised temporal graph representation learning in the general Riemannian space, supporting the time-varying curvature to shift among hyperspherical, Euclidean and hyperbolic spaces.
no code implementations • COLING 2022 • Xuming Hu, Zhijiang Guo, Yu Fu, Lijie Wen, Philip S. Yu
A scene graph is a semantic representation that expresses the objects, attributes, and relationships between objects in a scene.
no code implementations • 27 Sep 2022 • Yao Chen, Wensheng Gan, Yongdong Wu, Philip S. Yu
Contrast pattern mining (CPM) is an important and popular subfield of data mining.
no code implementations • 27 Sep 2022 • Chunkai Zhang, Maohua Lyu, Wensheng Gan, Philip S. Yu
TotalSR creates a utility table that can efficiently calculate antecedent support and a utility prefix sum list that can compute the remaining utility in O(1) time for a sequence.
no code implementations • 9 Oct 2022 • Yazhou Ren, Jingyu Pu, Zhimeng Yang, Jie Xu, Guofeng Li, Xiaorong Pu, Philip S. Yu, Lifang He
Finally, we discuss the open challenges and potential future opportunities in different fields of deep clustering.
no code implementations • 13 Oct 2022 • Jianpeng Chen, Yawen Ling, Jie Xu, Yazhou Ren, Shudong Huang, Xiaorong Pu, Zhifeng Hao, Philip S. Yu, Lifang He
The critical point of MGC is to better utilize the view-specific and view-common information in features and graphs of multiple views.
no code implementations • 19 Oct 2022 • Xuming Hu, Yong Jiang, Aiwei Liu, Zhongqiang Huang, Pengjun Xie, Fei Huang, Lijie Wen, Philip S. Yu
Data augmentation techniques have been used to alleviate the problem of scarce labeled data in various NER tasks (flat, nested, and discontinuous NER tasks).
no code implementations • 28 Oct 2022 • Byung-Hak Kim, Zhongfen Deng, Philip S. Yu, Varun Ganapathi
The medical codes prediction problem from clinical notes has received substantial interest in the NLP community, and several recent studies have shown the state-of-the-art (SOTA) code prediction results of full-fledged deep learning-based methods.
no code implementations • 11 Nov 2022 • Xuming Hu, Shiao Meng, Chenwei Zhang, Xiangli Yang, Lijie Wen, Irwin King, Philip S. Yu
Low-Resource Information Extraction (LRIE) strives to use unsupervised data, reducing the required resources and human annotation.
no code implementations • 30 Nov 2022 • Li Sun, Junda Ye, Hao Peng, Feiyang Wang, Philip S. Yu
On the one hand, existing methods work with the zero-curvature Euclidean space, and largely ignore the fact that curvature varies over the coming graph sequence.
no code implementations • 8 Dec 2022 • Hengrui Zhang, Qitian Wu, Yu Wang, Shaofeng Zhang, Junchi Yan, Philip S. Yu
Contrastive learning methods based on InfoNCE loss are popular in node representation learning tasks on graph-structured data.
no code implementations • 20 Dec 2022 • Gengsen Huang, Wensheng Gan, Philip S. Yu
An algorithm called Sequence Utility Maximization with Utility occupancy measure (SUMU) is proposed.
no code implementations • 30 Dec 2022 • Qingyun Sun, JianXin Li, Beining Yang, Xingcheng Fu, Hao Peng, Philip S. Yu
Most Graph Neural Networks follow the message-passing paradigm, assuming the observed structure depicts the ground-truth node relationships.
no code implementations • 10 Jan 2023 • Xiaohan Li, Yuqing Liu, Zheng Liu, Philip S. Yu
TA-HGAT is built in a hyperbolic space to learn the hierarchical structure of session graphs.
no code implementations • 22 Jan 2023 • Lilin Zhang, Ning Yang, Yanchao Sun, Philip S. Yu
Second, the existing AT methods often achieve adversarial robustness at the expense of standard generalizability (i. e., the accuracy on natural examples) because they make a tradeoff between them.
no code implementations • 31 Jan 2023 • Hengrui Zhang, Shen Wang, Vassilis N. Ioannidis, Soji Adeshina, Jiani Zhang, Xiao Qin, Christos Faloutsos, Da Zheng, George Karypis, Philip S. Yu
Graph Neural Networks (GNNs) are currently dominating in modeling graph-structure data, while their high reliance on graph structure for inference significantly impedes them from widespread applications.
no code implementations • 13 Feb 2023 • Song Wu, Yazhou Ren, Aodi Yang, Xinyue Chen, Xiaorong Pu, Jing He, Liqiang Nie, Philip S. Yu
In this survey, we investigate the main contributions of deep learning applications using medical images in fighting against COVID-19 from the aspects of image classification, lesion localization, and severity quantification, and review different deep learning architectures and some image preprocessing techniques for achieving a preciser diagnosis.
no code implementations • 18 Feb 2023 • Ce Zhou, Qian Li, Chen Li, Jun Yu, Yixin Liu, Guangjing Wang, Kai Zhang, Cheng Ji, Qiben Yan, Lifang He, Hao Peng, JianXin Li, Jia Wu, Ziwei Liu, Pengtao Xie, Caiming Xiong, Jian Pei, Philip S. Yu, Lichao Sun
This study provides a comprehensive review of recent research advancements, challenges, and opportunities for PFMs in text, image, graph, as well as other data modalities.
no code implementations • 22 Apr 2023 • Yu Wang, Zhiwei Liu, Liangwei Yang, Philip S. Yu
Generative models have attracted significant interest due to their ability to handle uncertainty by learning the inherent data distributions.
no code implementations • 5 May 2023 • Li Sun, Feiyang Wang, Junda Ye, Hao Peng, Philip S. Yu
On the other hand, contrastive learning boosts the deep graph clustering but usually struggles in either graph augmentation or hard sample mining.
no code implementations • 12 May 2023 • Ziwei Fan, Zhiwei Liu, Shelby Heinecke, JianGuo Zhang, Huan Wang, Caiming Xiong, Philip S. Yu
This paper presents a novel paradigm for the Zero-Shot Item-based Recommendation (ZSIR) task, which pre-trains a model on product knowledge graph (PKG) to refine the item features from PLMs.
no code implementations • 12 May 2023 • Yawen Yang, Xuming Hu, Fukun Ma, Shu'ang Li, Aiwei Liu, Lijie Wen, Philip S. Yu
Existing works for nested NER ignore the recognition order and boundary position relation of nested entities.
no code implementations • 21 May 2023 • Guangsi Shi, Daokun Zhang, Ming Jin, Shirui Pan, Philip S. Yu
To better comprehend the complex physical laws, this paper proposes a novel learning based simulation model- Graph Networks with Spatial-Temporal neural Ordinary Equations (GNSTODE)- that characterizes the varying spatial and temporal dependencies in particle systems using a united end-to-end framework.
no code implementations • 25 May 2023 • Xuming Hu, Zhijiang Guo, Zhiyang Teng, Irwin King, Philip S. Yu
Multimodal relation extraction (MRE) is the task of identifying the semantic relationships between two entities based on the context of the sentence image pair.
no code implementations • 25 May 2023 • Xuming Hu, Junzhe Chen, Zhijiang Guo, Philip S. Yu
Evidence plays a crucial role in automated fact-checking.
no code implementations • 26 May 2023 • Xuming Hu, Aiwei Liu, Zeqi Tan, Xin Zhang, Chenwei Zhang, Irwin King, Philip S. Yu
These techniques neither preserve the semantic consistency of the original sentences when rule-based augmentations are adopted, nor preserve the syntax structure of sentences when expressing relations using seq2seq models, resulting in less diverse augmentations.
no code implementations • 2 Jun 2023 • Liangqi Yuan, Lichao Sun, Philip S. Yu, Ziran Wang
Federated learning (FL) has been gaining attention for its ability to share knowledge while maintaining user data, protecting privacy, increasing learning efficiency, and reducing communication overhead.
no code implementations • 6 Jun 2023 • Heng Xu, Tianqing Zhu, Lefeng Zhang, Wanlei Zhou, Philip S. Yu
Machine learning has attracted widespread attention and evolved into an enabling technology for a wide range of highly successful applications, such as intelligent computer vision, speech recognition, medical diagnosis, and more.
no code implementations • 21 Jun 2023 • Ziwei Fan, Zhiwei Liu, Hao Peng, Philip S. Yu
We also establish a correlation between the ranks of sequence and item embeddings and the rank of the user-item preference prediction matrix, which can affect recommendation diversity.
no code implementations • 25 Jun 2023 • Huiqiang Chen, Tianqing Zhu, Tao Zhang, Wanlei Zhou, Philip S. Yu
Federated learning (FL) has been a hot topic in recent years.
no code implementations • 29 Jul 2023 • Yibo Wang, Yanbing Xue, Bo Liu, Musen Wen, Wenting Zhao, Stephen Guo, Philip S. Yu
Position bias, the phenomenon whereby users tend to focus on higher-ranked items of the search result list regardless of the actual relevance to queries, is prevailing in many ranking systems.
no code implementations • 16 Aug 2023 • Mingdai Yang, Zhiwei Liu, Liangwei Yang, Xiaolong Liu, Chen Wang, Hao Peng, Philip S. Yu
With the proliferation of social media, a growing number of users search for and join group activities in their daily life.
no code implementations • 7 Sep 2023 • Jiangshu Du, Congying Xia, Wenpeng Yin, TingTing Liang, Philip S. Yu
In intent detection tasks, leveraging meaningful semantic information from intent labels can be particularly beneficial for few-shot scenarios.
no code implementations • 20 Sep 2023 • Wenting Zhao, Ye Liu, Yao Wan, Yibo Wang, Zhongfen Deng, Philip S. Yu
Furthermore, TAG-QA outperforms the end-to-end model T5 by 16% and 12% on BLEU-4 and PARENT F-score, respectively.
1 code implementation • 28 Sep 2023 • Chunkai Zhang, Maohua Lyu, Huaijin Hao, Wensheng Gan, Philip S. Yu
For artificial intelligence, high-utility sequential rule mining (HUSRM) is a knowledge discovery method that can reveal the associations between events in the sequences.
no code implementations • 30 Sep 2023 • Haishuai Wang, Yang Gao, Xin Zheng, Peng Zhang, Hongyang Chen, Jiajun Bu, Philip S. Yu
In this paper, we integrate GPT-4 into GNAS and propose a new GPT-4 based Graph Neural Architecture Search method (GPT4GNAS for short).
no code implementations • 8 Oct 2023 • Xuming Hu, Junzhe Chen, Xiaochuan Li, Yufei Guo, Lijie Wen, Philip S. Yu, Zhijiang Guo
Large language models (LLMs) have recently driven striking performance improvements across a range of natural language processing tasks.
1 code implementation • 11 Oct 2023 • Zhongfen Deng, Wei-Te Chen, Lei Chen, Philip S. Yu
In this paper, we reformulate this task as a multi-label classification task that can be applied for real-world scenario in which only annotation of attribute values is available to train models (i. e., annotation of positional information of attribute values is not available).
no code implementations • 13 Oct 2023 • Chen Wang, Liangwei Yang, Zhiwei Liu, Xiaolong Liu, Mingdai Yang, Yueqing Liang, Philip S. Yu
However, PLMs often overlook the vital collaborative filtering signals, leading to challenges in merging collaborative and semantic representation spaces and fine-tuning semantic representations for better alignment with warm-start conditions.
no code implementations • 18 Oct 2023 • Jiawei Liu, Cheng Yang, Zhiyuan Lu, Junze Chen, Yibo Li, Mengmei Zhang, Ting Bai, Yuan Fang, Lichao Sun, Philip S. Yu, Chuan Shi
Foundation models have emerged as critical components in a variety of artificial intelligence applications, and showcase significant success in natural language processing and several other domains.
1 code implementation • 30 Oct 2023 • Jiaqian Ren, Hao Peng, Lei Jiang, Zhiwei Liu, Jia Wu, Zhengtao Yu, Philip S. Yu
While in our observation, compared to the rarity of classes, the calibrated uncertainty estimated from well-trained evidential deep learning networks better reflects model performance.
no code implementations • 31 Oct 2023 • Wenting Zhao, Ye Liu, Tong Niu, Yao Wan, Philip S. Yu, Shafiq Joty, Yingbo Zhou, Semih Yavuz
Moreover, a significant gap in the current landscape is the absence of a realistic benchmark for evaluating the effectiveness of grounding LLMs on heterogeneous knowledge sources (e. g., knowledge base and text).
no code implementations • 1 Nov 2023 • Jiangnan Xia, Yu Yang, Senzhang Wang, Hongzhi Yin, Jiannong Cao, Philip S. Yu
To this end, we investigate a novel problem of robust POI recommendation by considering the uncertainty factors of the user check-ins, and proposes a Bayes-enhanced Multi-view Attention Network.
no code implementations • 1 Nov 2023 • Xiangjue Dong, Yibo Wang, Philip S. Yu, James Caverlee
Large Language Models (LLMs) can generate biased and toxic responses.
no code implementations • 2 Nov 2023 • Hao Niu, Yun Xiong, Xiaosu Wang, Philip S. Yu
Furthermore, we propose a dual-level graph attention network as a global encoder by fully employing dependency tag information to capture long-distance information effectively.
no code implementations • 7 Nov 2023 • Zhongfen Deng, Seunghyun Yoon, Trung Bui, Franck Dernoncourt, Quan Hung Tran, Shuaiqi Liu, Wenting Zhao, Tao Zhang, Yibo Wang, Philip S. Yu
Then we merge the sentences selected for a specific aspect as the input for the summarizer to produce the aspect-based summary.