Search Results for author: Hao Peng

Found 226 papers, 144 papers with code

Improving Influence-based Instruction Tuning Data Selection for Balanced Learning of Diverse Capabilities

no code implementations21 Jan 2025 Qirun Dai, Dylan Zhang, Jiaqi W. Ma, Hao Peng

Our analysis further highlights the importance of both instance-level normalization and iterative optimization of selected data for balanced learning of diverse capabilities.

Hierarchical Superpixel Segmentation via Structural Information Theory

1 code implementation13 Jan 2025 Minhui Xie, Hao Peng, Pu Li, Guangjie Zeng, Shuhai Wang, Jia Wu, Peng Li, Philip S. Yu

Superpixel segmentation is a foundation for many higher-level computer vision tasks, such as image segmentation, object recognition, and scene understanding.

graph construction graph partitioning +5

Prompt-based Unifying Inference Attack on Graph Neural Networks

no code implementations20 Dec 2024 Yuecen Wei, Xingcheng Fu, Lingyun Liu, Qingyun Sun, Hao Peng, Chunming Hu

Specifically, ProIA retains the crucial topological information of the graph during pre-training, enhancing the background knowledge of the inference attack model.

Disentanglement Inference Attack

LongBench v2: Towards Deeper Understanding and Reasoning on Realistic Long-context Multitasks

1 code implementation19 Dec 2024 Yushi Bai, Shangqing Tu, Jiajie Zhang, Hao Peng, Xiaozhi Wang, Xin Lv, Shulin Cao, Jiazheng Xu, Lei Hou, Yuxiao Dong, Jie Tang, Juanzi Li

This paper introduces LongBench v2, a benchmark designed to assess the ability of LLMs to handle long-context problems requiring deep understanding and reasoning across real-world multitasks.

8k In-Context Learning +1

SocialED: A Python Library for Social Event Detection

1 code implementation18 Dec 2024 Kun Zhang, Xiaoyan Yu, Pu Li, Hao Peng, Philip S. Yu

SocialED is a comprehensive, open-source Python library designed to support social event detection (SED) tasks, integrating 19 detection algorithms and 14 diverse datasets.

Event Detection graph construction

Towards Effective, Efficient and Unsupervised Social Event Detection in the Hyperbolic Space

2 code implementations14 Dec 2024 Xiaoyan Yu, Yifan Wei, Shuaishuai Zhou, Zhiwei Yang, Li Sun, Hao Peng, Liehuang Zhu, Philip S. Yu

Specifically, the proposed framework first models social messages into semantic-based message anchors, and then leverages the structure of the anchor graph and the expressiveness of the hyperbolic space to acquire structure- and geometry-aware anchor representations.

Event Detection

Structural Entropy Guided Probabilistic Coding

1 code implementation12 Dec 2024 Xiang Huang, Hao Peng, Li Sun, Hui Lin, Chunyang Liu, Jiang Cao, Philip S. Yu

Probabilistic embeddings have several advantages over deterministic embeddings as they map each data point to a distribution, which better describes the uncertainty and complexity of data.

Natural Language Understanding regression +1

Graph-Sequential Alignment and Uniformity: Toward Enhanced Recommendation Systems

1 code implementation5 Dec 2024 Yuwei Cao, Liangwei Yang, Zhiwei Liu, Yuqing Liu, Chen Wang, Yueqing Liang, Hao Peng, Philip S. Yu

Graph-based and sequential methods are two popular recommendation paradigms, each excelling in its domain but lacking the ability to leverage signals from the other.

Graph Neural Network Transfer Learning

Free Process Rewards without Process Labels

1 code implementation2 Dec 2024 Lifan Yuan, Wendi Li, Huayu Chen, Ganqu Cui, Ning Ding, Kaiyan Zhang, BoWen Zhou, Zhiyuan Liu, Hao Peng

The only assumption is to parameterize the outcome reward as the log-likelihood ratios of the policy and reference models, which can be optimized regardless of the specific choice of loss objectives.

Math

Multi-View Incongruity Learning for Multimodal Sarcasm Detection

no code implementations1 Dec 2024 Diandian Guo, Cong Cao, Fangfang Yuan, Yanbing Liu, Guangjie Zeng, Xiaoyan Yu, Hao Peng, Philip S. Yu

Experimental results demonstrate the superiority of MICL on benchmark datasets, along with the analyses showcasing MICL's advancement in mitigating the effect of spurious correlation.

Contrastive Learning Data Augmentation +2

Constraint Back-translation Improves Complex Instruction Following of Large Language Models

no code implementations31 Oct 2024 Yunjia Qi, Hao Peng, Xiaozhi Wang, Bin Xu, Lei Hou, Juanzi Li

Following the conventional instruction-tuning practice, previous works conduct post-training on complex instruction-response pairs generated by feeding complex instructions to advanced LLMs.

Instruction Following Translation

Spiking Graph Neural Network on Riemannian Manifolds

1 code implementation23 Oct 2024 Li Sun, Zhenhao Huang, Qiqi Wan, Hao Peng, Philip S. Yu

Extensive experiments on common graphs show the proposed MSG achieves superior performance to previous spiking GNNs and energy efficiency to conventional GNNs.

Graph Neural Network

Scaling Diffusion Language Models via Adaptation from Autoregressive Models

1 code implementation23 Oct 2024 Shansan Gong, Shivam Agarwal, Yizhe Zhang, Jiacheng Ye, Lin Zheng, Mukai Li, Chenxin An, Peilin Zhao, Wei Bi, Jiawei Han, Hao Peng, Lingpeng Kong

Diffusion Language Models (DLMs) have emerged as a promising new paradigm for text generative modeling, potentially addressing limitations of autoregressive (AR) models.

In-Context Learning Language Modeling +1

Pre-training Distillation for Large Language Models: A Design Space Exploration

no code implementations21 Oct 2024 Hao Peng, Xin Lv, Yushi Bai, Zijun Yao, Jiajie Zhang, Lei Hou, Juanzi Li

Previous work applying KD in the field of large language models (LLMs) typically focused on the post-training phase, where the student LLM learns directly from instructions and corresponding responses generated by the teacher model.

Knowledge Distillation

CoreGuard: Safeguarding Foundational Capabilities of LLMs Against Model Stealing in Edge Deployment

no code implementations16 Oct 2024 Qinfeng Li, Yangfan Xie, Tianyu Du, Zhiqiang Shen, Zhenghan Qin, Hao Peng, Xinkui Zhao, Xianwei Zhu, Jianwei Yin, Xuhong Zhang

However, edge deployment of proprietary LLMs introduces new security threats: attackers who obtain an edge-deployed LLM can easily use it as a base model for various tasks due to its high generalization ability, which we call foundational capability stealing.

Effective Exploration Based on the Structural Information Principles

no code implementations9 Oct 2024 Xianghua Zeng, Hao Peng, Angsheng Li

Traditional information theory provides a valuable foundation for Reinforcement Learning, particularly through representation learning and entropy maximization for agent exploration.

Representation Learning

FactCheckmate: Preemptively Detecting and Mitigating Hallucinations in LMs

no code implementations3 Oct 2024 Deema Alnuhait, Neeraja Kirtane, Muhammad Khalifa, Hao Peng

Practically, both the detection and mitigation models in FactCheckMate are lightweight, adding little inference overhead; FactCheckMate proves a more efficient approach for mitigating hallucinations compared to many post-hoc alternatives.

Hallucination

A Little Goes a Long Way: Efficient Long Context Training and Inference with Partial Contexts

no code implementations2 Oct 2024 Suyu Ge, Xihui Lin, Yunan Zhang, Jiawei Han, Hao Peng

To address this, two critical steps are often required: a pretrained LLM typically undergoes a separate stage for context length extension by training on long-context data, followed by architectural modifications to reduce the overhead of KV cache during serving.

4k

DAMe: Personalized Federated Social Event Detection with Dual Aggregation Mechanism

1 code implementation1 Sep 2024 Xiaoyan Yu, Yifan Wei, Pu Li, Shuaishuai Zhou, Hao Peng, Li Sun, Liehuang Zhu, Philip S. Yu

We present a novel local aggregation strategy utilizing Bayesian optimization to incorporate global knowledge while retaining local characteristics.

Bayesian Optimization Event Detection +1

Multivariate Time-Series Anomaly Detection based on Enhancing Graph Attention Networks with Topological Analysis

1 code implementation23 Aug 2024 Zhe Liu, Xiang Huang, Jingyun Zhang, Zhifeng Hao, Li Sun, Hao Peng

Additionally, we present an augmented GAT to manage complex inter-feature dependencies, which incorporates graph topology into node features across multiple scales, a versatile, plug-and-play enhancement that significantly boosts the performance of GAT.

Graph Attention Time Series +2

OpenEP: Open-Ended Future Event Prediction

no code implementations13 Aug 2024 Yong Guan, Hao Peng, Xiaozhi Wang, Lei Hou, Juanzi Li

For question construction, we pose questions from seven perspectives, including location, time, event development, event outcome, event impact, event response, and other, to facilitate an in-depth analysis and understanding of the comprehensive evolution of events.

S2-Attention: Hardware-Aware Context Sharding Among Attention Heads

no code implementations25 Jul 2024 Xihui Lin, Yunan Zhang, Suyu Ge, Liliang Ren, Barun Patra, Vishrav Chaudhary, Hao Peng, Xia Song

S2-Attention achieves wall-clock speedup of 8. 79X, 15. 87X, 25. 3X compared to the strong FlashAttention-2 baseline with strong downstream performance on-par with full attention and perfect retrieval performance at a 128k context length.

OpenHands: An Open Platform for AI Software Developers as Generalist Agents

2 code implementations23 Jul 2024 Xingyao Wang, Boxuan Li, Yufan Song, Frank F. Xu, Xiangru Tang, Mingchen Zhuge, Jiayi Pan, Yueqi Song, Bowen Li, Jaskirat Singh, Hoang H. Tran, Fuqiang Li, Ren Ma, Mingzhang Zheng, Bill Qian, Yanjun Shao, Niklas Muennighoff, Yizhe Zhang, Binyuan Hui, Junyang Lin, Robert Brennan, Hao Peng, Heng Ji, Graham Neubig

OpenDevin), a platform for the development of powerful and flexible AI agents that interact with the world in similar ways to those of a human developer: by writing code, interacting with a command line, and browsing the web.

Adaptive Differentially Private Structural Entropy Minimization for Unsupervised Social Event Detection

no code implementations23 Jul 2024 Zhiwei Yang, Yuecen Wei, Haoran Li, Qian Li, Lei Jiang, Li Sun, Xiaoyan Yu, Chunming Hu, Hao Peng

In this process, our method can adaptively apply differential privacy based on the events occurring each day in an open environment to maximize the use of the privacy budget.

Event Detection

MAVEN-Fact: A Large-scale Event Factuality Detection Dataset

2 code implementations22 Jul 2024 Chunyang Li, Hao Peng, Xiaozhi Wang, Yunjia Qi, Lei Hou, Bin Xu, Juanzi Li

Thanks to the comprehensive annotations of event arguments and relations in MAVEN, MAVEN-Fact also supports some further analyses and we find that adopting event arguments and relations helps in event factuality detection for fine-tuned models but does not benefit LLMs.

Hallucination

Eliminating Position Bias of Language Models: A Mechanistic Approach

1 code implementation1 Jul 2024 Ziqi Wang, HANLIN ZHANG, Xiner Li, Kuan-Hao Huang, Chi Han, Shuiwang Ji, Sham M. Kakade, Hao Peng, Heng Ji

Position bias has proven to be a prevalent issue of modern language models (LMs), where the models prioritize content based on its position within the given context.

Math object-detection +4

GC-Bench: An Open and Unified Benchmark for Graph Condensation

1 code implementation30 Jun 2024 Qingyun Sun, Ziying Chen, Beining Yang, Cheng Ji, Xingcheng Fu, Sheng Zhou, Hao Peng, JianXin Li, Philip S. Yu

To fill this gap, we develop a comprehensive Graph Condensation Benchmark (GC-Bench) to analyze the performance of graph condensation in different scenarios systematically.

IGL-Bench: Establishing the Comprehensive Benchmark for Imbalanced Graph Learning

no code implementations14 Jun 2024 Jiawen Qin, Haonan Yuan, Qingyun Sun, Lyujin Xu, Jiaqi Yuan, Pengfeng Huang, Zhaonan Wang, Xingcheng Fu, Hao Peng, JianXin Li, Philip S. Yu

Deep graph learning has gained grand popularity over the past years due to its versatility and success in representing graph data across a wide range of domains.

Graph Learning

$\textbf{PLUM}$: Improving Code LMs with Execution-Guided On-Policy Preference Learning Driven By Synthetic Test Cases

no code implementations11 Jun 2024 Dylan Zhang, Shizhe Diao, Xueyan Zou, Hao Peng

Recent findings demonstrate that on-policy data is the key to successful preference learning, where the preference data is collected using the same policy LM being trained.

Code Generation HumanEval +1

M&M VTO: Multi-Garment Virtual Try-On and Editing

1 code implementation CVPR 2024 Luyang Zhu, Yingwei Li, Nan Liu, Hao Peng, Dawei Yang, Ira Kemelmacher-Shlizerman

We present M&M VTO, a mix and match virtual try-on method that takes as input multiple garment images, text description for garment layout and an image of a person.

Denoising Super-Resolution +1

R-ODE: Ricci Curvature Tells When You Will be Informed

no code implementations27 May 2024 Li Sun, Jingbin Hu, Mengjie Li, Hao Peng

Such limitation motivates us to pose the problem of the time-aware personalized information diffusion prediction for the first time, telling the time when the target user will be informed.

Blocking Graph Neural Network

LSEnet: Lorentz Structural Entropy Neural Network for Deep Graph Clustering

1 code implementation20 May 2024 Li Sun, Zhenhao Huang, Hao Peng, Yujie Wang, Chunyang Liu, Philip S. Yu

DSI is also theoretically presented as a new graph clustering objective, not requiring the predefined cluster number.

Clustering Deep Clustering +1

Event GDR: Event-Centric Generative Document Retrieval

no code implementations11 May 2024 Yong Guan, Dingxiao Liu, Jinchen Ma, Hao Peng, Xiaozhi Wang, Lei Hou, Ru Li

Inspired by this, we propose Event GDR, an event-centric generative document retrieval model, integrating event knowledge into this task.

Information Retrieval Retrieval

ADELIE: Aligning Large Language Models on Information Extraction

1 code implementation8 May 2024 Yunjia Qi, Hao Peng, Xiaozhi Wang, Bin Xu, Lei Hou, Juanzi Li

Large language models (LLMs) usually fall short on information extraction (IE) tasks and struggle to follow the complex instructions of IE tasks.

Hyperbolic Geometric Latent Diffusion Model for Graph Generation

1 code implementation6 May 2024 Xingcheng Fu, Yisen Gao, Yuecen Wei, Qingyun Sun, Hao Peng, JianXin Li, Xianxian Li

Diffusion models have made significant contributions to computer vision, sparking a growing interest in the community recently regarding the application of them to graph generation.

Graph Generation

Retrieval Head Mechanistically Explains Long-Context Factuality

1 code implementation24 Apr 2024 Wenhao Wu, Yizhong Wang, Guangxuan Xiao, Hao Peng, Yao Fu

Despite the recent progress in long-context language models, it remains elusive how transformer-based models exhibit the capability to retrieve relevant information from arbitrary locations within the long context.

Continual Pretraining Hallucination +3

BotDGT: Dynamicity-aware Social Bot Detection with Dynamic Graph Transformers

1 code implementation23 Apr 2024 Buyun He, Yingguang Yang, Qi Wu, Hao liu, Renyu Yang, Hao Peng, Xiang Wang, Yong Liao, Pengyuan Zhou

To tackle these challenges, we propose BotDGT, a novel framework that not only considers the topological structure, but also effectively incorporates dynamic nature of social network.

Misinformation

Effective Reinforcement Learning Based on Structural Information Principles

no code implementations15 Apr 2024 Xianghua Zeng, Hao Peng, Dingli Su, Angsheng Li

An innovative two-layer skill-based learning mechanism is introduced to compute the common path entropy of each state transition as its identified probability, thereby obviating the requirement for expert knowledge.

Decision Making reinforcement-learning +2

Relational Prompt-based Pre-trained Language Models for Social Event Detection

1 code implementation12 Apr 2024 Pu Li, Xiaoyan Yu, Hao Peng, Yantuan Xian, Linqin Wang, Li Sun, Jingyun Zhang, Philip S. Yu

In this paper, we approach social event detection from a new perspective based on Pre-trained Language Models (PLMs), and present RPLM_SED (Relational prompt-based Pre-trained Language Models for Social Event Detection).

Event Detection Graph Neural Network

Source-Aware Training Enables Knowledge Attribution in Language Models

1 code implementation1 Apr 2024 Muhammad Khalifa, David Wadden, Emma Strubell, Honglak Lee, Lu Wang, Iz Beltagy, Hao Peng

We investigate the problem of intrinsic source citation, where LLMs are required to cite the pretraining source supporting a generated response.

Data Augmentation

Instruction-based Hypergraph Pretraining

no code implementations28 Mar 2024 Mingdai Yang, Zhiwei Liu, Liangwei Yang, Xiaolong Liu, Chen Wang, Hao Peng, Philip S. Yu

However, the gap between training objectives and the discrepancy between data distributions in pretraining and downstream tasks hinders the transfer of the pretrained knowledge.

Diversity Graph Learning +2

Understanding the PULSAR Effect in Combined Radiotherapy and Immunotherapy through Attention Mechanisms with a Transformer Model

no code implementations7 Mar 2024 Hao Peng, Casey Moore, Debabrata Saha, Steve Jiang, Robert Timmerman

PULSAR (personalized, ultra-fractionated stereotactic adaptive radiotherapy) is the adaptation of stereotactic ablative radiotherapy towards personalized cancer management.

Management

Neeko: Leveraging Dynamic LoRA for Efficient Multi-Character Role-Playing Agent

1 code implementation21 Feb 2024 Xiaoyan Yu, Tongxu Luo, Yifan Wei, Fangyu Lei, Yiming Huang, Hao Peng, Liehuang Zhu

Large Language Models (LLMs) have revolutionized open-domain dialogue agents but encounter challenges in multi-character role-playing (MCRP) scenarios.

Incremental Learning

Event-level Knowledge Editing

1 code implementation20 Feb 2024 Hao Peng, Xiaozhi Wang, Chunyang Li, Kaisheng Zeng, Jiangshan Duo, Yixin Cao, Lei Hou, Juanzi Li

However, natural knowledge updates in the real world come from the occurrences of new events rather than direct changes in factual triplets.

knowledge editing Triplet

Data Engineering for Scaling Language Models to 128K Context

1 code implementation15 Feb 2024 Yao Fu, Rameswar Panda, Xinyao Niu, Xiang Yue, Hannaneh Hajishirzi, Yoon Kim, Hao Peng

We demonstrate that continual pretraining of the full model on 1B-5B tokens of such data is an effective and affordable strategy for scaling the context length of language models to 128K.

4k Continual Pretraining

PULSAR Effect: Revealing Potential Synergies in Combined Radiation Therapy and Immunotherapy via Differential Equations

no code implementations8 Feb 2024 Samiha Rouf, Casey Moore, Debabrata Saha, Dan Nguyen, MaryLena Bleile, Robert Timmerman, Hao Peng, Steve Jiang

Therefore, a synergistic effect between immunotherapy and PULSAR is observed when the pulses are spaced out by a certain number of days.

Executable Code Actions Elicit Better LLM Agents

2 code implementations1 Feb 2024 Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, Heng Ji

LLM agents are typically prompted to produce actions by generating JSON or text in a pre-defined format, which is usually limited by constrained action space (e. g., the scope of pre-defined tools) and restricted flexibility (e. g., inability to compose multiple tools).

Language Modelling Large Language Model

DARCS: Memory-Efficient Deep Compressed Sensing Reconstruction for Acceleration of 3D Whole-Heart Coronary MR Angiography

no code implementations1 Feb 2024 Zhihao Xue, Fan Yang, Juan Gao, Zhuo Chen, Hao Peng, Chao Zou, Hang Jin, Chenxi Hu

While unrolling-based deep reconstruction methods have achieved state-of-the-art performance on 2D image reconstruction, their application to 3D reconstruction is hindered by the large amount of memory needed to train an unrolled network.

3D Reconstruction De-aliasing +2

DeepRicci: Self-supervised Graph Structure-Feature Co-Refinement for Alleviating Over-squashing

no code implementations23 Jan 2024 Li Sun, Zhenhao Huang, Hua Wu, Junda Ye, Hao Peng, Zhengtao Yu, Philip S. Yu

Graph Neural Networks (GNNs) have shown great power for learning and mining on graphs, and Graph Structure Learning (GSL) plays an important role in boosting GNNs with a refined graph.

Contrastive Learning Graph structure learning

CLIP-Driven Semantic Discovery Network for Visible-Infrared Person Re-Identification

1 code implementation11 Jan 2024 Xiaoyan Yu, Neng Dong, Liehuang Zhu, Hao Peng, Dapeng Tao

Additionally, acknowledging the complementary nature of semantic details across different modalities, we integrate text features from the bimodal language descriptions to achieve comprehensive semantics.

Person Re-Identification

Motif-aware Riemannian Graph Neural Network with Generative-Contrastive Learning

1 code implementation2 Jan 2024 Li Sun, Zhenhao Huang, Zixi Wang, Feiyang Wang, Hao Peng, Philip Yu

In light of the issues above, we propose the problem of \emph{Motif-aware Riemannian Graph Representation Learning}, seeking a numerically stable encoder to capture motif regularity in a diverse-curvature manifold without labels.

Contrastive Learning Graph Neural Network +1

Mathematical Modeling of the Synergetic Effect between Radiotherapy and Immunotherapy

no code implementations28 Dec 2023 Yixun Xing, Casey Moore, Debabrata Saha, Dan Nguyen, MaryLena Bleile, Xun Jia, Robert Timmerman, Hao Peng, Steve Jiang

Achieving effective synergy between radiotherapy and immunotherapy is critical for optimizing tumor control and treatment outcomes.

Hierarchical and Incremental Structural Entropy Minimization for Unsupervised Social Event Detection

1 code implementation19 Dec 2023 Yuwei Cao, Hao Peng, Zhengtao Yu, Philip S. Yu

As a trending approach for social event detection, graph neural network (GNN)-based methods enable a fusion of natural language semantics and the complex social network structural information, thus showing SOTA performance.

Event Detection Graph Neural Network

Poincaré Differential Privacy for Hierarchy-Aware Graph Embedding

no code implementations19 Dec 2023 Yuecen Wei, Haonan Yuan, Xingcheng Fu, Qingyun Sun, Hao Peng, Xianxian Li, Chunming Hu

Specifically, PoinDP first learns the hierarchy weights for each entity based on the Poincar\'e model in hyperbolic space.

Graph Embedding Inductive Bias +3

Semi-Supervised Clustering via Structural Entropy with Different Constraints

1 code implementation18 Dec 2023 Guangjie Zeng, Hao Peng, Angsheng Li, Zhiwei Liu, Runze Yang, Chunyang Liu, Lifang He

In this work, we present Semi-supervised clustering via Structural Entropy (SSE), a novel method that can incorporate different types of constraints from diverse sources to perform both partitioning and hierarchical clustering.

Clustering

Prompt Based Tri-Channel Graph Convolution Neural Network for Aspect Sentiment Triplet Extraction

1 code implementation18 Dec 2023 Kun Peng, Lei Jiang, Hao Peng, Rui Liu, Zhengtao Yu, Jiaqian Ren, Zhifeng Hao, Philip S. Yu

Aspect Sentiment Triplet Extraction (ASTE) is an emerging task to extract a given sentence's triplets, which consist of aspects, opinions, and sentiments.

Aspect Sentiment Triplet Extraction Triplet

Adversarial Socialbots Modeling Based on Structural Information Principles

1 code implementation13 Dec 2023 Xianghua Zeng, Hao Peng, Angsheng Li

The importance of effective detection is underscored by the fact that socialbots imitate human behavior to propagate misinformation, leading to an ongoing competition between socialbots and detectors.

Misinformation

Measuring and Improving Attentiveness to Partial Inputs with Counterfactuals

no code implementations16 Nov 2023 Yanai Elazar, Bhargavi Paranjape, Hao Peng, Sarah Wiegreffe, Khyathi Raghavi, Vivek Srikumar, Sameer Singh, Noah A. Smith

Previous work has found that datasets with paired inputs are prone to correlations between a specific part of the input (e. g., the hypothesis in NLI) and the label; consequently, models trained only on those outperform chance.

counterfactual In-Context Learning +2

Examining LLMs' Uncertainty Expression Towards Questions Outside Parametric Knowledge

1 code implementation16 Nov 2023 Genglin Liu, Xingyao Wang, Lifan Yuan, Yangyi Chen, Hao Peng

Can large language models (LLMs) express their uncertainty in situations where they lack sufficient parametric knowledge to generate reasonable responses?

Question Answering valid

MAVEN-Arg: Completing the Puzzle of All-in-One Event Understanding Dataset with Event Argument Annotation

1 code implementation15 Nov 2023 Xiaozhi Wang, Hao Peng, Yong Guan, Kaisheng Zeng, Jianhui Chen, Lei Hou, Xu Han, Yankai Lin, Zhiyuan Liu, Ruobing Xie, Jie zhou, Juanzi Li

Understanding events in texts is a core objective of natural language understanding, which requires detecting event occurrences, extracting event arguments, and analyzing inter-event relationships.

Event Argument Extraction Event Detection +3

When does In-context Learning Fall Short and Why? A Study on Specification-Heavy Tasks

no code implementations15 Nov 2023 Hao Peng, Xiaozhi Wang, Jianhui Chen, Weikai Li, Yunjia Qi, Zimu Wang, Zhili Wu, Kaisheng Zeng, Bin Xu, Lei Hou, Juanzi Li

In this paper, we find that ICL falls short of handling specification-heavy tasks, which are tasks with complicated and extensive task specifications, requiring several hours for ordinary humans to master, such as traditional information extraction tasks.

In-Context Learning

JPAVE: A Generation and Classification-based Model for Joint Product Attribute Prediction and Value Extraction

1 code implementation7 Nov 2023 Zhongfen Deng, Hao Peng, Tao Zhang, Shuaiqi Liu, Wenting Zhao, Yibo Wang, Philip S. Yu

Furthermore, the copy mechanism in value generator and the value attention module in value classifier help our model address the data discrepancy issue by only focusing on the relevant part of input text and ignoring other information which causes the discrepancy issue such as sentence structure in the text.

Attribute Attribute Value Extraction +4

MultiSPANS: A Multi-range Spatial-Temporal Transformer Network for Traffic Forecast via Structural Entropy Optimization

1 code implementation6 Nov 2023 Dongcheng Zou, Senzhang Wang, Xuefeng Li, Hao Peng, Yuandong Wang, Chunyang Liu, Kehua Sheng, Bo Zhang

Based on this, we propose a relative structural entropy-based position encoding and a multi-head attention masking scheme based on multi-layer encoding trees.

Management Position +2

Uncertainty-guided Boundary Learning for Imbalanced Social Event Detection

1 code implementation30 Oct 2023 Jiaqian Ren, Hao Peng, Lei Jiang, Zhiwei Liu, Jia Wu, Zhengtao Yu, Philip S. Yu

While in our observation, compared to the rarity of classes, the calibrated uncertainty estimated from well-trained evidential deep learning networks better reflects model performance.

Contrastive Learning Event Detection

Language Models Hallucinate, but May Excel at Fact Verification

1 code implementation23 Oct 2023 Jian Guan, Jesse Dodge, David Wadden, Minlie Huang, Hao Peng

Recent progress in natural language processing (NLP) owes much to remarkable advances in large language models (LLMs).

Fact Verification Hallucination

Knowledge Graph Context-Enhanced Diversified Recommendation

1 code implementation20 Oct 2023 Xiaolong Liu, Liangwei Yang, Zhiwei Liu, Mingdai Yang, Chen Wang, Hao Peng, Philip S. Yu

Collectively, our contributions signify a substantial stride towards augmenting the panorama of recommendation diversity within the realm of KG-informed RecSys paradigms.

Diversity Knowledge Graphs +1

Multi-omics Sampling-based Graph Transformer for Synthetic Lethality Prediction

no code implementations17 Oct 2023 Xusheng Zhao, Hao liu, Qiong Dai, Hao Peng, Xu Bai, Huailiang Peng

We showcase the effectiveness of MSGT-SL on real-world SL tasks, demonstrating the empirical benefits gained from the graph transformer and multi-omics data.

Edge Classification

Privacy in Large Language Models: Attacks, Defenses and Future Directions

no code implementations16 Oct 2023 Haoran Li, Yulin Chen, Jinglong Luo, Jiecong Wang, Hao Peng, Yan Kang, Xiaojin Zhang, Qi Hu, Chunkit Chan, Zenglin Xu, Bryan Hooi, Yangqiu Song

The advancement of large language models (LLMs) has significantly enhanced the ability to effectively tackle various downstream NLP tasks and unify these tasks into generative pipelines.

TRAM: Bridging Trust Regions and Sharpness Aware Minimization

1 code implementation5 Oct 2023 Tom Sherborne, Naomi Saphra, Pradeep Dasigi, Hao Peng

We propose Trust Region Aware Minimization (TRAM), a SAM algorithm fine-tuning for low parameter sharpness and smooth, informative representations preserving pre-trained structure.

Domain Generalization Language Modeling +2

CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets

1 code implementation29 Sep 2023 Lifan Yuan, Yangyi Chen, Xingyao Wang, Yi R. Fung, Hao Peng, Heng Ji

It creates toolsets specifically curated for the tasks and equips LLMs with a component that retrieves tools from these sets to enhance their capability to solve complex tasks.

Language Modelling Mathematical Reasoning

OmniEvent: A Comprehensive, Fair, and Easy-to-Use Toolkit for Event Understanding

1 code implementation25 Sep 2023 Hao Peng, Xiaozhi Wang, Feng Yao, Zimu Wang, Chuzhao Zhu, Kaisheng Zeng, Lei Hou, Juanzi Li

Event understanding aims at understanding the content and relationship of events within texts, which covers multiple complicated information extraction tasks: event detection, event argument extraction, and event relation extraction.

Event Argument Extraction Event Detection +2

MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback

1 code implementation19 Sep 2023 Xingyao Wang, Zihan Wang, Jiateng Liu, Yangyi Chen, Lifan Yuan, Hao Peng, Heng Ji

However, current evaluation protocols often emphasize benchmark performance with single-turn exchanges, neglecting the nuanced interactions among the user, LLMs, and external tools, while also underestimating the importance of natural language feedback from users.

Decision Making

Unsupervised Skin Lesion Segmentation via Structural Entropy Minimization on Multi-Scale Superpixel Graphs

1 code implementation5 Sep 2023 Guangjie Zeng, Hao Peng, Angsheng Li, Zhiwei Liu, Chunyang Liu, Philip S. Yu, Lifang He

In this work, we propose a novel unsupervised Skin Lesion sEgmentation framework based on structural entropy and isolation forest outlier Detection, namely SLED.

Lesion Segmentation Outlier Detection +2

LM-Infinite: Zero-Shot Extreme Length Generalization for Large Language Models

1 code implementation30 Aug 2023 Chi Han, Qifan Wang, Hao Peng, Wenhan Xiong, Yu Chen, Heng Ji, Sinong Wang

As a result, their performance suffers drastically on inputs longer than those encountered during training, substantially limiting their applications in real-world tasks involving long contexts such as encoding scientific articles, code repositories, or long dialogues.

2k 4k +1

Multi-task Item-attribute Graph Pre-training for Strict Cold-start Item Recommendation

1 code implementation26 Jun 2023 Yuwei Cao, Liangwei Yang, Chen Wang, Zhiwei Liu, Hao Peng, Chenyu You, Philip S. Yu

We explore the role of the fine-grained item attributes in bridging the gaps between the existing and the SCS items and pre-train a knowledgeable item-attribute graph for SCS item recommendation.

Attribute Multi-Task Learning +1

Sequential Recommendation with Controllable Diversification: Representation Degeneration and Diversity

1 code implementation21 Jun 2023 Ziwei Fan, Zhiwei Liu, Hao Peng, Philip S. Yu

We then propose a novel Singular sPectrum sMoothing regularization for Recommendation (SPMRec), which acts as a controllable surrogate to alleviate the degeneration and achieve the balance between recommendation diversity and performance.

Diversity Sequential Recommendation

The Devil is in the Details: On the Pitfalls of Event Extraction Evaluation

1 code implementation12 Jun 2023 Hao Peng, Xiaozhi Wang, Feng Yao, Kaisheng Zeng, Lei Hou, Juanzi Li, Zhiyuan Liu, Weixing Shen

In this paper, we check the reliability of EE evaluations and identify three major pitfalls: (1) The data preprocessing discrepancy makes the evaluation results on the same dataset not directly comparable, but the data preprocessing details are not widely noted and specified in papers.

Event Argument Extraction Event Detection +1

Chain-of-Thought Hub: A Continuous Effort to Measure Large Language Models' Reasoning Performance

1 code implementation26 May 2023 Yao Fu, Litu Ou, Mingyu Chen, Yuhao Wan, Hao Peng, Tushar Khot

As large language models (LLMs) are continuously being developed, their evaluation becomes increasingly important yet challenging.

LeTI: Learning to Generate from Textual Interactions

1 code implementation17 May 2023 Xingyao Wang, Hao Peng, Reyhaneh Jabbarvand, Heng Ji

We explore LMs' potential to learn from textual interactions (LETI) that not only check their correctness with binary labels but also pinpoint and explain errors in their outputs through textual feedback.

Code Generation Event Argument Extraction +1

Improving Language Model Negotiation with Self-Play and In-Context Learning from AI Feedback

1 code implementation17 May 2023 Yao Fu, Hao Peng, Tushar Khot, Mirella Lapata

We study whether multiple large language models (LLMs) can autonomously improve each other in a negotiation game by playing, reflecting, and criticizing.

In-Context Learning Language Modeling +2

Contrastive Graph Clustering in Curvature Spaces

no code implementations5 May 2023 Li Sun, Feiyang Wang, Junda Ye, Hao Peng, Philip S. Yu

On the other hand, contrastive learning boosts the deep graph clustering but usually struggles in either graph augmentation or hard sample mining.

Clustering Contrastive Learning +1

Hierarchical State Abstraction Based on Structural Information Principles

1 code implementation24 Apr 2023 Xianghua Zeng, Hao Peng, Angsheng Li, Chunyang Liu, Lifang He, Philip S. Yu

State abstraction optimizes decision-making by ignoring irrelevant environmental information in reinforcement learning with rich observations.

continuous-control Continuous Control +2

Hyperbolic Geometric Graph Representation Learning for Hierarchy-imbalance Node Classification

1 code implementation11 Apr 2023 Xingcheng Fu, Yuecen Wei, Qingyun Sun, Haonan Yuan, Jia Wu, Hao Peng, JianXin Li

We find that training labeled nodes with different hierarchical properties have a significant impact on the node classification tasks and confirm it in our experiments.

Graph Representation Learning Node Classification

Graph Collaborative Signals Denoising and Augmentation for Recommendation

1 code implementation6 Apr 2023 Ziwei Fan, Ke Xu, Zhang Dong, Hao Peng, Jiawei Zhang, Philip S. Yu

Moreover, we show that the inclusion of user-user and item-item correlations can improve recommendations for users with both abundant and insufficient interactions.

Collaborative Filtering Denoising +1

Effective and Stable Role-Based Multi-Agent Collaboration by Structural Information Principles

1 code implementation3 Apr 2023 Xianghua Zeng, Hao Peng, Angsheng Li

Role-based learning is a promising approach to improving the performance of Multi-Agent Reinforcement Learning (MARL).

Multi-agent Reinforcement Learning Starcraft +1

Reinforcement Learning Guided Multi-Objective Exam Paper Generation

1 code implementation2 Mar 2023 Yuhu Shang, Xuexiong Luo, Lihong Wang, Hao Peng, Xiankun Zhang, Yimeng Ren, Kun Liang

To reduce the repetitive and complex work of instructors, exam paper generation (EPG) technique has become a salient topic in the intelligent education field, which targets at generating high-quality exam paper automatically according to instructor-specified assessment criteria.

Knowledge Tracing Paper generation +3

A Comprehensive Survey on Pretrained Foundation Models: A History from BERT to ChatGPT

no code implementations18 Feb 2023 Ce Zhou, Qian Li, Chen Li, Jun Yu, Yixin Liu, Guangjing Wang, Kai Zhang, Cheng Ji, Qiben Yan, Lifang He, Hao Peng, JianXin Li, Jia Wu, Ziwei Liu, Pengtao Xie, Caiming Xiong, Jian Pei, Philip S. Yu, Lichao Sun

This study provides a comprehensive review of recent research advancements, challenges, and opportunities for PFMs in text, image, graph, as well as other data modalities.

Graph Learning Language Modelling +1

A Comprehensive Survey on Automatic Knowledge Graph Construction

no code implementations10 Feb 2023 Lingfeng Zhong, Jia Wu, Qian Li, Hao Peng, Xindong Wu

A knowledge graph is built in three steps: knowledge acquisition, knowledge refinement, and knowledge evolution.

graph construction Survey

Specializing Smaller Language Models towards Multi-Step Reasoning

2 code implementations30 Jan 2023 Yao Fu, Hao Peng, Litu Ou, Ashish Sabharwal, Tushar Khot

by paying the price of decreased generic ability, we can clearly lift up the scaling curve of models smaller than 10B towards a specialized multi-step math reasoning ability.

Math Model Selection

Unbiased and Efficient Self-Supervised Incremental Contrastive Learning

1 code implementation28 Jan 2023 Cheng Ji, JianXin Li, Hao Peng, Jia Wu, Xingcheng Fu, Qingyun Sun, Phillip S. Yu

Contrastive Learning (CL) has been proved to be a powerful self-supervised approach for a wide range of domains, including computer vision and graph representation learning.

Contrastive Learning Graph Representation Learning +1

Mutual Wasserstein Discrepancy Minimization for Sequential Recommendation

1 code implementation28 Jan 2023 Ziwei Fan, Zhiwei Liu, Hao Peng, Philip S Yu

Wasserstein Discrepancy Measurement builds upon the 2-Wasserstein distance, which is more robust, more efficient in small batch sizes, and able to model the uncertainty of stochastic augmentation processes.

Contrastive Learning Mutual Information Estimation +2

Parallel Multi-Extended State Observers based {ADRC} with Application to High-Speed Precision Motion Stage

no code implementations18 Jan 2023 Guojie Tang, Wenchao Xue, Hao Peng, Yanlong Zhao, Zhijun Yang

In particular, the algorithm for calculating the tracking error caused by single ESO's estimation error is constructed.

Friction

Self-organization Preserved Graph Structure Learning with Principle of Relevant Information

no code implementations30 Dec 2022 Qingyun Sun, JianXin Li, Beining Yang, Xingcheng Fu, Hao Peng, Philip S. Yu

Most Graph Neural Networks follow the message-passing paradigm, assuming the observed structure depicts the ground-truth node relationships.

Graph structure learning

Self-Supervised Continual Graph Learning in Adaptive Riemannian Spaces

no code implementations30 Nov 2022 Li Sun, Junda Ye, Hao Peng, Feiyang Wang, Philip S. Yu

On the one hand, existing methods work with the zero-curvature Euclidean space, and largely ignore the fact that curvature varies over the coming graph sequence.

Graph Learning

MAVEN-ERE: A Unified Large-scale Dataset for Event Coreference, Temporal, Causal, and Subevent Relation Extraction

1 code implementation14 Nov 2022 Xiaozhi Wang, Yulin Chen, Ning Ding, Hao Peng, Zimu Wang, Yankai Lin, Xu Han, Lei Hou, Juanzi Li, Zhiyuan Liu, Peng Li, Jie zhou

It contains 103, 193 event coreference chains, 1, 216, 217 temporal relations, 57, 992 causal relations, and 15, 841 subevent relations, which is larger than existing datasets of all the ERE tasks by at least an order of magnitude.

Event Relation Extraction Relation +1

How Much Does Attention Actually Attend? Questioning the Importance of Attention in Pretrained Transformers

1 code implementation7 Nov 2022 Michael Hassid, Hao Peng, Daniel Rotem, Jungo Kasai, Ivan Montero, Noah A. Smith, Roy Schwartz

Our results motivate research on simpler alternatives to input-dependent attention, as well as on methods for better utilization of this mechanism in the Transformer architecture.

Ranking-based Group Identification via Factorized Attention on Social Tripartite Graph

1 code implementation2 Nov 2022 Mingdai Yang, Zhiwei Liu, Liangwei Yang, Xiaolong Liu, Chen Wang, Hao Peng, Philip S. Yu

PA layers efficiently learn the relatedness of non-neighbor nodes to improve the information propagation to users.

Sequential Recommendation with Auxiliary Item Relationships via Multi-Relational Transformer

1 code implementation24 Oct 2022 Ziwei Fan, Zhiwei Liu, Chen Wang, Peijie Huang, Hao Peng, Philip S. Yu

However, it remains a significant challenge to model auxiliary item relationships in SR. To simultaneously model high-order item-item transitions in sequences and auxiliary item relationships, we propose a Multi-relational Transformer capable of modeling auxiliary item relationships for SR (MT4SR).

Sequential Recommendation

DAGAD: Data Augmentation for Graph Anomaly Detection

1 code implementation18 Oct 2022 Fanzhen Liu, Xiaoxiao Ma, Jia Wu, Jian Yang, Shan Xue, Amin Beheshti, Chuan Zhou, Hao Peng, Quan Z. Sheng, Charu C. Aggarwal

To bridge the gaps, this paper devises a novel Data Augmentation-based Graph Anomaly Detection (DAGAD) framework for attributed graphs, equipped with three specially designed modules: 1) an information fusion module employing graph neural network encoders to learn representations, 2) a graph data augmentation module that fertilizes the training set with generated samples, and 3) an imbalance-tailored learning module to discriminate the distributions of the minority (anomalous) and majority (normal) classes.

Data Augmentation Graph Anomaly Detection +1

Modeling Context With Linear Attention for Scalable Document-Level Translation

1 code implementation16 Oct 2022 Zhaofeng Wu, Hao Peng, Nikolaos Pappas, Noah A. Smith

Document-level machine translation leverages inter-sentence dependencies to produce more coherent and consistent translations.

Document Level Machine Translation Document Translation +4

Transparency Helps Reveal When Language Models Learn Meaning

1 code implementation14 Oct 2022 Zhaofeng Wu, William Merrill, Hao Peng, Iz Beltagy, Noah A. Smith

Many current NLP systems are built from language models trained to optimize unsupervised objectives on large amounts of raw text.

Complexity-Based Prompting for Multi-Step Reasoning

no code implementations3 Oct 2022 Yao Fu, Hao Peng, Ashish Sabharwal, Peter Clark, Tushar Khot

In this work, we propose complexity-based prompting, a simple and effective example selection scheme for multi-step reasoning.

Date Understanding GSM8K +2

Heterogeneous Graph Neural Network for Privacy-Preserving Recommendation

1 code implementation2 Oct 2022 Yuecen Wei, Xingcheng Fu, Qingyun Sun, Hao Peng, Jia Wu, Jinyan Wang, Xianxian Li

To address this issue, we propose a novel heterogeneous graph neural network privacy-preserving method based on a differential privacy mechanism named HeteDP, which provides a double guarantee on graph features and topology.

Graph Neural Network Privacy Preserving

Information Extraction and Human-Robot Dialogue towards Real-life Tasks: A Baseline Study with the MobileCS Dataset

1 code implementation27 Sep 2022 Hong Liu, Hao Peng, Zhijian Ou, Juanzi Li, Yi Huang, Junlan Feng

Recently, there have merged a class of task-oriented dialogue (TOD) datasets collected through Wizard-of-Oz simulated games.

Cross-Network Social User Embedding with Hybrid Differential Privacy Guarantees

1 code implementation4 Sep 2022 Jiaqian Ren, Lei Jiang, Hao Peng, Lingjuan Lyu, Zhiwei Liu, Chaochao Chen, Jia Wu, Xu Bai, Philip S. Yu

Integrating multiple online social networks (OSNs) has important implications for many downstream social mining tasks, such as user preference modelling, recommendation, and link prediction.

Attribute Link Prediction +2

A Self-supervised Riemannian GNN with Time Varying Curvature for Temporal Graph Learning

no code implementations30 Aug 2022 Li Sun, Junda Ye, Hao Peng, Philip S. Yu

To bridge this gap, we make the first attempt to study the problem of self-supervised temporal graph representation learning in the general Riemannian space, supporting the time-varying curvature to shift among hyperspherical, Euclidean and hyperbolic spaces.

Graph Learning Graph Neural Network +2

Position-aware Structure Learning for Graph Topology-imbalance by Relieving Under-reaching and Over-squashing

1 code implementation17 Aug 2022 Qingyun Sun, JianXin Li, Haonan Yuan, Xingcheng Fu, Hao Peng, Cheng Ji, Qian Li, Philip S. Yu

Topology-imbalance is a graph-specific imbalance problem caused by the uneven topology positions of labeled nodes, which significantly damages the performance of GNNs.

Graph Learning Graph structure learning +2

Automating DBSCAN via Deep Reinforcement Learning

2 code implementations9 Aug 2022 Ruitong Zhang, Hao Peng, Yingtong Dou, Jia Wu, Qingyun Sun, Jingyi Zhang, Philip S. Yu

DBSCAN is widely used in many scientific and engineering fields because of its simplicity and practicality.

Clustering Computational Efficiency +4

A Challenge on Semi-Supervised and Reinforced Task-Oriented Dialog Systems

1 code implementation6 Jul 2022 Zhijian Ou, Junlan Feng, Juanzi Li, Yakun Li, Hong Liu, Hao Peng, Yi Huang, Jiangjiang Zhao

A challenge on Semi-Supervised and Reinforced Task-Oriented Dialog Systems, Co-located with EMNLP2022 SereTOD Workshop.

BOND: Benchmarking Unsupervised Outlier Node Detection on Static Attributed Graphs

2 code implementations21 Jun 2022 Kay Liu, Yingtong Dou, Yue Zhao, Xueying Ding, Xiyang Hu, Ruitong Zhang, Kaize Ding, Canyu Chen, Hao Peng, Kai Shu, Lichao Sun, Jundong Li, George H. Chen, Zhihao Jia, Philip S. Yu

To bridge this gap, we present--to the best of our knowledge--the first comprehensive benchmark for unsupervised outlier node detection on static attributed graphs called BOND, with the following highlights.

Anomaly Detection Benchmarking +2

Graph-level Neural Networks: Current Progress and Future Directions

no code implementations31 May 2022 Ge Zhang, Jia Wu, Jian Yang, Shan Xue, Wenbin Hu, Chuan Zhou, Hao Peng, Quan Z. Sheng, Charu Aggarwal

To frame this survey, we propose a systematic taxonomy covering GLNNs upon deep neural networks, graph neural networks, and graph pooling.

Survey