Search Results for author: Yongqiang Chen

Found 26 papers, 15 papers with code

Retrieval-Augmented Generation with Hierarchical Knowledge

1 code implementation13 Mar 2025 Haoyu Huang, Yongfeng Huang, Junjie Yang, Zhenyu Pan, Yongqiang Chen, Kaili Ma, Hongzhi Chen, James Cheng

Graph-based Retrieval-Augmented Generation (RAG) methods have significantly enhanced the performance of large language models (LLMs) in domain-specific tasks.

Multi-hop Question Answering Question Answering +2

Can Large Language Models Help Experimental Design for Causal Discovery?

no code implementations3 Mar 2025 Junyi Li, Yongqiang Chen, Chenxi Liu, Qianyi Cai, Tongliang Liu, Bo Han, Kun Zhang, Hui Xiong

Designing proper experiments and selecting optimal intervention targets is a longstanding problem in scientific or causal discovery.

Causal Discovery Experimental Design +3

DivIL: Unveiling and Addressing Over-Invariance for Out-of- Distribution Generalization

1 code implementation18 Feb 2025 Jiaqi Wang, YuHang Zhou, Zhixiong Zhang, Qiguang Chen, Yongqiang Chen, James Cheng

A popular approach to addressing this issue is invariant learning (IL), in which the model is compiled to focus on invariant features instead of spurious features by adding strong constraints during training.

Contrastive Learning Out-of-Distribution Generalization

BrainOOD: Out-of-distribution Generalizable Brain Network Analysis

1 code implementation2 Feb 2025 Jiaxing Xu, Yongqiang Chen, Xia Dong, Mengcheng Lan, Tiancheng Huang, Qingtian Bian, James Cheng, Yiping Ke

Graph Neural Networks (GNNs) have shown promising in analyzing brain networks, but there are two major challenges in using GNNs: (1) distribution shifts in multi-site brain network data, leading to poor Out-of-Distribution (OOD) generalization, and (2) limited interpretability in identifying key brain regions critical to neurological disorders.

Eliciting Causal Abilities in Large Language Models for Reasoning Tasks

1 code implementation19 Dec 2024 Yajing Wang, Zongwei Luo, Jingzhe Wang, Zhanke Zhou, Yongqiang Chen, Bo Han

In SCIE, the instructions are treated as the treatment, and textual features are used to process natural language, establishing causal relationships through treatments between instructions and downstream tasks.

Causal Inference

On the Comparison between Multi-modal and Single-modal Contrastive Learning

no code implementations5 Nov 2024 Wei Huang, Andi Han, Yongqiang Chen, Yuan Cao, Zhiqiang Xu, Taiji Suzuki

Our analysis provides a unified framework that can characterize the optimization and generalization of both single-modal and multi-modal contrastive learning.

Contrastive Learning Learning Theory

UniMoT: Unified Molecule-Text Language Model with Discrete Token Representation

no code implementations1 Aug 2024 Juzheng Zhang, Yatao Bian, Yongqiang Chen, Quanming Yao

Equipped with this tokenizer, UniMoT can unify molecule and text modalities under a shared token representation and an autoregressive training paradigm, enabling it to interpret molecules as a foreign language and generate them as text.

Language Modeling Language Modelling +1

Empowering Graph Invariance Learning with Deep Spurious Infomax

1 code implementation13 Jul 2024 Tianjun Yao, Yongqiang Chen, Zhenhao Chen, Kai Hu, Zhiqiang Shen, Kun Zhang

To bridge this gap, we introduce a novel graph invariance learning paradigm, which induces a robust and general inductive bias.

Inductive Bias

HIGHT: Hierarchical Graph Tokenization for Graph-Language Alignment

no code implementations20 Jun 2024 Yongqiang Chen, Quanming Yao, Juzheng Zhang, James Cheng, Yatao Bian

As LLMs are predominantly trained with 1D text data, most existing approaches adopt a graph neural network to represent a graph as a series of node tokens and feed these tokens to LLMs for graph-language alignment.

Graph Neural Network Hallucination

How Interpretable Are Interpretable Graph Neural Networks?

1 code implementation12 Jun 2024 Yongqiang Chen, Yatao Bian, Bo Han, James Cheng

Extracting the desired interpretable subgraph requires an accurate approximation of SubMT, yet we find that the existing XGNNs can have a huge gap in fitting SubMT.

Graph Classification

A Sober Look at the Robustness of CLIPs to Spurious Features

no code implementations18 Mar 2024 Qizhou Wang, Yong Lin, Yongqiang Chen, Ludwig Schmidt, Bo Han, Tong Zhang

Large vision language models, such as CLIP, demonstrate impressive robustness to spurious features than single-modal models trained on ImageNet.

Benchmarking

Discovery of the Hidden World with Large Language Models

no code implementations6 Feb 2024 Chenxi Liu, Yongqiang Chen, Tongliang Liu, Mingming Gong, James Cheng, Bo Han, Kun Zhang

Meanwhile, COAT also adopts CDs to find causal relations among the identified variables as well as to provide feedback to LLMs to iteratively refine the proposed factors.

Causal Discovery

Enhancing Neural Subset Selection: Integrating Background Information into Set Representations

no code implementations5 Feb 2024 Binghui Xie, Yatao Bian, Kaiwen Zhou, Yongqiang Chen, Peilin Zhao, Bo Han, Wei Meng, James Cheng

Learning neural subset selection tasks, such as compound selection in AI-aided drug discovery, have become increasingly pivotal across diverse applications.

Drug Discovery

Enhancing Evolving Domain Generalization through Dynamic Latent Representations

no code implementations16 Jan 2024 Binghui Xie, Yongqiang Chen, Jiaqi Wang, Kaiwen Zhou, Bo Han, Wei Meng, James Cheng

However, in non-stationary tasks where new domains evolve in an underlying continuous structure, such as time, merely extracting the invariant features is insufficient for generalization to the evolving new domains.

Evolving Domain Generalization

Positional Information Matters for Invariant In-Context Learning: A Case Study of Simple Function Classes

no code implementations30 Nov 2023 Yongqiang Chen, Binghui Xie, Kaiwen Zhou, Bo Han, Yatao Bian, James Cheng

Surprisingly, DeepSet outperforms transformers across a variety of distribution shifts, implying that preserving permutation invariance symmetry to input demonstrations is crucial for OOD ICL.

In-Context Learning

Does Invariant Graph Learning via Environment Augmentation Learn Invariance?

1 code implementation NeurIPS 2023 Yongqiang Chen, Yatao Bian, Kaiwen Zhou, Binghui Xie, Bo Han, James Cheng

Invariant graph representation learning aims to learn the invariance among data from different environments for out-of-distribution generalization on graphs.

Graph Learning Graph Representation Learning +1

Towards out-of-distribution generalizable predictions of chemical kinetics properties

1 code implementation4 Oct 2023 ZiHao Wang, Yongqiang Chen, Yang Duan, Weijiang Li, Bo Han, James Cheng, Hanghang Tong

Under this framework, we create comprehensive datasets to benchmark (1) the state-of-the-art ML approaches for reaction prediction in the OOD setting and (2) the state-of-the-art graph OOD methods in kinetics property prediction problems.

Prediction Property Prediction

Understanding and Improving Feature Learning for Out-of-Distribution Generalization

1 code implementation NeurIPS 2023 Yongqiang Chen, Wei Huang, Kaiwen Zhou, Yatao Bian, Bo Han, James Cheng

Moreover, when fed the ERM learned features to the OOD objectives, the invariant feature learning quality significantly affects the final OOD performance, as OOD objectives rarely learn new features.

Out-of-Distribution Generalization

Dataset and Baseline System for Multi-lingual Extraction and Normalization of Temporal and Numerical Expressions

1 code implementation31 Mar 2023 Sanxing Chen, Yongqiang Chen, Börje F. Karlsson

Temporal and numerical expression understanding is of great importance in many downstream Natural Language Processing (NLP) and Information Retrieval (IR) tasks.

Date Understanding Information Retrieval +2

RLTP: Reinforcement Learning to Pace for Delayed Impression Modeling in Preloaded Ads

no code implementations6 Feb 2023 Penghui Wei, Yongqiang Chen, Shaoguo Liu, Liang Wang, Bo Zheng

In a whole delivery period, advertisers usually desire a certain impression count for the ads, and they also expect that the delivery performance is as good as possible (e. g., obtaining high click-through rate).

reinforcement-learning Reinforcement Learning (RL)

Understanding and Improving Graph Injection Attack by Promoting Unnoticeability

1 code implementation ICLR 2022 Yongqiang Chen, Han Yang, Yonggang Zhang, Kaili Ma, Tongliang Liu, Bo Han, James Cheng

Recently Graph Injection Attack (GIA) emerges as a practical attack scenario on Graph Neural Networks (GNNs), where the adversary can merely inject few malicious nodes instead of modifying existing nodes or edges, i. e., Graph Modification Attack (GMA).

Learning Causally Invariant Representations for Out-of-Distribution Generalization on Graphs

3 code implementations11 Feb 2022 Yongqiang Chen, Yonggang Zhang, Yatao Bian, Han Yang, Kaili Ma, Binghui Xie, Tongliang Liu, Bo Han, James Cheng

Despite recent success in using the invariance principle for out-of-distribution (OOD) generalization on Euclidean data (e. g., images), studies on graph data are still limited.

Drug Discovery Graph Learning +1

Calibrating and Improving Graph Contrastive Learning

1 code implementation27 Jan 2021 Kaili Ma, Haochen Yang, Han Yang, Yongqiang Chen, James Cheng

To assess the discrepancy between the prediction and the ground-truth in the downstream tasks for these contrastive pairs, we adapt the expected calibration error (ECE) to graph contrastive learning.

Contrastive Learning Graph Clustering +4

Self-Enhanced GNN: Improving Graph Neural Networks Using Model Outputs

1 code implementation18 Feb 2020 Han Yang, Xiao Yan, Xinyan Dai, Yongqiang Chen, James Cheng

In this paper, we propose self-enhanced GNN (SEG), which improves the quality of the input data using the outputs of existing GNN models for better performance on semi-supervised node classification.

General Classification Node Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.