Search Results for author: Ling Yang

Found 24 papers, 12 papers with code

Motion-aware Latent Diffusion Models for Video Frame Interpolation

no code implementations21 Apr 2024 Zhilin Huang, Yijie Yu, Ling Yang, Chujun Qin, Bing Zheng, Xiawu Zheng, Zikun Zhou, YaoWei Wang, Wenming Yang

With the advancement of AIGC, video frame interpolation (VFI) has become a crucial component in existing video generation frameworks, attracting widespread research interest.

Motion Estimation Video Frame Interpolation +1

Distribution-Aware Data Expansion with Diffusion Models

1 code implementation11 Mar 2024 Haowei Zhu, Ling Yang, Jun-Hai Yong, Wentao Zhang, Bin Wang

In this paper, we propose DistDiff, an effective data expansion framework based on the distribution-aware diffusion model.

Image Generation Informativeness

Retrieval-Augmented Generation for AI-Generated Content: A Survey

2 code implementations29 Feb 2024 Penghao Zhao, Hailin Zhang, Qinhan Yu, Zhengren Wang, Yunteng Geng, Fangcheng Fu, Ling Yang, Wentao Zhang, Jie Jiang, Bin Cui

We first classify RAG foundations according to how the retriever augments the generator, distilling the fundamental abstractions of the augmentation methodologies for various retrievers and generators.

Information Retrieval Large Language Model +2

Structure-Guided Adversarial Training of Diffusion Models

no code implementations27 Feb 2024 Ling Yang, Haotian Qian, Zhilong Zhang, Jingwei Liu, Bin Cui

In this pioneering approach, we compel the model to learn manifold structures between samples in each training batch.

Conditional Image Generation Denoising

Cross-Modal Contextualized Diffusion Models for Text-Guided Visual Generation and Editing

1 code implementation26 Feb 2024 Ling Yang, Zhilong Zhang, Zhaochen Yu, Jingwei Liu, Minkai Xu, Stefano Ermon, Bin Cui

To address this issue, we propose a novel and general contextualized diffusion model (ContextDiff) by incorporating the cross-modal context encompassing interactions and alignments between text condition and visual sample into forward and reverse processes.

Text-to-Image Generation Text-to-Video Editing +1

RealCompo: Dynamic Equilibrium between Realism and Compositionality Improves Text-to-Image Diffusion Models

2 code implementations20 Feb 2024 Xinchen Zhang, Ling Yang, Yaqi Cai, Zhaochen Yu, Jiake Xie, Ye Tian, Minkai Xu, Yong Tang, Yujiu Yang, Bin Cui

In this paper, we propose a new training-free and transferred-friendly text-to-image generation framework, namely RealCompo, which aims to leverage the advantages of text-to-image and layout-to-image models to enhance both realism and compositionality of the generated images.

Denoising Text-to-Image Generation

Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs

1 code implementation22 Jan 2024 Ling Yang, Zhaochen Yu, Chenlin Meng, Minkai Xu, Stefano Ermon, Bin Cui

In this paper, we propose a brand new training-free text-to-image generation/editing framework, namely Recaption, Plan and Generate (RPG), harnessing the powerful chain-of-thought reasoning ability of multimodal LLMs to enhance the compositionality of text-to-image diffusion models.

Diffusion Personalization Tuning Free Large Language Model

Binding-Adaptive Diffusion Models for Structure-Based Drug Design

1 code implementation15 Jan 2024 Zhilin Huang, Ling Yang, Zaixi Zhang, Xiangxin Zhou, Yu Bao, Xiawu Zheng, Yuwei Yang, Yu Wang, Wenming Yang

Then the selected protein-ligand subcomplex is processed with SE(3)-equivariant neural networks, and transmitted back to each atom of the complex for augmenting the target-aware 3D molecule diffusion generation with binding interaction information.

Avg

MedXChat: Bridging CXR Modalities with a Unified Multimodal Large Model

no code implementations4 Dec 2023 Ling Yang, Zhanyu Wang, Luping Zhou

Despite the success of Large Language Models (LLMs) in general image tasks, a gap persists in the medical field for a multimodal large model adept at handling the nuanced diversity of medical images.

Instruction Following Question Answering +1

VQGraph: Rethinking Graph Representation Space for Bridging GNNs and MLPs

1 code implementation4 Aug 2023 Ling Yang, Ye Tian, Minkai Xu, Zhongyi Liu, Shenda Hong, Wei Qu, Wentao Zhang, Bin Cui, Muhan Zhang, Jure Leskovec

To address this issue, we propose to learn a new powerful graph representation space by directly labeling nodes' diverse local structures for GNN-to-MLP distillation.

Knowledge Distillation Quantization +1

Individual and Structural Graph Information Bottlenecks for Out-of-Distribution Generalization

1 code implementation28 Jun 2023 Ling Yang, Jiayi Zheng, Heyuan Wang, Zhongyi Liu, Zhilin Huang, Shenda Hong, Wentao Zhang, Bin Cui

To remove class spurious feature caused by distribution shifts, we propose Individual Graph Information Bottleneck (I-GIB) which discards irrelevant information by minimizing the mutual information between the input graph and its embeddings.

Graph Learning Out-of-Distribution Generalization

Diffusion-Based Scene Graph to Image Generation with Masked Contrastive Pre-Training

1 code implementation21 Nov 2022 Ling Yang, Zhilin Huang, Yang song, Shenda Hong, Guohao Li, Wentao Zhang, Bin Cui, Bernard Ghanem, Ming-Hsuan Yang

Generating images from graph-structured inputs, such as scene graphs, is uniquely challenging due to the difficulty of aligning nodes and connections in graphs with objects and their relations in images.

Image Generation

The Importance of Suppressing Complete Reconstruction in Autoencoders for Unsupervised Outlier Detection

no code implementations6 Nov 2022 Yafei Shen, Ling Yang

The core idea of our scheme is that in order to better detect high leverage points, we should suppress the complete reconstruction of the dataset to convert high leverage points into influential points, and it is also necessary to ensure that the differences between the eigenvalues of the covariance matrix of the original dataset and their corresponding reconstructed results in the direction of each principal component are equal.

Outlier Detection regression

Dimensionality Reduced Antenna Array for Beamforming/steering

no code implementations28 Oct 2022 Shiyi Xia, Mingyang Zhao, Qian Ma, Xunnan Zhang, Ling Yang, Yazhi Pi, Hyunchul Chung, Ad Reniers, A. M. J. Koonen, Zizheng Cao

Finally, the 16/8/4 -array beam steering was demonstrated by using 4/3/2 active controllers, respectively.

Diffusion Models: A Comprehensive Survey of Methods and Applications

2 code implementations2 Sep 2022 Ling Yang, Zhilong Zhang, Yang song, Shenda Hong, Runsheng Xu, Yue Zhao, Yingxia Shao, Wentao Zhang, Bin Cui, Ming-Hsuan Yang

This survey aims to provide a contextualized, in-depth look at the state of diffusion models, identifying the key areas of focus and pointing to potential areas for further exploration.

Image Super-Resolution Text-to-Image Generation +1

Omni-Granular Ego-Semantic Propagation for Self-Supervised Graph Representation Learning

no code implementations31 May 2022 Ling Yang, Shenda Hong

Unsupervised/self-supervised graph representation learning is critical for downstream node- and graph-level classification tasks.

Graph Representation Learning

Self-Supervised Time Series Representation Learning via Cross Reconstruction Transformer

1 code implementation20 May 2022 Wenrui Zhang, Ling Yang, Shijia Geng, Shenda Hong

In this paper, we aim at learning representations for time series from a new perspective and propose Cross Reconstruction Transformer (CRT) to solve the aforementioned problems in a unified way.

Contrastive Learning Representation Learning +2

Spatial Autoregressive Coding for Graph Neural Recommendation

no code implementations19 May 2022 Jiayi Zheng, Ling Yang, Heyuan Wang, Cheng Yang, Yinghong Li, Xiaowei Hu, Shenda Hong

To adequately leverage neighbor proximity and high-order information, we design a novel spatial autoregressive paradigm.

Graph Embedding

Spectral Propagation Graph Network for Few-shot Time Series Classification

no code implementations8 Feb 2022 Ling Yang, Shenda Hong, Luxia Zhang

To the best of our knowledge, SPGN is the first to utilize spectral comparisons in different intervals and involve spectral propagation across all time series with graph networks for few-shot TSC.

Classification Time Series +2

Unsupervised Time-Series Representation Learning with Iterative Bilinear Temporal-Spectral Fusion

no code implementations8 Feb 2022 Ling Yang, Shenda Hong

Unsupervised/self-supervised time series representation learning is a challenging problem because of its complex dynamics and sparse annotations.

Anomaly Detection Contrastive Learning +3

Iterative Bilinear Temporal-Spectral Fusion for Unsupervised Representation Learning in Time Series

no code implementations29 Sep 2021 Ling Yang, Shenda Hong, Luxia Zhang

First, we revisit the augmentation methods for time series of existing works and note that they mostly use segment-level augmentation derived from time slicing, which may bring about sampling bias and incorrect optimization with false negatives due to the loss of global context.

Anomaly Detection Contrastive Learning +4

DPGN: Distribution Propagation Graph Network for Few-shot Learning

1 code implementation CVPR 2020 Ling Yang, Liangliang Li, Zilun Zhang, Xinyu Zhou, Erjin Zhou, Yu Liu

To combine the distribution-level relations and instance-level relations for all examples, we construct a dual complete graph network which consists of a point graph and a distribution graph with each node standing for an example.

Few-Shot Learning Relation

DSTP-RNN: a dual-stage two-phase attention-based recurrent neural networks for long-term and multivariate time series prediction

no code implementations16 Apr 2019 Yeqi Liu, Chuanyang Gong, Ling Yang, Yingyi Chen

The key to solve this problem is to capture the spatial correlations at the same time, the spatio-temporal relationships at different times and the long-term dependence of the temporal relationships between different series.

Time Series Time Series Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.