Search Results for author: Lin Zheng

Found 22 papers, 13 papers with code

Scaling Diffusion Language Models via Adaptation from Autoregressive Models

1 code implementation23 Oct 2024 Shansan Gong, Shivam Agarwal, Yizhe Zhang, Jiacheng Ye, Lin Zheng, Mukai Li, Chenxin An, Peilin Zhao, Wei Bi, Jiawei Han, Hao Peng, Lingpeng Kong

Diffusion Language Models (DLMs) have emerged as a promising new paradigm for text generative modeling, potentially addressing limitations of autoregressive (AR) models.

In-Context Learning Language Modeling +1

Beyond Autoregression: Discrete Diffusion for Complex Reasoning and Planning

1 code implementation18 Oct 2024 Jiacheng Ye, Jiahui Gao, Shansan Gong, Lin Zheng, Xin Jiang, Zhenguo Li, Lingpeng Kong

Our work highlights the potential of diffusion-based approaches in advancing AI capabilities for sophisticated language understanding and problem-solving tasks.

SubgoalXL: Subgoal-based Expert Learning for Theorem Proving

1 code implementation20 Aug 2024 Xueliang Zhao, Lin Zheng, Haige Bo, Changran Hu, Urmish Thakker, Lingpeng Kong

This paper introduces SubgoalXL, a novel approach that synergizes subgoal-based proofs with expert learning to enhance LLMs' capabilities in formal theorem proving within the Isabelle environment.

Ranked #3 on Automated Theorem Proving on miniF2F-test (using extra training data)

Automated Theorem Proving

Diffusion of Thoughts: Chain-of-Thought Reasoning in Diffusion Language Models

1 code implementation12 Feb 2024 Jiacheng Ye, Shansan Gong, Liheng Chen, Lin Zheng, Jiahui Gao, Han Shi, Chuan Wu, Xin Jiang, Zhenguo Li, Wei Bi, Lingpeng Kong

Recently, diffusion models have garnered significant interest in the field of text processing due to their many potential advantages compared to conventional autoregressive models.

Language Modeling Language Modelling +1

Linear Attention via Orthogonal Memory

no code implementations18 Dec 2023 Jun Zhang, Shuyang Jiang, Jiangtao Feng, Lin Zheng, Lingpeng Kong

Given that orthogonal memory compresses global information, we further dissect the context to amplify fine-grained local information.

Causal Language Modeling Computational Efficiency +2

Automated interpretation of congenital heart disease from multi-view echocardiograms

no code implementations30 Nov 2023 Jing Wang, Xiaofeng Liu, Fangyun Wang, Lin Zheng, Fengqiao Gao, Hanwen Zhang, Xin Zhang, Wanqing Xie, Binbin Wang

Our video-based model can diagnose with an accuracy of 93. 9\% (binary classification), and 92. 1\% (3-class classification) in a collected 2D video testing set, which does not need key-frame selection and view annotation in testing.

Binary Classification

Self-Infilling Code Generation

1 code implementation29 Nov 2023 Lin Zheng, Jianbo Yuan, Zhi Zhang, Hongxia Yang, Lingpeng Kong

This work introduces self-infilling code generation, a general framework that incorporates infilling operations into auto-regressive decoding.

Code Generation

Attentive Multi-Layer Perceptron for Non-autoregressive Generation

1 code implementation14 Oct 2023 Shuyang Jiang, Jun Zhang, Jiangtao Feng, Lin Zheng, Lingpeng Kong

Furthermore, we marry AMLP with popular NAR models, deriving a highly efficient NAR-AMLP architecture with linear time and space complexity.

Machine Translation Speech Synthesis +2

K-means Clustering Based Feature Consistency Alignment for Label-free Model Evaluation

no code implementations17 Apr 2023 Shuyu Miao, Lin Zheng, Jingjing Liu, and Hong Jin

The label-free model evaluation aims to predict the model performance on various test sets without relying on ground truths.

Clustering

Retrieved Sequence Augmentation for Protein Representation Learning

1 code implementation24 Feb 2023 Chang Ma, Haiteng Zhao, Lin Zheng, Jiayi Xin, Qintong Li, Lijun Wu, Zhihong Deng, Yang Lu, Qi Liu, Lingpeng Kong

RSA links query protein sequences to a set of sequences with similar structures or properties in the database and combines these sequences for downstream prediction.

Prediction Property Prediction +2

A Reparameterized Discrete Diffusion Model for Text Generation

1 code implementation11 Feb 2023 Lin Zheng, Jianbo Yuan, Lei Yu, Lingpeng Kong

This work studies discrete diffusion probabilistic models with applications to natural language generation.

model Text Generation

Efficient Attention via Control Variates

1 code implementation9 Feb 2023 Lin Zheng, Jianbo Yuan, Chong Wang, Lingpeng Kong

Built upon previous progress of RFA, we characterize this gap through the lens of control variates and show that RFA can be decomposed into a sum of multiple control variate estimators for each element in the sequence.

CAB: Comprehensive Attention Benchmarking on Long Sequence Modeling

1 code implementation14 Oct 2022 Jun Zhang, Shuyang Jiang, Jiangtao Feng, Lin Zheng, Lingpeng Kong

In this paper, we propose Comprehensive Attention Benchmark (CAB) under a fine-grained attention taxonomy with four distinguishable attention patterns, namely, noncausal self, causal self, noncausal cross, and causal cross attentions.

Benchmarking Language Modeling +1

Learning Feature Disentanglement and Dynamic Fusion for Recaptured Image Forensic

no code implementations13 Jun 2022 Shuyu Miao, Lin Zheng, Hong Jin

Image recapture seriously breaks the fairness of artificial intelligent (AI) systems, which deceives the system by recapturing others' images.

Disentanglement Fairness

Poincaré Heterogeneous Graph Neural Networks for Sequential Recommendation

no code implementations16 May 2022 Naicheng Guo, Xiaolei Liu, Shaoshuai Li, Qiongxu Ma, Kaixin Gao, Bing Han, Lin Zheng, Xiaobo Guo

In this paper, we propose a Poincar\'{e}-based heterogeneous graph neural network named PHGR to model the sequential pattern information as well as hierarchical information contained in the data of SR scenarios simultaneously.

Graph Neural Network Graph Representation Learning +1

Linear Complexity Randomized Self-attention Mechanism

1 code implementation10 Apr 2022 Lin Zheng, Chong Wang, Lingpeng Kong

By combining the expressiveness in RA and the efficiency in RFA, we develop a novel linear complexity self-attention mechanism called linear randomized attention (LARA).

Ripple Attention for Visual Perception with Sub-quadratic Complexity

no code implementations6 Oct 2021 Lin Zheng, Huijie Pan, Lingpeng Kong

Transformer architectures are now central to sequence modeling tasks.

Cascaded Head-colliding Attention

1 code implementation ACL 2021 Lin Zheng, Zhiyong Wu, Lingpeng Kong

Transformers have advanced the field of natural language processing (NLP) on a variety of important tasks.

Language Modeling Language Modelling +2

Generative Semantic Hashing Enhanced via Boltzmann Machines

no code implementations ACL 2020 Lin Zheng, Qinliang Su, Dinghan Shen, Changyou Chen

Generative semantic hashing is a promising technique for large-scale information retrieval thanks to its fast retrieval speed and small memory footprint.

Information Retrieval Retrieval

Cannot find the paper you are looking for? You can Submit a new open access paper.