Search Results for author: Linyuan Gong

Found 10 papers, 7 papers with code

MC-BERT: Efficient Language Pre-Training via a Meta Controller

1 code implementation10 Jun 2020 Zhenhui Xu, Linyuan Gong, Guolin Ke, Di He, Shuxin Zheng, Li-Wei Wang, Jiang Bian, Tie-Yan Liu

Pre-trained contextual representations (e. g., BERT) have become the foundation to achieve state-of-the-art results on many NLP tasks.

Binary Classification Cloze Test +4

Joint Language Semantic and Structure Embedding for Knowledge Graph Completion

1 code implementation COLING 2022 Jianhao Shen, Chenguang Wang, Linyuan Gong, Dawn Song

Unlike previous approaches that rely on either the structures or semantics of the knowledge graphs, we propose to jointly embed the semantics in the natural language description of the knowledge triplets with their structure information.

Link Prediction

AST-T5: Structure-Aware Pretraining for Code Generation and Understanding

1 code implementation5 Jan 2024 Linyuan Gong, Mostafa Elhoushi, Alvin Cheung

Large language models (LLMs) have made significant advancements in code-related tasks, yet many LLMs treat code as simple sequences, neglecting its structured nature.

Code Generation

Anytime Sampling for Autoregressive Models via Ordered Autoencoding

1 code implementation ICLR 2021 Yilun Xu, Yang song, Sahaj Garg, Linyuan Gong, Rui Shu, Aditya Grover, Stefano Ermon

Experimentally, we demonstrate in several image and audio generation tasks that sample quality degrades gracefully as we reduce the computational budget for sampling.

Audio Generation Computational Efficiency

Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers

1 code implementation21 May 2023 Linyuan Gong, Chenyan Xiong, Xiaodong Liu, Payal Bajaj, Yiqing Xie, Alvin Cheung, Jianfeng Gao, Xia Song

This paper explores the effectiveness of model-generated signals in improving zero-shot generalization of text-to-text Transformers such as T5.

Zero-shot Generalization

Evaluation of LLMs on Syntax-Aware Code Fill-in-the-Middle Tasks

1 code implementation7 Mar 2024 Linyuan Gong, Sida Wang, Mostafa Elhoushi, Alvin Cheung

We introduce Syntax-Aware Fill-In-the-Middle (SAFIM), a new benchmark for evaluating Large Language Models (LLMs) on the code Fill-in-the-Middle (FIM) task.

Code Completion

What Does a TextCNN Learn?

no code implementations19 Jan 2018 Linyuan Gong, Ruyi Ji

TextCNN, the convolutional neural network for text, is a useful deep learning algorithm for sentence classification tasks such as sentiment analysis and question classification.

Classification General Classification +4

ADELT: Transpilation Between Deep Learning Frameworks

no code implementations7 Mar 2023 Linyuan Gong, Jiayi Wang, Alvin Cheung

We propose the Adversarial DEep Learning Transpiler (ADELT), a novel approach to source-to-source transpilation between deep learning frameworks.

Cannot find the paper you are looking for? You can Submit a new open access paper.