Search Results for author: Yufan Huang

Found 8 papers, 2 papers with code

Is Next Token Prediction Sufficient for GPT? Exploration on Code Logic Comprehension

no code implementations13 Apr 2024 MengNan Qi, Yufan Huang, Yongqiang Yao, Maoquan Wang, Bin Gu, Neel Sundaresan

Our experimental results reveal that following this pretraining, both Code Llama and StarCoder, the prevalent code domain pretraining models, display significant improvements on our logically equivalent code selection task and the code completion task.

Code Completion Sentence +2

Rethinking the Instruction Quality: LIFT is What You Need

no code implementations12 Dec 2023 Yang Xu, Yongqiang Yao, Yufan Huang, MengNan Qi, Maoquan Wang, Bin Gu, Neel Sundaresan

Instruction tuning, a specialized technique to enhance large language model (LLM) performance via instruction datasets, relies heavily on the quality of employed data.

Code Generation Instruction Following +3

SUT: Active Defects Probing for Transcompiler Models

no code implementations22 Oct 2023 MengNan Qi, Yufan Huang, Maoquan Wang, Yongqiang Yao, Zihan Liu, Bin Gu, Colin Clement, Neel Sundaresan

In this paper we introduce a new metrics for programming language translation and these metrics address these basic syntax errors.

Translation

Program Translation via Code Distillation

no code implementations17 Oct 2023 Yufan Huang, MengNan Qi, Yongqiang Yao, Maoquan Wang, Bin Gu, Colin Clement, Neel Sundaresan

Distilled code serves as a translation pivot for any programming language, leading by construction to parallel corpora which scale to all available source code by simply applying the distillation compiler.

Machine Translation Translation

A flexible PageRank-based graph embedding framework closely related to spectral eigenvector embeddings

1 code implementation22 Jul 2022 Disha Shur, Yufan Huang, David F. Gleich

We study a simple embedding technique based on a matrix of personalized PageRank vectors seeded on a random set of nodes.

Graph Embedding

Stochastic-Sign SGD for Federated Learning with Theoretical Guarantees

no code implementations25 Feb 2020 Richeng Jin, Yufan Huang, Xiaofan He, Huaiyu Dai, Tianfu Wu

We present Stochastic-Sign SGD which utilizes novel stochastic-sign based gradient compressors enabling the aforementioned properties in a unified framework.

Federated Learning Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.