Search Results for author: Binhua Li

Found 27 papers, 17 papers with code

Thinking Longer, Not Larger: Enhancing Software Engineering Agents via Scaling Test-Time Compute

1 code implementation31 Mar 2025 Yingwei Ma, Yongbin Li, Yihong Dong, Xue Jiang, Rongyu Cao, Jue Chen, Fei Huang, Binhua Li

Recent advancements in software engineering agents have demonstrated promising capabilities in automating program improvements.

Fault localization

Do Code LLMs Understand Design Patterns?

no code implementations8 Jan 2025 Zhenyu Pan, Xuefeng Song, Yunkun Wang, Rongyu Cao, Binhua Li, Yongbin Li, Han Liu

Code Large Language Models (LLMs) demonstrate great versatility in adapting to various downstream tasks, including code generation and completion, as well as bug detection and fixing.

Code Generation

LLMs as Continuous Learners: Improving the Reproduction of Defective Code in Software Issues

no code implementations21 Nov 2024 Yalan Lin, Yingwei Ma, Rongyu Cao, Binhua Li, Fei Huang, Xiaodong Gu, Yongbin Li

Reproducing buggy code is the first and crucially important step in issue resolving, as it aids in identifying the underlying problems and validating that generated patches resolve the problem.

Lingma SWE-GPT: An Open Development-Process-Centric Language Model for Automated Software Improvement

1 code implementation1 Nov 2024 Yingwei Ma, Rongyu Cao, Yongchang Cao, Yue Zhang, Jue Chen, Yibo Liu, Yuchen Liu, Binhua Li, Fei Huang, Yongbin Li

The results demonstrate that Lingma SWE-GPT 72B successfully resolves 30. 20% of the GitHub issues, marking a significant improvement in automatic issue resolution (22. 76% relative improvement compared to Llama 3. 1 405B), approaching the performance of closed-source models (31. 80\% issues of GPT-4o resolved).

Language Modeling Language Modelling

EvoCodeBench: An Evolving Code Generation Benchmark with Domain-Specific Evaluations

no code implementations30 Oct 2024 Jia Li, Ge Li, Xuanming Zhang, YunFei Zhao, Yihong Dong, Zhi Jin, Binhua Li, Fei Huang, Yongbin Li

These evaluations help practitioners select superior LLMs in specific domains and discover the shortcomings of existing LLMs.

Code Generation Fairness

Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?

1 code implementation2 Oct 2024 Zhenyu Pan, Rongyu Cao, Yongchang Cao, Yingwei Ma, Binhua Li, Fei Huang, Han Liu, Yongbin Li

Code completion, a key downstream task in code generation, is one of the most frequent and impactful methods for enhancing developer productivity in software development.

Code Completion Code Generation

In-Context Transfer Learning: Demonstration Synthesis by Transferring Similar Tasks

1 code implementation2 Oct 2024 Dingzirui Wang, Xuanliang Zhang, Qiguang Chen, Longxu Dou, Xiao Xu, Rongyu Cao, Yingwei Ma, Qingfu Zhu, Wanxiang Che, Binhua Li, Fei Huang, Yongbin Li

To address this, inspired by transfer learning, we propose In-Context Transfer Learning (ICTL), which synthesizes target task demonstrations by transferring labeled demonstrations from similar source tasks.

In-Context Learning Transfer Learning

Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining

1 code implementation1 Oct 2024 Jie Cheng, Ruixi Qiao, Gang Xiong, Qinghai Miao, Yingwei Ma, Binhua Li, Yongbin Li, Yisheng Lv

Experimental results indicate that our largest agent, with 150 million parameters, achieves 78. 9% human-level performance on pretrained games using only 10% subsampled offline data, outperforming existing state-of-the-art large-scale offline RL baselines by 31. 6% on averange.

Atari Games model +3

Alibaba LingmaAgent: Improving Automated Issue Resolution via Comprehensive Repository Exploration

no code implementations3 Jun 2024 Yingwei Ma, Qingping Yang, Rongyu Cao, Binhua Li, Fei Huang, Yongbin Li

This paper presents Alibaba LingmaAgent, a novel Automated Software Engineering method designed to comprehensively understand and utilize whole software repositories for issue resolution.

Language Modelling Large Language Model

Unifying Structured Data as Graph for Data-to-Text Pre-Training

1 code implementation2 Jan 2024 Shujie Li, Liang Li, Ruiying Geng, Min Yang, Binhua Li, Guanghu Yuan, Wanwei He, Shao Yuan, Can Ma, Fei Huang, Yongbin Li

In this paper, we unify different types of structured data (i. e., table, key-value data, knowledge graph) into the graph format and cast different data-to-text generation tasks as graph-to-text generation.

Data-to-Text Generation

CATS: A Pragmatic Chinese Answer-to-Sequence Dataset with Large Scale and High Quality

no code implementations20 Jun 2023 Liang Li, Ruiying Geng, Chengyang Fang, Bing Li, Can Ma, Rongyu Cao, Binhua Li, Fei Huang, Yongbin Li

To alleviate these limitations, in this paper, we present CATS, a pragmatic Chinese answer-to-sequence dataset with large scale and high quality.

Iterative Forward Tuning Boosts In-Context Learning in Language Models

1 code implementation22 May 2023 Jiaxi Yang, Binyuan Hui, Min Yang, Bailin Wang, Bowen Li, Binhua Li, Fei Huang, Yongbin Li

Despite the advancements in in-context learning (ICL) for large language models (LLMs), current research centers on specific prompt engineering, such as demonstration selection, with the expectation that a single iteration of demonstrations processing can generalize effectively to a given test sample.

Decision Making In-Context Learning +2

Can LLM Already Serve as A Database Interface? A BIg Bench for Large-Scale Database Grounded Text-to-SQLs

1 code implementation NeurIPS 2023 Jinyang Li, Binyuan Hui, Ge Qu, Jiaxi Yang, Binhua Li, Bowen Li, Bailin Wang, Bowen Qin, Rongyu Cao, Ruiying Geng, Nan Huo, Xuanhe Zhou, Chenhao Ma, Guoliang Li, Kevin C. C. Chang, Fei Huang, Reynold Cheng, Yongbin Li

Our emphasis on database values highlights the new challenges of dirty database contents, external knowledge between NL questions and database contents, and SQL efficiency, particularly in the context of massive databases.

SQL Parsing Text to SQL +1

Plan-then-Seam: Towards Efficient Table-to-Text Generation

1 code implementation10 Feb 2023 Liang Li, Ruiying Geng, Chengyang Fang, Bing Li, Can Ma, Binhua Li, Yongbin Li

Table-to-text generation aims at automatically generating text to help people conveniently obtain salient information in tables.

Table-to-Text Generation

Large Language Models are Versatile Decomposers: Decompose Evidence and Questions for Table-based Reasoning

2 code implementations31 Jan 2023 Yunhu Ye, Binyuan Hui, Min Yang, Binhua Li, Fei Huang, Yongbin Li

To alleviate the above challenges, we exploit large language models (LLMs) as decomposers for effective table-based reasoning, which (i) decompose huge evidence (a huge table) into sub-evidence (a small table) to mitigate the interference of useless information for table reasoning; and (ii) decompose complex questions into simpler sub-questions for text reasoning.

Hallucination Semantic Parsing +1

Towards Generalizable and Robust Text-to-SQL Parsing

1 code implementation23 Oct 2022 Chang Gao, Bowen Li, Wenxuan Zhang, Wai Lam, Binhua Li, Fei Huang, Luo Si, Yongbin Li

Text-to-SQL parsing tackles the problem of mapping natural language questions to executable SQL queries.

SQL Parsing Text to SQL +1

STAR: SQL Guided Pre-Training for Context-dependent Text-to-SQL Parsing

1 code implementation21 Oct 2022 ZeFeng Cai, Xiangyu Li, Binyuan Hui, Min Yang, Bowen Li, Binhua Li, Zheng Cao, Weijie Li, Fei Huang, Luo Si, Yongbin Li

Concretely, we propose two novel pre-training objectives which respectively explore the context-dependent interactions of NL utterances and SQL queries within each text-to-SQL conversation: (i) schema state tracking (SST) objective that tracks and explores the schema states of context-dependent SQL queries in the form of schema-states by predicting and updating the value of each schema slot during interaction; (ii) utterance dependency tracking (UDT) objective that employs weighted contrastive learning to pull together two semantically similar NL utterances and push away the representations of semantically dissimilar NL utterances within each conversation.

Contrastive Learning SQL Parsing +2

SUN: Exploring Intrinsic Uncertainties in Text-to-SQL Parsers

1 code implementation COLING 2022 Bowen Qin, Lihan Wang, Binyuan Hui, Bowen Li, Xiangpeng Wei, Binhua Li, Fei Huang, Luo Si, Min Yang, Yongbin Li

To improve the generalizability and stability of neural text-to-SQL parsers, we propose a model uncertainty constraint to refine the query representations by enforcing the output representations of different perturbed encoding networks to be consistent with each other.

SQL Parsing Text to SQL +1

A Survey on Text-to-SQL Parsing: Concepts, Methods, and Future Directions

no code implementations29 Aug 2022 Bowen Qin, Binyuan Hui, Lihan Wang, Min Yang, Jinyang Li, Binhua Li, Ruiying Geng, Rongyu Cao, Jian Sun, Luo Si, Fei Huang, Yongbin Li

In recent years, deep neural networks have significantly advanced this task by neural generation models, which automatically learn a mapping function from an input NL question to an output SQL query.

SQL Parsing Text to SQL +1

Proton: Probing Schema Linking Information from Pre-trained Language Models for Text-to-SQL Parsing

2 code implementations28 Jun 2022 Lihan Wang, Bowen Qin, Binyuan Hui, Bowen Li, Min Yang, Bailin Wang, Binhua Li, Fei Huang, Luo Si, Yongbin Li

The importance of building text-to-SQL parsers which can be applied to new databases has long been acknowledged, and a critical step to achieve this goal is schema linking, i. e., properly recognizing mentions of unseen columns or tables when generating SQLs.

SQL Parsing Text to SQL +1

Improving Text-to-SQL with Schema Dependency Learning

no code implementations7 Mar 2021 Binyuan Hui, Xiang Shi, Ruiying Geng, Binhua Li, Yongbin Li, Jian Sun, Xiaodan Zhu

In this paper, we present the Schema Dependency guided multi-task Text-to-SQL model (SDSQL) to guide the network to effectively capture the interactions between questions and schemas.

Text to SQL Text-To-SQL

Cannot find the paper you are looking for? You can Submit a new open access paper.