Search Results for author: Jie M. Zhang

Found 11 papers, 3 papers with code

LLM-Powered Test Case Generation for Detecting Tricky Bugs

no code implementations16 Apr 2024 Kaibo Liu, Yiyang Liu, Zhenpeng Chen, Jie M. Zhang, Yudong Han, Yun Ma, Ge Li, Gang Huang

Conventional automated test generation tools struggle to generate test oracles and tricky bug-revealing test inputs.

EffiBench: Benchmarking the Efficiency of Automatically Generated Code

no code implementations3 Feb 2024 Dong Huang, Jie M. Zhang, Yuhao QING, Heming Cui

This paper presents EffiBench, a benchmark with 1, 000 efficiency-critical coding problems for assessing the efficiency of code generated by code generation models.

Benchmarking Code Completion +1

AgentCoder: Multi-Agent-based Code Generation with Iterative Testing and Optimisation

no code implementations20 Dec 2023 Dong Huang, Qingwen Bu, Jie M. Zhang, Michael Luck, Heming Cui

The advancement of natural language processing (NLP) has been significantly boosted by the development of transformer-based large language models (LLMs).

Code Generation Prompt Engineering

ConDefects: A New Dataset to Address the Data Leakage Concern for LLM-based Fault Localization and Program Repair

no code implementations25 Oct 2023 Yonghao Wu, Zheng Li, Jie M. Zhang, Yong liu

With the growing interest on Large Language Models (LLMs) for fault localization and program repair, ensuring the integrity and generalizability of the LLM-based methods becomes paramount.

Benchmarking Fault localization

Bias Behind the Wheel: Fairness Analysis of Autonomous Driving Systems

no code implementations5 Aug 2023 Xinyue Li, Zhenpeng Chen, Jie M. Zhang, Federica Sarro, Ying Zhang, Xuanzhe Liu

This paper analyzes fairness in automated pedestrian detection, a crucial but under-explored issue in autonomous driving systems.

Autonomous Driving Fairness +1

Fairness Improvement with Multiple Protected Attributes: How Far Are We?

1 code implementation25 Jul 2023 Zhenpeng Chen, Jie M. Zhang, Federica Sarro, Mark Harman

Existing research mostly improves the fairness of Machine Learning (ML) software regarding a single protected attribute at a time, but this is unrealistic given that many users have multiple protected attributes.

Attribute Fairness

A Comprehensive Empirical Study of Bias Mitigation Methods for Machine Learning Classifiers

2 code implementations7 Jul 2022 Zhenpeng Chen, Jie M. Zhang, Federica Sarro, Mark Harman

We find that (1) the bias mitigation methods significantly decrease ML performance in 53% of the studied scenarios (ranging between 42%~66% according to different ML performance metrics); (2) the bias mitigation methods significantly improve fairness measured by the 4 used metrics in 46% of all the scenarios (ranging between 24%~59% according to different fairness metrics); (3) the bias mitigation methods even lead to decrease in both fairness and ML performance in 25% of the scenarios; (4) the effectiveness of the bias mitigation methods depends on tasks, models, the choice of protected attributes, and the set of metrics used to assess fairness and ML performance; (5) there is no bias mitigation method that can achieve the best trade-off in all the scenarios.

Fairness

Model Validation Using Mutated Training Labels: An Exploratory Study

no code implementations24 May 2019 Jie M. Zhang, Mark Harman, Benjamin Guedj, Earl T. Barr, John Shawe-Taylor

MV mutates training data labels, retrains the model against the mutated data, then uses the metamorphic relation that captures the consequent training performance changes to assess model fit.

BIG-bench Machine Learning General Classification +1

Cannot find the paper you are looking for? You can Submit a new open access paper.