no code implementations • 15 Apr 2024 • Mohammed Latif Siddiq, Simantika Dristi, Joy Saha, Joanna C. S. Santos
We found that code generation evaluation benchmarks mainly focused on Python and coding exercises and had very limited contextual dependencies to challenge the model.
no code implementations • 1 Nov 2023 • Mohammed Latif Siddiq, Joanna C. S. Santos
This framework has three major components: a novel dataset of security-centric Python prompts, an evaluation environment to test the generated code, and novel metrics to evaluate the models' performance from the perspective of secure code generation.
no code implementations • 17 Jul 2023 • Mohammed Latif Siddiq, Beatrice Casey, Joanna C. S. Santos
FRANC includes a static filter to make the generated code compilable with heuristics and a quality-aware ranker to sort the code snippets based on a quality score.
1 code implementation • 30 Apr 2023 • Mohammed Latif Siddiq, Joanna C. S. Santos, Ridwanul Hasan Tanvir, Noshin Ulfat, Fahmid Al Rifat, Vinicius Carvalho Lopes
A code generation model generates code by taking a prompt from a code comment, existing code, or a combination of both.
no code implementations • 16 Apr 2023 • Rishov Paul, Md. Mohib Hossain, Mohammed Latif Siddiq, Masum Hasan, Anindya Iqbal, Joanna C. S. Santos
We applied PLBART and CodeT5, two state-of-the-art language models that are pre-trained with both PL and NL, on two such natural language-based program repair datasets and found that the pre-trained language models fine-tuned with datasets containing both code review and subsequent code changes notably outperformed each of the previous models.
1 code implementation • 1st International Workshop on Mining Software Repositories Applications for Privacy and Security (MSR4P&S ’22) 2022 • Mohammed Latif Siddiq, Joanna C.S. Santos
For this reason, evaluating the security of a code generation model is a must.