Search Results for author: Yuriy Brun

Found 13 papers, 6 papers with code

Rango: Adaptive Retrieval-Augmented Proving for Automated Software Verification

1 code implementation18 Dec 2024 Kyle Thompson, Nuno Saavedra, Pedro Carrott, Kevin Fisher, Alex Sanchez-Stern, Yuriy Brun, João F. Ferreira, Sorin Lerner, Emily First

We present Rango, a fully automated proof synthesis tool for Coq that automatically identifies relevant premises and also similar proofs from the current project and uses them during synthesis.

Retrieval

Cobblestone: Iterative Automation for Formal Verification

no code implementations25 Oct 2024 Saketh Ram Kasibatla, Arpan Agarwal, Yuriy Brun, Sorin Lerner, Talia Ringer, Emily First

We introduce Cobblestone, a new proof-synthesis approach that improves on the state of the art by taking advantage of partial progress in proof synthesis attempts.

Large Language Model

QEDCartographer: Automating Formal Verification Using Reward-Free Reinforcement Learning

no code implementations17 Aug 2024 Alex Sanchez-Stern, Abhishek Varghese, Zhanna Kaufman, Dylan Zhang, Talia Ringer, Yuriy Brun

To address this problem, we create QEDCartographer, an automated proof-synthesis tool that combines supervised and reinforcement learning to more effectively explore the proof space.

reinforcement-learning Reinforcement Learning

Thinking Forward: Memory-Efficient Federated Finetuning of Language Models

1 code implementation24 May 2024 Kunjal Panchal, Nisarg Parikh, Sunav Choudhary, Lijun Zhang, Yuriy Brun, Hui Guan

We also derive Spry's convergence rate, showing that the gradients decrease inversely proportional to the number of FL rounds, indicating the convergence up to the limits of heterogeneity.

Federated Learning

Attack-Resilient Image Watermarking Using Stable Diffusion

1 code implementation8 Jan 2024 Lijun Zhang, Xiao Liu, Antoni Viros Martin, Cindy Xiong Bearfield, Yuriy Brun, Hui Guan

To address this problem, we present ZoDiac, which uses a pre-trained stable diffusion model to inject a watermark into the trainable latent space, resulting in watermarks that can be reliably detected in the latent vector even when attacked.

Denoising

Baldur: Whole-Proof Generation and Repair with Large Language Models

no code implementations8 Mar 2023 Emily First, Markus N. Rabe, Talia Ringer, Yuriy Brun

Recent work has developed methods to automate formal verification using proof assistants, such as Coq and Isabelle/HOL, e. g., by training a model to predict one proof step at a time, and using that model to search through the space of possible proofs.

Enforcing Delayed-Impact Fairness Guarantees

no code implementations24 Aug 2022 Aline Weber, Blossom Metevier, Yuriy Brun, Philip S. Thomas, Bruno Castro da Silva

Recent research has shown that seemingly fair machine learning models, when used to inform decisions that have an impact on peoples' lives or well-being (e. g., applications involving education, employment, and lending), can inadvertently increase social inequality in the long term.

Fairness

Fairness Guarantees under Demographic Shift

no code implementations ICLR 2022 Stephen Giguere, Blossom Metevier, Yuriy Brun, Philip S. Thomas, Scott Niekum, Bruno Castro da Silva

Recent studies have demonstrated that using machine learning for social applications can lead to injustice in the form of racist, sexist, and otherwise unfair and discriminatory outcomes.

Fairness

Blindspots in Python and Java APIs Result in Vulnerable Code

no code implementations10 Mar 2021 Yuriy Brun, Tian Lin, Jessie Elise Somerville, Elisha Myers, Natalie C. Ebner

We find that using APIs with blindspots statistically significantly reduces the developers' ability to correctly reason about the APIs in both languages, but that the effect is more pronounced for Python.

Software Engineering Cryptography and Security

Fairkit, Fairkit, on the Wall, Who's the Fairest of Them All? Supporting Data Scientists in Training Fair Models

1 code implementation17 Dec 2020 Brittany Johnson, Jesse Bartola, Rico Angell, Katherine Keith, Sam Witty, Stephen J. Giguere, Yuriy Brun

To address bias in machine learning, data scientists need tools that help them understand the trade-offs between model quality and fairness in their specific data domains.

All BIG-bench Machine Learning +1

Fairness Testing: Testing Software for Discrimination

no code implementations11 Sep 2017 Sainyam Galhotra, Yuriy Brun, Alexandra Meliou

This paper defines software fairness and discrimination and develops a testing-based method for measuring if and how much software discriminates, focusing on causality in discriminatory behavior.

Fairness valid

Effectiveness of Anonymization in Double-Blind Review

1 code implementation5 Sep 2017 Claire Le Goues, Yuriy Brun, Sven Apel, Emery Berger, Sarfraz Khurshid, Yannis Smaragdakis

Double-blind review relies on the authors' ability and willingness to effectively anonymize their submissions.

Digital Libraries General Literature Software Engineering

Cannot find the paper you are looking for? You can Submit a new open access paper.