no code implementations • 10 Dec 2023 • Luca Marzari, Gabriele Roncolato, Alessandro Farinelli
Deep Neural Networks (DNNs) are powerful tools that have shown extraordinary results in many scenarios, ranging from pattern recognition to complex robotic problems.
no code implementations • 18 Aug 2023 • Luca Marzari, Davide Corsi, Enrico Marchesini, Alessandro Farinelli, Ferdinando Cicalese
Identifying safe areas is a key point to guarantee trust for systems that are based on Deep Neural Networks (DNNs).
no code implementations • 20 Feb 2023 • Enrico Marchesini, Luca Marzari, Alessandro Farinelli, Christopher Amato
In this paper, we investigate an alternative approach that uses domain knowledge to quantify the risk in the proximity of such states by defining a violation metric.
no code implementations • 13 Feb 2023 • Luca Marzari, Enrico Marchesini, Alessandro Farinelli
Our evaluation compares the benefits of computing the number of violations using standard hard-coded properties and the ones generated with CROP.
no code implementations • 17 Jan 2023 • Luca Marzari, Davide Corsi, Ferdinando Cicalese, Alessandro Farinelli
Due to the #P-completeness of the problem, we also propose a randomized, approximate method that provides a provable probabilistic bound of the correct count while significantly reducing computational requirements.
no code implementations • 26 May 2022 • Guy Amir, Davide Corsi, Raz Yerushalmi, Luca Marzari, David Harel, Alessandro Farinelli, Guy Katz
Our work is the first to establish the usefulness of DNN verification in identifying and filtering out suboptimal DRL policies in real-world robots, and we believe that the methods presented here are applicable to a wide range of systems that incorporate deep-learning-based agents.
1 code implementation • 23 Dec 2021 • Luca Marzari, Davide Corsi, Enrico Marchesini, Alessandro Farinelli
To this end, we present a CL approach that leverages Transfer of Learning (ToL) and fine-tuning in a Unity-based simulation with the Robotnik Kairos as a robotic agent.