no code implementations • 7 Sep 2021 • Greg Fields, Mohammad Samragh, Mojan Javaheripi, Farinaz Koushanfar, Tara Javidi
Deep neural networks have been shown to be vulnerable to backdoor, or trojan, attacks where an adversary has embedded a trigger in the network at training time such that the model correctly classifies all standard inputs, but generates a targeted, incorrect classification on any input which contains the trigger.
no code implementations • NeurIPS 2021 • Shubhanshu Shekhar, Greg Fields, Mohammad Ghavamzadeh, Tara Javidi
Machine learning models trained on uncurated datasets can often end up adversely affecting inputs belonging to underrepresented groups.