no code implementations • 31 Jul 2024 • Gandalf Nicolas, Aylin Caliskan
Finally, the taxonomy predicted the LLMs' internal evaluations of social categories (e. g., how positively/negatively the categories were represented), supporting the relevance of a multidimensional taxonomy for characterizing LLM stereotypes.
no code implementations • 29 Jul 2024 • Kyra Wilson, Aylin Caliskan
We find that the MTEs are biased, significantly favoring White-associated names in 85. 1\% of cases and female-associated names in only 11. 1\% of cases, with a minority of cases showing no statistically significant differences.
no code implementations • 20 Jul 2024 • Sourojit Ghosh, Pranav Narayanan Venkit, Sanjana Gautam, Shomir Wilson, Aylin Caliskan
Our research investigates the impact of Generative Artificial Intelligence (GAI) models, specifically text-to-image generators (T2Is), on the representation of non-Western cultures, with a focus on Indian contexts.
no code implementations • 2 Jul 2024 • Chahat Raj, Anjishnu Mukherjee, Aylin Caliskan, Antonios Anastasopoulos, Ziwei Zhu
We propose a unique debiasing technique, Social Contact Debiasing (SCD), that instruction-tunes these models with unbiased responses to prompts.
1 code implementation • 2 Jul 2024 • Chahat Raj, Anjishnu Mukherjee, Aylin Caliskan, Antonios Anastasopoulos, Ziwei Zhu
Existing works examining Vision-Language Models (VLMs) for social biases predominantly focus on a limited set of documented bias associations, such as gender:profession or race:crime.
no code implementations • 20 Jun 2024 • Steven A. Lehr, Aylin Caliskan, Suneragiri Liyanage, Mahzarin R. Banaji
How good a research scientist is ChatGPT?
no code implementations • CVPR 2024 • Yiwei Yang, Anthony Z. Liu, Robert Wolfe, Aylin Caliskan, Bill Howe
We propose Concept Correction a framework that represents a concept as a curated set of images from any source then labels each training sample by its similarity to the concept set to control spurious correlations.
no code implementations • 30 Oct 2023 • Sourojit Ghosh, Aylin Caliskan
We observe how Stable Diffusion outputs of `a person' without any additional gender/nationality information correspond closest to images of men and least with persons of nonbinary gender, and to persons from Europe/North America over Africa/Asia, pointing towards Stable Diffusion having a concerning representation of personhood to be a European/North American man.
1 code implementation • 29 Oct 2023 • Isaac Slaughter, Craig Greenberg, Reva Schwartz, Aylin Caliskan
We compare biases found in pre-trained models to biases in downstream models adapted to the task of Speech Emotion Recognition (SER) and find that in 66 of the 96 tests performed (69%), the group that is more associated with positive valence as indicated by the SpEAT also tends to be predicted as speaking with higher valence by the downstream model.
no code implementations • 30 Aug 2023 • Inyoung Cheong, Aylin Caliskan, Tadayoshi Kohno
Our interdisciplinary study investigates how effectively U. S. laws confront the challenges posed by Generative AI to human values.
1 code implementation • 7 Jul 2023 • Shiva Omrani Sabbaghi, Robert Wolfe, Aylin Caliskan
Adapting the projection-based approach to embedding association tests that quantify bias, we find that language models exhibit the most biased attitudes against gender identity, social class, and sexual orientation signals in language.
1 code implementation • 8 Jun 2023 • Katelyn X. Mei, Sonia Fereidooni, Aylin Caliskan
We investigate bias against these groups in English pre-trained Masked Language Models (MLMs) and their downstream sentiment classification tasks.
no code implementations • 17 May 2023 • Sourojit Ghosh, Aylin Caliskan
In this multicultural age, language translation is one of the most performed tasks, and it is becoming increasingly AI-moderated and automated.
1 code implementation • 21 Dec 2022 • Robert Wolfe, Yiwei Yang, Bill Howe, Aylin Caliskan
A first experiment uses standardized images of women from the Sexual OBjectification and EMotion Database, and finds that human characteristics are disassociated from images of objectified women: the model's recognition of emotional state is mediated by whether the subject is fully or partially clothed.
1 code implementation • 7 Nov 2022 • Federico Bianchi, Pratyusha Kalluri, Esin Durmus, Faisal Ladhak, Myra Cheng, Debora Nozza, Tatsunori Hashimoto, Dan Jurafsky, James Zou, Aylin Caliskan
For example, we find cases of prompting for basic traits or social roles resulting in images reinforcing whiteness as ideal, prompting for occupations resulting in amplification of racial and gender disparities, and prompting for objects resulting in reification of American norms.
no code implementations • 1 Jul 2022 • Robert Wolfe, Aylin Caliskan
In an image captioning task, BLIP remarks upon the race of Asian individuals as much as 36% of the time, but never remarks upon race for White individuals.
no code implementations • 7 Jun 2022 • Aylin Caliskan, Pimparkar Parth Ajay, Tessa Charlesworth, Robert Wolfe, Mahzarin R. Banaji
Using the Single-Category Word Embedding Association Test, we demonstrate the widespread prevalence of gender biases that also show differences in: (1) frequencies of words associated with men versus women; (b) part-of-speech tags in gender-associated words; (c) semantic categories in gender-associated words; and (d) valence, arousal, and dominance in gender-associated words.
1 code implementation • 3 Jun 2022 • Shiva Omrani Sabbaghi, Aylin Caliskan
We demonstrate that word embeddings learn the association between a noun and its grammatical gender in grammatically gendered languages, which can skew social gender bias measurements.
1 code implementation • 23 May 2022 • Robert Wolfe, Aylin Caliskan
The model is more likely to rank the unmarked "person" label higher than labels denoting gender for Male individuals (26. 7% of the time) vs.
1 code implementation • 22 May 2022 • Robert Wolfe, Mahzarin R. Banaji, Aylin Caliskan
We examine the state-of-the-art multimodal "visual semantic" model CLIP ("Contrastive Language Image Pretraining") for the rule of hypodescent, or one-drop rule, whereby multiracial people are more likely to be assigned a racial or ethnic label corresponding to a minority or disadvantaged racial or ethnic group than to the equivalent majority or advantaged group.
no code implementations • ACL 2022 • Robert Wolfe, Aylin Caliskan
We find that contrastive visual semantic pretraining significantly mitigates the anisotropy found in contextualized word embeddings from GPT-2, such that the intra-layer self-similarity (mean pairwise cosine similarity) of CLIP word embeddings is under . 25 in all layers, compared to greater than . 95 in the top layer of GPT-2.
1 code implementation • 14 Mar 2022 • Robert Wolfe, Aylin Caliskan
VAST, the Valence-Assessing Semantics Test, is a novel intrinsic evaluation task for contextualized word embeddings (CWEs).
no code implementations • EMNLP 2021 • Robert Wolfe, Aylin Caliskan
Moreover, we find Spearman's r between racial bias and name frequency in BERT of . 492, indicating that lower-frequency minority group names are more associated with unpleasantness.
1 code implementation • 28 Oct 2020 • Ryan Steed, Aylin Caliskan
Recent advances in machine learning leverage massive datasets of unlabeled images from the web to learn general-purpose image representations for tasks from image classification to face recognition.
no code implementations • 8 Jun 2020 • Akshat Pandey, Aylin Caliskan
An analysis of 100 million ridehailing samples from the city of Chicago indicates a significant disparate impact in fare pricing of neighborhoods due to AI bias learned from ridehailing utilization patterns associated with demographic attributes.
2 code implementations • EMNLP 2021 • Autumn Toney-Wails, Aylin Caliskan
Word embeddings learn implicit biases from linguistic regularities captured by word co-occurrence statistics.
2 code implementations • 6 Jun 2020 • Wei Guo, Aylin Caliskan
Furthermore, we develop two methods, Intersectional Bias Detection (IBD) and Emergent Intersectional Bias Detection (EIBD), to automatically identify the intersectional biases and emergent intersectional biases from static word embeddings in addition to measuring them in contextualized word embeddings.
1 code implementation • 18 Apr 2020 • Autumn Toney, Akshat Pandey, Wei Guo, David Broniatowski, Aylin Caliskan
This paper considers the problem of automatically characterizing overall attitudes and biases that may be associated with emerging information operations via artificial intelligence.
1 code implementation • 13 Feb 2020 • Ryan Steed, Aylin Caliskan
We shed light on the ways in which appearance bias could be embedded in face processing technology and cast further doubt on the practice of predicting subjective traits based on appearances.
no code implementations • 20 Jan 2017 • Edwin Dauber, Aylin Caliskan, Richard Harang, Gregory Shearer, Michael Weisman, Frederica Nelson, Rachel Greenstadt
We show that we can also use these calibration curves in the case that we do not have linking information and thus are forced to classify individual samples directly.
1 code implementation • 25 Aug 2016 • Aylin Caliskan, Joanna J. Bryson, Arvind Narayanan
Here we show for the first time that human-like semantic biases result from the application of standard machine learning to ordinary language---the same sort of language humans are exposed to every day.
3 code implementations • 28 Dec 2015 • Aylin Caliskan, Fabian Yamaguchi, Edwin Dauber, Richard Harang, Konrad Rieck, Rachel Greenstadt, Arvind Narayanan
Many distinguishing features present in source code, e. g. variable names, are removed in the compilation process, and compiler optimization may alter the structure of a program, further obscuring features that are known to be useful in determining authorship.
Cryptography and Security