Search Results for author: Aylin Caliskan

Found 26 papers, 16 papers with code

'Person' == Light-skinned, Western Man, and Sexualization of Women of Color: Stereotypes in Stable Diffusion

no code implementations30 Oct 2023 Sourojit Ghosh, Aylin Caliskan

We observe how Stable Diffusion outputs of `a person' without any additional gender/nationality information correspond closest to images of men and least with persons of nonbinary gender, and to persons from Europe/North America over Africa/Asia, pointing towards Stable Diffusion having a concerning representation of personhood to be a European/North American man.

Language Modelling

Pre-trained Speech Processing Models Contain Human-Like Biases that Propagate to Speech Emotion Recognition

1 code implementation29 Oct 2023 Isaac Slaughter, Craig Greenberg, Reva Schwartz, Aylin Caliskan

We compare biases found in pre-trained models to biases in downstream models adapted to the task of Speech Emotion Recognition (SER) and find that in 66 of the 96 tests performed (69%), the group that is more associated with positive valence as indicated by the SpEAT also tends to be predicted as speaking with higher valence by the downstream model.

Speech Emotion Recognition

Is the U.S. Legal System Ready for AI's Challenges to Human Values?

no code implementations30 Aug 2023 Inyoung Cheong, Aylin Caliskan, Tadayoshi Kohno

Our interdisciplinary study investigates how effectively U. S. laws confront the challenges posed by Generative AI to human values.

Evaluating Biased Attitude Associations of Language Models in an Intersectional Context

1 code implementation7 Jul 2023 Shiva Omrani Sabbaghi, Robert Wolfe, Aylin Caliskan

Adapting the projection-based approach to embedding association tests that quantify bias, we find that language models exhibit the most biased attitudes against gender identity, social class, and sexual orientation signals in language.

Sentence Word Embeddings

Bias Against 93 Stigmatized Groups in Masked Language Models and Downstream Sentiment Classification Tasks

1 code implementation8 Jun 2023 Katelyn X. Mei, Sonia Fereidooni, Aylin Caliskan

We investigate bias against these groups in English pre-trained Masked Language Models (MLMs) and their downstream sentiment classification tasks.

Sentiment Analysis Sentiment Classification

Contrastive Language-Vision AI Models Pretrained on Web-Scraped Multimodal Data Exhibit Sexual Objectification Bias

1 code implementation21 Dec 2022 Robert Wolfe, Yiwei Yang, Bill Howe, Aylin Caliskan

A first experiment uses standardized images of women from the Sexual OBjectification and EMotion Database, and finds that human characteristics are disassociated from images of objectified women: the model's recognition of emotional state is mediated by whether the subject is fully or partially clothed.

Easily Accessible Text-to-Image Generation Amplifies Demographic Stereotypes at Large Scale

1 code implementation7 Nov 2022 Federico Bianchi, Pratyusha Kalluri, Esin Durmus, Faisal Ladhak, Myra Cheng, Debora Nozza, Tatsunori Hashimoto, Dan Jurafsky, James Zou, Aylin Caliskan

For example, we find cases of prompting for basic traits or social roles resulting in images reinforcing whiteness as ideal, prompting for occupations resulting in amplification of racial and gender disparities, and prompting for objects resulting in reification of American norms.

Text-to-Image Generation

American == White in Multimodal Language-and-Image AI

no code implementations1 Jul 2022 Robert Wolfe, Aylin Caliskan

In an image captioning task, BLIP remarks upon the race of Asian individuals as much as 36% of the time, but never remarks upon race for White individuals.

Image Captioning Question Answering +1

Gender Bias in Word Embeddings: A Comprehensive Analysis of Frequency, Syntax, and Semantics

no code implementations7 Jun 2022 Aylin Caliskan, Pimparkar Parth Ajay, Tessa Charlesworth, Robert Wolfe, Mahzarin R. Banaji

Using the Single-Category Word Embedding Association Test, we demonstrate the widespread prevalence of gender biases that also show differences in: (1) frequencies of words associated with men versus women; (b) part-of-speech tags in gender-associated words; (c) semantic categories in gender-associated words; and (d) valence, arousal, and dominance in gender-associated words.

Word Embeddings

Measuring Gender Bias in Word Embeddings of Gendered Languages Requires Disentangling Grammatical Gender Signals

1 code implementation3 Jun 2022 Shiva Omrani Sabbaghi, Aylin Caliskan

We demonstrate that word embeddings learn the association between a noun and its grammatical gender in grammatically gendered languages, which can skew social gender bias measurements.

Word Embeddings

Markedness in Visual Semantic AI

1 code implementation23 May 2022 Robert Wolfe, Aylin Caliskan

The model is more likely to rank the unmarked "person" label higher than labels denoting gender for Male individuals (26. 7% of the time) vs.

Evidence for Hypodescent in Visual Semantic AI

1 code implementation22 May 2022 Robert Wolfe, Mahzarin R. Banaji, Aylin Caliskan

We examine the state-of-the-art multimodal "visual semantic" model CLIP ("Contrastive Language Image Pretraining") for the rule of hypodescent, or one-drop rule, whereby multiracial people are more likely to be assigned a racial or ethnic label corresponding to a minority or disadvantaged racial or ethnic group than to the equivalent majority or advantaged group.

MORPH

VAST: The Valence-Assessing Semantics Test for Contextualizing Language Models

1 code implementation14 Mar 2022 Robert Wolfe, Aylin Caliskan

VAST, the Valence-Assessing Semantics Test, is a novel intrinsic evaluation task for contextualized word embeddings (CWEs).

Word Embeddings Word Similarity

Contrastive Visual Semantic Pretraining Magnifies the Semantics of Natural Language Representations

no code implementations ACL 2022 Robert Wolfe, Aylin Caliskan

We find that contrastive visual semantic pretraining significantly mitigates the anisotropy found in contextualized word embeddings from GPT-2, such that the intra-layer self-similarity (mean pairwise cosine similarity) of CLIP word embeddings is under . 25 in all layers, compared to greater than . 95 in the top layer of GPT-2.

Image Captioning Semantic Textual Similarity +3

Low Frequency Names Exhibit Bias and Overfitting in Contextualizing Language Models

no code implementations EMNLP 2021 Robert Wolfe, Aylin Caliskan

Moreover, we find Spearman's r between racial bias and name frequency in BERT of . 492, indicating that lower-frequency minority group names are more associated with unpleasantness.

Image Representations Learned With Unsupervised Pre-Training Contain Human-like Biases

1 code implementation28 Oct 2020 Ryan Steed, Aylin Caliskan

Recent advances in machine learning leverage massive datasets of unlabeled images from the web to learn general-purpose image representations for tasks from image classification to face recognition.

BIG-bench Machine Learning Face Recognition +2

Disparate Impact of Artificial Intelligence Bias in Ridehailing Economy's Price Discrimination Algorithms

no code implementations8 Jun 2020 Akshat Pandey, Aylin Caliskan

An analysis of 100 million ridehailing samples from the city of Chicago indicates a significant disparate impact in fare pricing of neighborhoods due to AI bias learned from ridehailing utilization patterns associated with demographic attributes.

Decision Making Fairness

Detecting Emergent Intersectional Biases: Contextualized Word Embeddings Contain a Distribution of Human-like Biases

1 code implementation6 Jun 2020 Wei Guo, Aylin Caliskan

Furthermore, we develop two methods, Intersectional Bias Detection (IBD) and Emergent Intersectional Bias Detection (EIBD), to automatically identify the intersectional biases and emergent intersectional biases from static word embeddings in addition to measuring them in contextualized word embeddings.

Bias Detection Sentence +1

Automatically Characterizing Targeted Information Operations Through Biases Present in Discourse on Twitter

1 code implementation18 Apr 2020 Autumn Toney, Akshat Pandey, Wei Guo, David Broniatowski, Aylin Caliskan

This paper considers the problem of automatically characterizing overall attitudes and biases that may be associated with emerging information operations via artificial intelligence.

A Set of Distinct Facial Traits Learned by Machines Is Not Predictive of Appearance Bias in the Wild

1 code implementation13 Feb 2020 Ryan Steed, Aylin Caliskan

We shed light on the ways in which appearance bias could be embedded in face processing technology and cast further doubt on the practice of predicting subjective traits based on appearances.

Face Recognition Transfer Learning

Git Blame Who?: Stylistic Authorship Attribution of Small, Incomplete Source Code Fragments

no code implementations20 Jan 2017 Edwin Dauber, Aylin Caliskan, Richard Harang, Gregory Shearer, Michael Weisman, Frederica Nelson, Rachel Greenstadt

We show that we can also use these calibration curves in the case that we do not have linking information and thus are forced to classify individual samples directly.

Authorship Attribution

Semantics derived automatically from language corpora contain human-like biases

1 code implementation25 Aug 2016 Aylin Caliskan, Joanna J. Bryson, Arvind Narayanan

Here we show for the first time that human-like semantic biases result from the application of standard machine learning to ordinary language---the same sort of language humans are exposed to every day.

BIG-bench Machine Learning Ethics +1

When Coding Style Survives Compilation: De-anonymizing Programmers from Executable Binaries

3 code implementations28 Dec 2015 Aylin Caliskan, Fabian Yamaguchi, Edwin Dauber, Richard Harang, Konrad Rieck, Rachel Greenstadt, Arvind Narayanan

Many distinguishing features present in source code, e. g. variable names, are removed in the compilation process, and compiler optimization may alter the structure of a program, further obscuring features that are known to be useful in determining authorship.

Cryptography and Security

Cannot find the paper you are looking for? You can Submit a new open access paper.