Search Results for author: Alex Andonian

Found 16 papers, 10 papers with code

Locating and Editing Factual Associations in GPT

2 code implementations10 Feb 2022 Kevin Meng, David Bau, Alex Andonian, Yonatan Belinkov

To test our hypothesis that these computations correspond to factual association recall, we modify feed-forward weights to update specific factual associations using Rank-One Model Editing (ROME).

counterfactual Model Editing +2

Temporal Relational Reasoning in Videos

5 code implementations ECCV 2018 Bolei Zhou, Alex Andonian, Aude Oliva, Antonio Torralba

Temporal relational reasoning, the ability to link meaningful transformations of objects or entities over time, is a fundamental property of intelligent species.

Action Classification Action Recognition In Videos +4

Mass-Editing Memory in a Transformer

2 code implementations13 Oct 2022 Kevin Meng, Arnab Sen Sharma, Alex Andonian, Yonatan Belinkov, David Bau

Recent work has shown exciting promise in updating large language models with new memories, so as to replace obsolete information or add specialized knowledge.

Language Modelling

Cross-view Semantic Segmentation for Sensing Surroundings

1 code implementation9 Jun 2019 Bowen Pan, Jiankai Sun, Ho Yin Tiga Leung, Alex Andonian, Bolei Zhou

Our further experiment on a LoCoBot robot shows that our model enables the surrounding sensing capability from 2D image input.

Domain Adaptation Semantic Segmentation

GANalyze: Toward Visual Definitions of Cognitive Image Properties

1 code implementation ICCV 2019 Authors, :, Lore Goetschalckx, Alex Andonian, Aude Oliva, Phillip Isola

We introduce a framework that uses Generative Adversarial Networks (GANs) to study cognitive properties like memorability, aesthetics, and emotional valence.

Unsupervised Learning from Video with Deep Neural Embeddings

1 code implementation CVPR 2020 Chengxu Zhuang, Tianwei She, Alex Andonian, Max Sobol Mark, Daniel Yamins

Because of the rich dynamical structure of videos and their ubiquity in everyday life, it is a natural idea that video data could serve as a powerful unsupervised learning signal for training visual representations in deep neural networks.

Action Recognition Object Recognition

Contrastive Feature Loss for Image Prediction

1 code implementation12 Nov 2021 Alex Andonian, Taesung Park, Bryan Russell, Phillip Isola, Jun-Yan Zhu, Richard Zhang

Training supervised image synthesis models requires a critic to compare two images: the ground truth to the result.

Image Generation

VA-RED$^2$: Video Adaptive Redundancy Reduction

no code implementations ICLR 2021 Bowen Pan, Rameswar Panda, Camilo Fosco, Chung-Ching Lin, Alex Andonian, Yue Meng, Kate Saenko, Aude Oliva, Rogerio Feris

An inherent property of real-world videos is the high correlation of information across frames which can translate into redundancy in either temporal or spatial feature maps of the models, or both.

Robust Cross-Modal Representation Learning with Progressive Self-Distillation

no code implementations CVPR 2022 Alex Andonian, Shixing Chen, Raffay Hamid

The learning objective of vision-language approach of CLIP does not effectively account for the noisy many-to-many correspondences found in web-harvested image captioning datasets, which contributes to its compute and data inefficiency.

Ranked #98 on Image Classification on ObjectNet (using extra training data)

Contrastive Learning Image Captioning +5

Deepfake Caricatures: Amplifying attention to artifacts increases deepfake detection by humans and machines

no code implementations1 Jun 2022 Camilo Fosco, Emilie Josephs, Alex Andonian, Allen Lee, Xi Wang, Aude Oliva

Second, they allow us to generate novel "Deepfake Caricatures": transformations of the deepfake that exacerbate artifacts to improve human detection.

DeepFake Detection Face Swapping +2

Three ways to improve feature alignment for open vocabulary detection

no code implementations23 Mar 2023 Relja Arandjelović, Alex Andonian, Arthur Mensch, Olivier J. Hénaff, Jean-Baptiste Alayrac, Andrew Zisserman

The core problem in zero-shot open vocabulary detection is how to align visual and text features, so that the detector performs well on unseen classes.

Language Modelling

Cannot find the paper you are looking for? You can Submit a new open access paper.