Search Results for author: Andrei Barbu

Found 31 papers, 11 papers with code

Brain Treebank: Large-scale intracranial recordings from naturalistic language stimuli

no code implementations13 Nov 2024 Christopher Wang, Adam Uri Yaari, Aaditya K Singh, Vighnesh Subramaniam, Dana Rosenfarb, Jan DeWitt, Pranav Misra, Joseph R. Madsen, Scellig Stone, Gabriel Kreiman, Boris Katz, Ignacio Cases, Andrei Barbu

We present the Brain Treebank, a large-scale dataset of electrophysiological neural responses, recorded from intracranial probes while 10 subjects watched one or more Hollywood movies.

Sentence

BrainBits: How Much of the Brain are Generative Reconstruction Methods Using?

no code implementations5 Nov 2024 David Mayo, Christopher Wang, Asa Harbin, Abdulrahman Alabdulkareem, Albert Eaton Shaw, Boris Katz, Andrei Barbu

When evaluating stimuli reconstruction results it is tempting to assume that higher fidelity text and image generation is due to an improved understanding of the brain or more powerful signal extraction from neural recordings.

Image Generation

Using Multimodal Deep Neural Networks to Disentangle Language from Visual Aesthetics

no code implementations31 Oct 2024 Colin Conwell, Christopher Hamblin, Chelsea Boccagno, David Mayo, Jesse Cummings, Leyla Isik, Andrei Barbu

We show that unimodal vision models (e. g. SimCLR) account for the vast majority of explainable variance in these ratings.

Baba Is AI: Break the Rules to Beat the Benchmark

no code implementations18 Jul 2024 Nathan Cloos, Meagan Jens, Michelangelo Naim, Yen-Ling Kuo, Ignacio Cases, Andrei Barbu, Christopher J. Cueva

Humans solve problems by following existing rules and procedures, and also by leaps of creativity to redefine those rules and objectives.

Revealing Vision-Language Integration in the Brain with Multimodal Networks

1 code implementation20 Jun 2024 Vighnesh Subramaniam, Colin Conwell, Christopher Wang, Gabriel Kreiman, Boris Katz, Ignacio Cases, Andrei Barbu

We operationalize sites of multimodal integration as regions where a multimodal vision-language model predicts recordings better than unimodal language, unimodal vision, or linearly-integrated language-vision models.

Contrastive Learning Language Modelling

SecureLLM: Using Compositionality to Build Provably Secure Language Models for Private, Sensitive, and Secret Data

1 code implementation16 May 2024 Abdulrahman Alabdulkareem, Christian M Arnold, Yerim Lee, Pieter M Feenstra, Boris Katz, Andrei Barbu

We reflect the compositional nature of such security mechanisms back into the structure of LLMs to build a provably secure LLM; that we term SecureLLM.

Translation

Can Machines Imitate Humans? Integrative Turing Tests for Vision and Language Demonstrate a Narrowing Gap

no code implementations23 Nov 2022 Mengmi Zhang, Giorgia Dellaferrera, Ankur Sikarwar, Caishun Chen, Marcelo Armendariz, Noga Mudrik, Prachi Agrawal, Spandan Madan, Mranmay Shetty, Andrei Barbu, Haochen Yang, Tanishq Kumar, Shui'Er Han, Aman RAJ Singh, Meghna Sadwani, Stella Dellaferrera, Michele Pizzochero, Brandon Tang, Yew Soon Ong, Hanspeter Pfister, Gabriel Kreiman

To address this question, we turn to the Turing test and systematically benchmark current AIs in their abilities to imitate humans in three language tasks (Image captioning, Word association, and Conversation) and three vision tasks (Object detection, Color estimation, and Attention prediction).

Image Captioning object-detection +1

Query The Agent: Improving sample efficiency through epistemic uncertainty estimation

no code implementations5 Oct 2022 Julian Alverio, Boris Katz, Andrei Barbu

Curricula for goal-conditioned reinforcement learning agents typically rely on poor estimates of the agent's epistemic uncertainty or fail to consider the agents' epistemic uncertainty altogether, resulting in poor sample efficiency.

reinforcement-learning Reinforcement Learning (RL)

On the use of Cortical Magnification and Saccades as Biological Proxies for Data Augmentation

1 code implementation NeurIPS Workshop SVRHM 2021 Binxu Wang, David Mayo, Arturo Deza, Andrei Barbu, Colin Conwell

Critically, we find that random cropping can be substituted by cortical magnification, and saccade-like sampling of the image could also assist the representation learning.

Data Augmentation Representation Learning +1

Neural Regression, Representational Similarity, Model Zoology & Neural Taskonomy at Scale in Rodent Visual Cortex

1 code implementation NeurIPS 2021 Colin Conwell, David Mayo, Andrei Barbu, Michael Buice, George Alvarez, Boris Katz

Using our benchmark as an atlas, we offer preliminary answers to overarching questions about levels of analysis (e. g. do models that better predict the representations of individual neurons also predict representational similarity across neural populations?

Benchmarking Object Recognition +1

Trajectory Prediction with Linguistic Representations

no code implementations19 Oct 2021 Yen-Ling Kuo, Xin Huang, Andrei Barbu, Stephen G. McGill, Boris Katz, John J. Leonard, Guy Rosman

Language allows humans to build mental models that interpret what is happening around them resulting in more accurate long-term predictions.

Trajectory Prediction

Spoken ObjectNet: A Bias-Controlled Spoken Caption Dataset

1 code implementation14 Oct 2021 Ian Palmer, Andrew Rouditchenko, Andrei Barbu, Boris Katz, James Glass

These results show that models trained on other datasets and then evaluated on Spoken ObjectNet tend to perform poorly due to biases in other datasets that the models have learned.

Image Retrieval Language Modelling +1

PHASE: PHysically-grounded Abstract Social Events for Machine Social Perception

no code implementations NeurIPS Workshop SVRHM 2020 Aviv Netanyahu, Tianmin Shu, Boris Katz, Andrei Barbu, Joshua B. Tenenbaum

The ability to perceive and reason about social interactions in the context of physical environments is core to human social intelligence and human-machine cooperation.

Learning a natural-language to LTL executable semantic parser for grounded robotics

no code implementations7 Aug 2020 Christopher Wang, Candace Ross, Yen-Ling Kuo, Boris Katz, Andrei Barbu

We take a step toward robots that can do the same by training a grounded semantic parser, which discovers latent linguistic representations that can be used for the execution of natural-language commands.

Sentence

Compositional Networks Enable Systematic Generalization for Grounded Language Understanding

1 code implementation Findings (EMNLP) 2021 Yen-Ling Kuo, Boris Katz, Andrei Barbu

Recent work has shown that while deep networks can mimic some human language abilities when presented with novel sentences, systematic variation uncovers the limitations in the language-understanding abilities of networks.

Systematic Generalization

Encoding formulas as deep networks: Reinforcement learning for zero-shot execution of LTL formulas

1 code implementation1 Jun 2020 Yen-Ling Kuo, Boris Katz, Andrei Barbu

We demonstrate a reinforcement learning agent which uses a compositional recurrent neural network that takes as input an LTL formula and determines satisfying actions.

Minecraft Multi-Task Learning +2

Measuring Social Biases in Grounded Vision and Language Embeddings

1 code implementation NAACL 2021 Candace Ross, Boris Katz, Andrei Barbu

We generalize the notion of social biases from language embeddings to grounded vision and language embeddings.

Word Embeddings

Deep compositional robotic planners that follow natural language commands

no code implementations12 Feb 2020 Yen-Ling Kuo, Boris Katz, Andrei Barbu

We demonstrate how a sampling-based robotic planner can be augmented to learn to understand a sequence of natural language commands in a continuous configuration space to move and manipulate objects.

ObjectNet: A large-scale bias-controlled dataset for pushing the limits of object recognition models

no code implementations NeurIPS 2019 Andrei Barbu, David Mayo, Julian Alverio, William Luo, Christopher Wang, Dan Gutfreund, Josh Tenenbaum, Boris Katz

Although we focus on object recognition here, data with controls can be gathered at scale using automated tools throughout machine learning to generate datasets that exercise models in new ways thus providing valuable feedback to researchers.

Ranked #51 on Image Classification on ObjectNet (using extra training data)

BIG-bench Machine Learning Image Classification +2

Anchoring and Agreement in Syntactic Annotations

no code implementations EMNLP 2016 Yevgeni Berzak, Yan Huang, Andrei Barbu, Anna Korhonen, Boris Katz

Our agreement results control for parser bias, and are consequential in that they are on par with state of the art parsing performance for English newswire.

Decision Making Dependency Parsing

Do You See What I Mean? Visual Resolution of Linguistic Ambiguities

no code implementations EMNLP 2015 Yevgeni Berzak, Andrei Barbu, Daniel Harari, Boris Katz, Shimon Ullman

Understanding language goes hand in hand with the ability to integrate complex contextual information obtained via perception.

Sentence

Saying What You're Looking For: Linguistics Meets Video Search

no code implementations20 Sep 2013 Andrei Barbu, N. Siddharth, Jeffrey Mark Siskind

We present an approach to searching large video corpora for video clips which depict a natural-language query in the form of a sentence.

object-detection Object Detection +1

Seeing What You're Told: Sentence-Guided Activity Recognition In Video

no code implementations CVPR 2014 N. Siddharth, Andrei Barbu, Jeffrey Mark Siskind

We present a system that demonstrates how the compositional structure of events, in concert with the compositional structure of language, can interplay with the underlying focusing mechanisms in video action recognition, thereby providing a medium, not only for top-down and bottom-up integration, but also for multi-modal integration between vision and language.

Action Recognition Sentence +1

Cannot find the paper you are looking for? You can Submit a new open access paper.