Search Results for author: Barry-John Theobald

Found 20 papers, 3 papers with code

Sample-Efficient Preference-based Reinforcement Learning with Dynamics Aware Rewards

1 code implementation28 Feb 2024 Katherine Metcalf, Miguel Sarabia, Natalie Mackraz, Barry-John Theobald

Preference-based reinforcement learning (PbRL) aligns a robot behavior with human preferences via a reward function learned from binary feedback over agent behaviors.

reinforcement-learning

REALM: Robust Entropy Adaptive Loss Minimization for Improved Single-Sample Test-Time Adaptation

no code implementations7 Sep 2023 Skyler Seto, Barry-John Theobald, Federico Danieli, Navdeep Jaitly, Dan Busbridge

In online F-TTA, a pre-trained model is adapted using a stream of test samples by minimizing a self-supervised objective, such as entropy minimization.

Test-time Adaptation

Spatial LibriSpeech: An Augmented Dataset for Spatial Audio Learning

1 code implementation18 Aug 2023 Miguel Sarabia, Elena Menyaylenko, Alessandro Toso, Skyler Seto, Zakaria Aldeneh, Shadi Pirhosseinloo, Luca Zappella, Barry-John Theobald, Nicholas Apostoloff, Jonathan Sheaffer

We present Spatial LibriSpeech, a spatial audio dataset with over 650 hours of 19-channel audio, first-order ambisonics, and optional distractor noise.

8k Position

Understanding the Robustness of Multi-Exit Models under Common Corruptions

no code implementations3 Dec 2022 Akshay Mehra, Skyler Seto, Navdeep Jaitly, Barry-John Theobald

Furthermore, the lack of calibration increases the inconsistency in the predictions of the model across exits, leading to both inefficient inference and more misclassifications compared with evaluation on in-distribution data.

Rewards Encoding Environment Dynamics Improves Preference-based Reinforcement Learning

no code implementations12 Nov 2022 Katherine Metcalf, Miguel Sarabia, Barry-John Theobald

In this work, we demonstrate that encoding environment dynamics in the reward function (REED) dramatically reduces the number of preference labels required in state-of-the-art preference-based RL frameworks.

reinforcement-learning Reinforcement Learning (RL)

Naturalistic Head Motion Generation from Speech

no code implementations26 Oct 2022 Trisha Mittal, Zakaria Aldeneh, Masha Fedzechkina, Anurag Ranjan, Barry-John Theobald

Synthesizing natural head motion to accompany speech for an embodied conversational agent is necessary for providing a rich interactive experience.

On the role of Lip Articulation in Visual Speech Perception

no code implementations18 Mar 2022 Zakaria Aldeneh, Masha Fedzechkina, Skyler Seto, Katherine Metcalf, Miguel Sarabia, Nicholas Apostoloff, Barry-John Theobald

Previous research has shown that traditional metrics used to optimize and assess models for generating lip motion from speech are not a good indicator of subjective opinion of animation quality.

FedEmbed: Personalized Private Federated Learning

no code implementations18 Feb 2022 Andrew Silva, Katherine Metcalf, Nicholas Apostoloff, Barry-John Theobald

Federated learning enables the deployment of machine learning to problems for which centralized data collection is impractical.

Federated Learning

MorphGAN: One-Shot Face Synthesis GAN for Detecting Recognition Bias

no code implementations9 Dec 2020 Nataniel Ruiz, Barry-John Theobald, Anurag Ranjan, Ahmed Hussein Abdelaziz, Nicholas Apostoloff

Images generated using MorphGAN conserve the identity of the person in the original image, and the provided control over head pose and facial expression allows test sets to be created to identify robustness issues of a facial recognition deep network with respect to pose and expression.

Data Augmentation Face Generation +2

Modality Dropout for Improved Performance-driven Talking Faces

no code implementations27 May 2020 Ahmed Hussen Abdelaziz, Barry-John Theobald, Paul Dixon, Reinhard Knothe, Nicholas Apostoloff, Sachin Kajareker

We use subjective testing to demonstrate: 1) the improvement of audiovisual-driven animation over the equivalent video-only approach, and 2) the improvement in the animation of speech-related facial movements after introducing modality dropout.

Mirroring to Build Trust in Digital Assistants

no code implementations2 Apr 2019 Katherine Metcalf, Barry-John Theobald, Garrett Weinberg, Robert Lee, Ing-Marie Jonsson, Russ Webb, Nicholas Apostoloff

We describe experiments towards building a conversational digital assistant that considers the preferred conversational style of the user.

Learning Sharing Behaviors with Arbitrary Numbers of Agents

no code implementations10 Dec 2018 Katherine Metcalf, Barry-John Theobald, Nicholas Apostoloff

We model the individual behavior for each agent in an interaction and then use a multi-agent fusion model to generate a summary over the expected actions of the group to render the model independent of the number of agents.

Q-Learning

Some observations on computer lip-reading: moving from the dream to the reality

no code implementations3 Oct 2017 Helen L. Bear, Gari Owen, Richard Harvey, Barry-John Theobald

In the quest for greater computer lip-reading performance there are a number of tacit assumptions which are either present in the datasets (high resolution for example) or in the methods (recognition of spoken visual units called visemes for example).

Lip Reading

Which phoneme-to-viseme maps best improve visual-only computer lip-reading?

no code implementations3 Oct 2017 Helen L. Bear, Richard W. Harvey, Barry-John Theobald, Yuxuan Lan

A critical assumption of all current visual speech recognition systems is that there are visual speech units called visemes which can be mapped to units of acoustic speech, the phonemes.

Lip Reading speech-recognition +1

Resolution limits on visual speech recognition

no code implementations3 Oct 2017 Helen L. Bear, Richard Harvey, Barry-John Theobald, Yuxuan Lan

Visual-only speech recognition is dependent upon a number of factors that can be difficult to control, such as: lighting; identity; motion; emotion and expression.

Lip Reading speech-recognition +1

Cannot find the paper you are looking for? You can Submit a new open access paper.