Search Results for author: David C. Hogg

Found 9 papers, 1 papers with code

Exploring the GLIDE model for Human Action Effect Prediction

no code implementations PVLAM (LREC) 2022 Fangjun Li, David C. Hogg, Anthony G. Cohn

GLIDE is a generative neural network that can synthesize (inpaint) masked areas of an image, conditioned on a short piece of text.

Advancing Spatial Reasoning in Large Language Models: An In-Depth Evaluation and Enhancement Using the StepGame Benchmark

1 code implementation8 Jan 2024 Fangjun Li, David C. Hogg, Anthony G. Cohn

We analyze GPT's spatial reasoning performance on the rectified benchmark, identifying proficiency in mapping natural language text to spatial relations but limitations in multi-hop reasoning.

Relation Mapping Text Generation

Bayesian uncertainty-weighted loss for improved generalisability on polyp segmentation task

no code implementations13 Sep 2023 Rebecca S. Stone, Pedro E. Chavarrias-Solano, Andrew J. Bulpitt, David C. Hogg, Sharib Ali

While several previous studies have devised methods for segmentation of polyps, most of these methods are not rigorously assessed on multi-center datasets.

Segmentation

3D shape reconstruction of semi-transparent worms

no code implementations CVPR 2023 Thomas P. Ilett, Omer Yuval, Thomas Ranner, Netta Cohen, David C. Hogg

3D shape reconstruction typically requires identifying object features or textures in multiple images of a subject.

3D Shape Reconstruction

Talking Head from Speech Audio using a Pre-trained Image Generator

no code implementations9 Sep 2022 Mohammed M. Alghamdi, He Wang, Andrew J. Bulpitt, David C. Hogg

We train a recurrent neural network to map from speech utterances to displacements in the latent space of the image generator.

SSIM

Exploring the GLIDE model for Human Action-effect Prediction

no code implementations1 Aug 2022 Fangjun Li, David C. Hogg, Anthony G. Cohn

GLIDE is a generative neural network that can synthesize (inpaint) masked areas of an image, conditioned on a short piece of text.

O2A: One-shot Observational learning with Action vectors

no code implementations17 Oct 2018 Leo Pauly, Wisdom C. Agboh, David C. Hogg, Raul Fuentes

The distance between the action vectors from the observed third-person demonstration and trial robot executions is used as a reward for reinforcement learning of the demonstrated task.

One-Shot Learning

CLAD: A Complex and Long Activities Dataset with Rich Crowdsourced Annotations

no code implementations11 Sep 2017 Jawad Tayyub, Majd Hawasly, David C. Hogg, Anthony G. Cohn

This paper introduces a novel activity dataset which exhibits real-life and diverse scenarios of complex, temporally-extended human activities and actions.

Activity Recognition object-detection +1

Natural Language Grounding and Grammar Induction for Robotic Manipulation Commands

no code implementations WS 2017 Muhannad Alomari, Paul Duckworth, Majd Hawasly, David C. Hogg, Anthony G. Cohn

This is achieved by first learning a set of visual {`}concepts{'} that abstract the visual feature spaces into concepts that have human-level meaning.

Cannot find the paper you are looking for? You can Submit a new open access paper.