Search Results for author: Matt Huenerfauth

Found 15 papers, 1 papers with code

Sign Language Video Anonymization

no code implementations SignLang (LREC) 2022 Zhaoyang Xia, Yuxiao Chen, Qilong Zhangli, Matt Huenerfauth, Carol Neidle, Dimitri Metaxas

We modify a motion-based image animation model to generate high-resolution videos with the signer identity changed, but with the preservation of linguistically significant motions and facial expressions.

Image Animation Optical Flow Estimation

An Isolated-Signing RGBD Dataset of 100 American Sign Language Signs Produced by Fluent ASL Signers

no code implementations LREC 2020 Saad Hassan, Larwan Berke, Elahe Vahdani, Longlong Jing, YingLi Tian, Matt Huenerfauth

We have collected a new dataset consisting of color and depth videos of fluent American Sign Language (ASL) signers performing sequences of 100 ASL signs from a Kinect v2 sensor.

Recognizing American Sign Language Nonmanual Signal Grammar Errors in Continuous Videos

no code implementations1 May 2020 Elahe Vahdani, Longlong Jing, YingLi Tian, Matt Huenerfauth

Our system is able to recognize grammatical elements on ASL-HW-RGBD from manual gestures, facial expressions, and head movements and successfully detect 8 ASL grammatical mistakes.

Sign Language Recognition, Generation, and Translation: An Interdisciplinary Perspective

1 code implementation22 Aug 2019 Danielle Bragg, Oscar Koller, Mary Bellard, Larwan Berke, Patrick Boudrealt, Annelies Braffort, Naomi Caselli, Matt Huenerfauth, Hernisa Kacorri, Tessa Verhoef, Christian Vogler, Meredith Ringel Morris

Developing successful sign language recognition, generation, and translation systems requires expertise in a wide range of fields, including computer vision, computer graphics, natural language processing, human-computer interaction, linguistics, and Deaf culture.

Cultural Vocal Bursts Intensity Prediction Sign Language Recognition +1

Recognizing American Sign Language Manual Signs from RGB-D Videos

no code implementations7 Jun 2019 Longlong Jing, Elahe Vahdani, Matt Huenerfauth, YingLi Tian

In this paper, we propose a 3D Convolutional Neural Network (3DCNN) based multi-stream framework to recognize American Sign Language (ASL) manual signs (consisting of movements of the hands, as well as non-manual face movements in some cases) in real-time from RGB-D videos, by fusing multimodality features including hand gestures, facial expressions, and body poses from multi-channels (RGB, depth, motion, and skeleton joints).

Modeling Acoustic-Prosodic Cues for Word Importance Prediction in Spoken Dialogues

no code implementations WS 2019 Sushant Kafle, Cecilia O. Alm, Matt Huenerfauth

Word error rate, a traditional metric for evaluating automatic speech recognition, fails to capture that some words are more important for a system to transcribe correctly than others.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

A Corpus for Modeling Word Importance in Spoken Dialogue Transcripts

no code implementations LREC 2018 Sushant Kafle, Matt Huenerfauth

Motivated by a project to create a system for people who are deaf or hard-of-hearing that would use automatic speech recognition (ASR) to produce real-time text captions of spoken English during in-person meetings with hearing individuals, we have augmented a transcript of the Switchboard conversational dialogue corpus with an overlay of word-importance annotations, with a numeric score for each word, to indicate its importance to the meaning of each dialogue turn.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

Cannot find the paper you are looking for? You can Submit a new open access paper.