Sign Language Recognition

54 papers with code • 7 benchmarks • 19 datasets

Sign Language Recognition is a computer vision and natural language processing task that involves automatically recognizing and translating sign language gestures into written or spoken language. The goal of sign language recognition is to develop algorithms that can understand and interpret sign language, enabling people who use sign language as their primary mode of communication to communicate more easily with non-signers.

( Image credit: Word-level Deep Sign Language Recognition from Video: A New Large-scale Dataset and Methods Comparison )

Most implemented papers

Learning to Estimate 3D Hand Pose from Single RGB Images

lmb-freiburg/hand3d ICCV 2017

Low-cost consumer depth cameras and deep learning have enabled reasonable 3D hand pose estimation from single depth images.

BlazePose: On-device Real-time Body Pose tracking

google/mediapipe 17 Jun 2020

We present BlazePose, a lightweight convolutional neural network architecture for human pose estimation that is tailored for real-time inference on mobile devices.

Skeleton Aware Multi-modal Sign Language Recognition

jackyjsy/CVPR21Chal-SLR 16 Mar 2021

Sign language is commonly used by deaf or speech impaired people to communicate but requires significant effort to master.

A Simple Multi-Modality Transfer Learning Baseline for Sign Language Translation

FangyunWei/SLRT CVPR 2022

Concretely, we pretrain the sign-to-gloss visual network on the general domain of human actions and the within-domain of a sign-to-gloss dataset, and pretrain the gloss-to-text translation network on the general domain of a multilingual corpus and the within-domain of a gloss-to-text corpus.

Fingerspelling recognition in the wild with iterative visual attention

chevalierNoir/asl-iter-attn ICCV 2019

In this paper we focus on recognition of fingerspelling sequences in American Sign Language (ASL) videos collected in the wild, mainly from YouTube and Deaf social media.

Word-level Deep Sign Language Recognition from Video: A New Large-scale Dataset and Methods Comparison

dxli94/WLASL 24 Oct 2019

Based on this new large-scale dataset, we are able to experiment with several deep learning methods for word-level sign recognition and evaluate their performances in large scale scenarios.

TSPNet: Hierarchical Feature Learning via Temporal Semantic Pyramid for Sign Language Translation

verashira/TSPNet NeurIPS 2020

Sign language translation (SLT) aims to interpret sign video sequences into text-based natural language sentences.

Context Matters: Self-Attention for Sign Language Recognition

faresbs/slrt 12 Jan 2021

For that reason, we apply attention to synchronize and help capture entangled dependencies between the different sign language components.

Visual Alignment Constraint for Continuous Sign Language Recognition

ycmin95/VAC_CSLR ICCV 2021

Specifically, the proposed VAC comprises two auxiliary losses: one focuses on visual features only, and the other enforces prediction alignment between the feature extractor and the alignment module.

Sign Language Recognition via Skeleton-Aware Multi-Model Ensemble

jackyjsy/sam-slr-v2 12 Oct 2021

Current Sign Language Recognition (SLR) methods usually extract features via deep neural networks and suffer overfitting due to limited and noisy data.