Human Interaction Recognition

8 papers with code • 8 benchmarks • 8 datasets

Human Interaction Recognition (HIR) is a field of study that involves the development of computer algorithms to detect and recognize human interactions in videos, images, or other multimedia content. The goal of HIR is to automatically identify and analyze the social interactions between people, their body language, and facial expressions.

Most implemented papers

Slow-Fast Auditory Streams For Audio Recognition

ekazakos/auditory-slow-fast 5 Mar 2021

We propose a two-stream convolutional network for audio recognition, that operates on time-frequency spectrogram inputs.

Interaction Relational Network for Mutual Action Recognition

mauriciolp/inter-rel-net 11 Oct 2019

Our solution is able to achieve state-of-the-art performance on the traditional interaction recognition datasets SBU and UT, and also on the mutual actions from the large-scale dataset NTU RGB+D.

Two-person Graph Convolutional Network for Skeleton-based Human Interaction Recognition

mgiant/2p-gcn 12 Aug 2022

To overcome the above shortcoming, we introduce a novel unified two-person graph to represent inter-body and intra-body correlations between joints.

Interactive Spatiotemporal Token Attention Network for Skeleton-based General Interactive Action Recognition

Necolizer/ISTA-Net 14 Jul 2023

To address these problems, we propose an Interactive Spatiotemporal Token Attention Network (ISTA-Net), which simultaneously model spatial, temporal, and interactive relations.

Learning Mutual Excitation for Hand-to-Hand and Human-to-Human Interaction Recognition

nkliuyifang/me-gcn 4 Feb 2024

Recognizing interactive actions, including hand-to-hand interaction and human-to-human interaction, has attracted increasing attention for various applications in the field of video analysis and human-robot interaction.

SkateFormer: Skeletal-Temporal Transformer for Human Action Recognition

KAIST-VICLab/SkateFormer 14 Mar 2024

We categorize the key skeletal-temporal relations for action recognition into a total of four distinct types.

Empathic Grounding: Explorations using Multimodal Interaction and Large Language Models with Conversational Agents

arminarj/empathic-grounding 1 Jul 2024

We introduce the concept of "empathic grounding" in conversational agents as an extension of Clark's conceptualization of grounding in conversation in which the grounding criterion includes listener empathy for the speaker's affective state.

CHASE: Learning Convex Hull Adaptive Shift for Skeleton-based Multi-Entity Action Recognition

Necolizer/CHASE 9 Oct 2024

To this end, we introduce a Convex Hull Adaptive Shift based multi-Entity action recognition method (CHASE), which mitigates inter-entity distribution gaps and unbiases subsequent backbones.