Search Results for author: Shijian Deng

Found 4 papers, 0 papers with code

Hear Me, See Me, Understand Me: Audio-Visual Autism Behavior Recognition

no code implementations22 Mar 2024 Shijian Deng, Erin E. Kosloski, Siddhi Patel, Zeke A. Barnett, Yiyang Nan, Alexander Kaplan, Sisira Aarukapalli, William T. Doan, Matthew Wang, Harsh Singh, Pamela R. Rollins, Yapeng Tian

To pave the way for further research on this new problem, we intensively explored leveraging foundation models and multimodal large language models across different modalities.

Language Modelling Large Language Model +1

Separating Invisible Sounds Toward Universal Audiovisual Scene-Aware Sound Separation

no code implementations18 Oct 2023 Yiyang Su, Ali Vosoughi, Shijian Deng, Yapeng Tian, Chenliang Xu

The audio-visual sound separation field assumes visible sources in videos, but this excludes invisible sounds beyond the camera's view.

cross-modal alignment

Unveiling Cross Modality Bias in Visual Question Answering: A Causal View with Possible Worlds VQA

no code implementations31 May 2023 Ali Vosoughi, Shijian Deng, Songyang Zhang, Yapeng Tian, Chenliang Xu, Jiebo Luo

In this paper, we first model a confounding effect that causes language and vision bias simultaneously, then propose a counterfactual inference to remove the influence of this effect.

counterfactual Counterfactual Inference +2

Cannot find the paper you are looking for? You can Submit a new open access paper.