Search Results for author: Onur Cezmi Mutlu

Found 6 papers, 1 papers with code

TikTokActions: A TikTok-Derived Video Dataset for Human Action Recognition

no code implementations14 Feb 2024 Yang Qian, Yinan Sun, Ali Kargarandehkordi, Onur Cezmi Mutlu, Saimourya Surabhi, Pingyi Chen, Zain Jabbar, Dennis Paul Wall, Peter Washington

We find that the performance of the model pre-trained using our Tik-Tok dataset is comparable to models trained on larger action recognition datasets (95. 3% on UCF101 and 53. 24% on HMDB51).

Action Recognition Temporal Action Localization

TempT: Temporal consistency for Test-time adaptation

no code implementations19 Mar 2023 Onur Cezmi Mutlu, Mohammadmahdi Honarmand, Saimourya Surabhi, Dennis P. Wall

We introduce Temporal consistency for Test-time adaptation (TempT) a novel method for test-time adaptation on videos through the use of temporal coherence of predictions across sequential frames as a self-supervision signal.

Facial Expression Recognition Facial Expression Recognition (FER) +2

Classification of Abnormal Hand Movement for Aiding in Autism Detection: Machine Learning Study

1 code implementation18 Aug 2021 Anish Lakkapragada, Aaron Kline, Onur Cezmi Mutlu, Kelley Paskov, Brianna Chrisman, Nate Stockham, Peter Washington, Dennis Wall

This work aims to demonstrate the feasibility of deep learning technologies for detecting hand flapping from unstructured home videos as a first step towards validating whether models and digital technologies can be leveraged to aid with autism diagnoses.

Action Detection Activity Detection +2

Activity Recognition with Moving Cameras and Few Training Examples: Applications for Detection of Autism-Related Headbanging

no code implementations10 Jan 2021 Peter Washington, Aaron Kline, Onur Cezmi Mutlu, Emilie Leblanc, Cathy Hou, Nate Stockham, Kelley Paskov, Brianna Chrisman, Dennis P. Wall

Activity recognition computer vision algorithms can be used to detect the presence of autism-related behaviors, including what are termed "restricted and repetitive behaviors", or stimming, by diagnostic instruments.

Action Detection Activity Detection +1

Training Affective Computer Vision Models by Crowdsourcing Soft-Target Labels

no code implementations10 Jan 2021 Peter Washington, Onur Cezmi Mutlu, Emilie Leblanc, Aaron Kline, Cathy Hou, Brianna Chrisman, Nate Stockham, Kelley Paskov, Catalin Voss, Nick Haber, Dennis Wall

While the F1-score for a one-hot encoded classifier is much higher (94. 33% vs. 78. 68%) with respect to the ground truth CAFE labels, the output probability vector of the crowd-trained classifier more closely resembles the distribution of human labels (t=3. 2827, p=0. 0014).

BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.