Search Results for author: Richard P. Wildes

Found 28 papers, 12 papers with code

On Diverse Asynchronous Activity Anticipation

no code implementations ECCV 2020 He Zhao, Richard P. Wildes

We investigate the joint anticipation of long-term activity labels and their corresponding times with the aim of improving both the naturalness and diversity of predictions.

Visual Concept Connectome (VCC): Open World Concept Discovery and their Interlayer Connections in Deep Models

no code implementations2 Apr 2024 Matthew Kowal, Richard P. Wildes, Konstantinos G. Derpanis

Understanding what deep network models capture in their learned representations is a fundamental challenge in computer vision.

Image Classification

Understanding Video Transformers for Segmentation: A Survey of Application and Interpretability

no code implementations18 Oct 2023 Rezaul Karim, Richard P. Wildes

In this survey, we address the above with a thorough discussion of various categories of video segmentation, a component-wise discussion of the state-of-the-art transformer-based models, and a review of related interpretability methods.

Segmentation Video Segmentation +1

A Unified Multiscale Encoder-Decoder Transformer for Video Segmentation

no code implementations CVPR 2023 Rezaul Karim, He Zhao, Richard P. Wildes, Mennatullah Siam

In this paper, we present an end-to-end trainable unified multiscale encoder-decoder transformer that is focused on dense prediction tasks in video.

Action Segmentation Optical Flow Estimation +6

Sports Video Analysis on Large-Scale Data

1 code implementation9 Aug 2022 Dekun Wu, He Zhao, Xingce Bao, Richard P. Wildes

In this paper, we propose a novel large-scale NBA dataset for Sports Video Analysis (NSVA) with a focus on captioning, to address the above challenges.

Action Recognition

Is Appearance Free Action Recognition Possible?

1 code implementation13 Jul 2022 Filip Ilic, Thomas Pock, Richard P. Wildes

Presently, a methodology and corresponding dataset to isolate the effects of dynamic information in video are missing.

Action Recognition Optical Flow Estimation +1

P3IV: Probabilistic Procedure Planning from Instructional Videos with Weak Supervision

1 code implementation CVPR 2022 He Zhao, Isma Hadji, Nikita Dvornik, Konstantinos G. Derpanis, Richard P. Wildes, Allan D. Jepson

Our model is based on a transformer equipped with a memory module, which maps the start and goal observations to a sequence of plausible actions.

Temporal Transductive Inference for Few-Shot Video Object Segmentation

1 code implementation27 Mar 2022 Mennatullah Siam, Konstantinos G. Derpanis, Richard P. Wildes

In this paper, we present a simple but effective temporal transductive inference (TTI) approach that leverages temporal consistency in the unlabelled video frames during few-shot inference.

Meta-Learning Object +3

Interpretable Deep Feature Propagation for Early Action Recognition

no code implementations11 Jul 2021 He Zhao, Richard P. Wildes

Early action recognition (action prediction) from limited preliminary observations plays a critical role for streaming vision systems that demand real-time inference, as video actions often possess elongated temporal spans which cause undesired latency.

Action Recognition

Detecting Biological Locomotion in Video: A Computational Approach

no code implementations26 May 2021 Soo Min Kang, Richard P. Wildes

An algorithm is developed to measure the presence of these traits in tracked objects to determine if they correspond to a biological entity in locomotion.

Navigate

Where Are You Heading? Dynamic Trajectory Prediction With Expert Goal Examples

1 code implementation ICCV 2021 He Zhao, Richard P. Wildes

Goal-conditioned approaches recently have been found very useful to human trajectory prediction, when adequate goal estimates are provided.

Trajectory Prediction

Why Convolutional Networks Learn Oriented Bandpass Filters: Theory and Empirical Support

no code implementations30 Nov 2020 Isma Hadji, Richard P. Wildes

A standard explanation of this result is that these filters reflect the structure of the images that they have been exposed to during training: Natural images typically are locally composed of oriented contours at various scales and oriented bandpass filters are matched to such structure.

Why Convolutional Networks Learn Oriented Bandpass Filters: A Hypothesis

no code implementations25 Sep 2019 Richard P. Wildes

A standard explanation of this result is that these filters reflect the structure of the images that they have been exposed to during training: Natural images typically are locally composed of oriented contours at various scales and oriented bandpass filters are matched to such structure.

What Do We Understand About Convolutional Networks?

3 code implementations23 Mar 2018 Isma Hadji, Richard P. Wildes

This document will review the most prominent proposals using multilayer convolutional architectures.

What have we learned from deep representations for action recognition?

no code implementations CVPR 2018 Christoph Feichtenhofer, Axel Pinz, Richard P. Wildes, Andrew Zisserman

In this paper, we shed light on deep spatiotemporal representations by visualizing what two-stream models have learned in order to recognize actions in video.

Action Recognition Temporal Action Localization

A Spatiotemporal Oriented Energy Network for Dynamic Texture Recognition

1 code implementation ICCV 2017 Isma Hadji, Richard P. Wildes

Another key aspect of the network is its recurrent nature, whereby the output of each layer of processing feeds back to the input.

Dynamic Texture Recognition

Spatiotemporal Multiplier Networks for Video Action Recognition

1 code implementation CVPR 2017 Christoph Feichtenhofer, Axel Pinz, Richard P. Wildes

This paper presents a general ConvNet architecture for video action recognition based on multiplicative interactions of spacetime features.

Action Recognition General Classification +1

Temporal Residual Networks for Dynamic Scene Recognition

1 code implementation CVPR 2017 Christoph Feichtenhofer, Axel Pinz, Richard P. Wildes

Finally, our temporal ResNet boosts recognition performance and establishes a new state-of-the-art on dynamic scene recognition, as well as on the complementary task of action recognition.

Action Recognition Scene Recognition +1

Review of Action Recognition and Detection Methods

1 code implementation21 Oct 2016 Soo Min Kang, Richard P. Wildes

In this report, a thorough review of various action recognition and detection algorithms in computer vision is provided by analyzing the two-step process of a typical action recognition and detection algorithm: (i) extraction and encoding of features, and (ii) classifying features into action classes.

Action Detection Action Recognition +1

Dynamically Encoded Actions Based on Spacetime Saliency

no code implementations CVPR 2015 Christoph Feichtenhofer, Axel Pinz, Richard P. Wildes

By using the resulting definition of saliency during feature pooling we show that action recognition performance achieves state-of-the-art levels on three widely considered action recognition datasets.

Action Recognition Temporal Action Localization

Cannot find the paper you are looking for? You can Submit a new open access paper.