no code implementations • ICLR 2019 • Pascal Mettes, Elise van der Pol, Cees G. M. Snoek
The structure is defined by polar prototypes, points on the hypersphere of the output space.
1 code implementation • 17 Jun 2022 • Tejaswi Kasarla, Gertjan J. Burghouts, Max van Spengler, Elise van der Pol, Rita Cucchiara, Pascal Mettes
This paper proposes a simple alternative: encoding maximum separation as an inductive bias in the network by adding one fixed matrix multiplication before computing the softmax activations.
no code implementations • 19 Apr 2022 • Pengwan Yang, Yuki M. Asano, Pascal Mettes, Cees G. M. Snoek
The goal of this paper is to bypass the need for labelled examples in few-shot video understanding at run time.
1 code implementation • CVPR 2022 • Mina GhadimiAtigh, Julian Schoep, Erman Acar, Nanne van Noord, Pascal Mettes
For image segmentation, the current standard is to perform pixel-level optimization and inference in Euclidean output embedding spaces through linear hyperplanes.
no code implementations • 8 Mar 2022 • Pascal Mettes
For universal action models, we first seek to find a hyperspherical optimal transport mapping from unseen action prototypes to the set of all projected test videos.
1 code implementation • 26 Oct 2021 • Carlo Bretti, Pascal Mettes
This paper investigates the problem of zero-shot action recognition, in the setting where no training videos with seen actions are available.
1 code implementation • 25 Oct 2021 • Shuo Chen, Pascal Mettes, Cees G. M. Snoek
Video relation detection forms a new and challenging problem in computer vision, where subjects and objects need to be localized spatio-temporally and a predicate label needs to be assigned if and only if there is an interaction between the two.
no code implementations • 29 Sep 2021 • Pascal Mettes
For universal object models, we outline a weighted transport variant from unseen action embeddings to object embeddings directly.
1 code implementation • ICCV 2021 • Shuo Chen, Zenglin Shi, Pascal Mettes, Cees G. M. Snoek
We also propose Social Fabric: an encoding that represents a pair of object tubelets as a composition of interaction primitives.
1 code implementation • 19 Jul 2021 • Zenglin Shi, Pascal Mettes, Guoyan Zheng, Cees Snoek
In this paper, we find that all existing approaches share a common limitation: reconstruction breaks down in and around the high-frequency parts of CT images.
1 code implementation • 2 Jul 2021 • Zenglin Shi, Pascal Mettes, Subhransu Maji, Cees G. M. Snoek
The deep image prior showed that a randomly initialized network with a suitable architecture can be trained to solve inverse imaging problems by simply optimizing it's parameters to reconstruct a single degraded image.
1 code implementation • NeurIPS 2021 • Mina Ghadimi Atigh, Martin Keller-Ressel, Pascal Mettes
To be able to compute proximities to ideal prototypes, we introduce the penalised Busemann loss.
1 code implementation • 2 Jun 2021 • Zenglin Shi, Yunlu Chen, Efstratios Gavves, Pascal Mettes, Cees G. M. Snoek
The state-of-the-art leverages deep networks to estimate the two core coefficients of the guided filter.
1 code implementation • 10 Apr 2021 • Pascal Mettes, William Thong, Cees G. M. Snoek
This work strives for the classification and localization of human actions in videos, without the need for any labeled video training examples.
no code implementations • CVPR 2021 • Pengwan Yang, Pascal Mettes, Cees G. M. Snoek
This paper introduces the task of few-shot common action localization in time and space.
1 code implementation • 1 Nov 2020 • Victor Zuanazzi, Joris van Vugt, Olaf Booij, Pascal Mettes
This work proposes a metric learning approach for self-supervised scene flow estimation.
1 code implementation • ECCV 2020 • Yunlu Chen, Vincent Tao Hu, Efstratios Gavves, Thomas Mensink, Pascal Mettes, Pengwan Yang, Cees G. M. Snoek
In this paper, we define data augmentation between point clouds as a shortest path linear interpolation.
Ranked #3 on
3D Point Cloud Data Augmentation
on ModelNet40
3D Point Cloud Classification
3D Point Cloud Data Augmentation
+2
1 code implementation • ECCV 2020 • Pengwan Yang, Vincent Tao Hu, Pascal Mettes, Cees G. M. Snoek
The start and end of an action in a long untrimmed video is determined based on just a hand-full of trimmed video examples containing the same action, without knowing their common class label.
2 code implementations • 19 Nov 2019 • William Thong, Pascal Mettes, Cees G. M. Snoek
In this paper, we make the step towards an open setting where multiple visual domains are available.
1 code implementation • 22 Oct 2019 • Andrew Brown, Pascal Mettes, Marcel Worring
Interestingly, when incorporating shifts to all point-wise convolutions in residual networks, 4-connected shifts outperform 8-connected shifts.
1 code implementation • ICCV 2019 • Zenglin Shi, Pascal Mettes, Cees G. M. Snoek
To assist both the density estimation and the focus from segmentation, we also introduce an improved kernel size estimator for the point annotations.
1 code implementation • NeurIPS 2019 • Pascal Mettes, Elise van der Pol, Cees G. M. Snoek
This paper introduces hyperspherical prototype networks, which unify classification and regression with prototypes on hyperspherical output spaces.
no code implementations • 8 Jul 2018 • Pascal Mettes, Cees G. M. Snoek
Rather than disconnecting the spatio-temporal learning from the training, we propose Spatio-Temporal Instance Learning, which enables action localization directly from box proposals in video frames.
Multiple Instance Learning
Spatio-Temporal Action Localization
+1
no code implementations • 29 May 2018 • Pascal Mettes, Cees G. M. Snoek
Experimental evaluation on three action localization datasets shows our pointly-supervised approach (i) is as effective as traditional box-supervision at a fraction of the annotation cost, (ii) is robust to sparse and noisy point annotations, (iii) benefits from pseudo-points during inference, and (iv) outperforms recent weakly-supervised alternatives.
no code implementations • 28 Jul 2017 • Pascal Mettes, Cees G. M. Snoek, Shih-Fu Chang
The goal of this paper is to determine the spatio-temporal location of actions in video.
no code implementations • ICCV 2017 • Pascal Mettes, Cees G. M. Snoek
Action localization and classification experiments on four contemporary action video datasets support our proposal.
no code implementations • 26 Apr 2016 • Pascal Mettes, Jan C. van Gemert, Cees G. M. Snoek
Rather than annotating boxes, we propose to annotate actions in video with points on a sparse subset of frames only.
no code implementations • 23 Feb 2016 • Pascal Mettes, Dennis C. Koelma, Cees G. M. Snoek
To deal with the problems of over-specific classes and classes with few images, we introduce a bottom-up and top-down approach for reorganization of the ImageNet hierarchy based on all its 21, 814 classes and more than 14 million images.
no code implementations • 2 Nov 2015 • Pascal Mettes, Robby T. Tan, Remco C. Veltkamp
Experimental evaluation on the Video Water Database and the DynTex database indicates the effectiveness of the proposed algorithm, outperforming multiple algorithms for dynamic texture recognition and material recognition by ca.
no code implementations • 16 Oct 2015 • Pascal Mettes, Jan C. van Gemert, Cees G. M. Snoek
This work aims for image categorization using a representation of distinctive parts.