no code implementations • 17 Dec 2024 • Yifei Zhou, Qianlan Yang, Kaixiang Lin, Min Bai, Xiong Zhou, Yu-Xiong Wang, Sergey Levine, Erran Li
We validate PAE on challenging vision-based web navigation, using both real-world and self-hosted websites from WebVoyager and WebArena. To the best of our knowledge, this work represents the first effective learning system to apply autonomous task proposal with RL for agents that generalizes real-world human-annotated benchmarks with SOTA performances.
no code implementations • 11 Dec 2024 • Michal Shlapentokh-Rothman, Yu-Xiong Wang, Derek Hoiem
Visual programming prompts LLMs (large language mod-els) to generate executable code for visual tasks like visual question answering (VQA).
no code implementations • 5 Dec 2024 • Vinayak Gupta, Yunze Man, Yu-Xiong Wang
To address these limitations, we present PaintScene4D, a novel text-to-4D scene generation framework that departs from conventional multi-view generative models in favor of a streamlined architecture that harnesses video generative models trained on diverse real-world datasets.
no code implementations • 2 Dec 2024 • Ziqi Pang, Tianyuan Zhang, Fujun Luan, Yunze Man, Hao Tan, Kai Zhang, William T. Freeman, Yu-Xiong Wang
We introduce RandAR, a decoder-only visual autoregressive (AR) model capable of generating images in arbitrary token orders.
no code implementations • 7 Nov 2024 • Shuhong Zheng, Zhipeng Bao, Ruoyu Zhao, Martial Hebert, Yu-Xiong Wang
Beyond high-fidelity image synthesis, diffusion models have recently exhibited promising results in dense visual perception tasks.
no code implementations • 7 Nov 2024 • Jun-Kun Chen, Yu-Xiong Wang
This paper proposes ProEdit - a simple yet effective framework for high-quality 3D scene editing guided by diffusion distillation in a novel progressive manner.
1 code implementation • 31 Oct 2024 • Kai Yan, Alexander G. Schwing, Yu-Xiong Wang
As suggested by our analysis, in our experiments, we hence find that simply adding TD3 gradients to the finetuning process of ODT effectively improves the online finetuning performance of ODT, especially if ODT is pretrained with low-reward offline data.
no code implementations • 30 Oct 2024 • Anurag Bagchi, Zhipeng Bao, Yu-Xiong Wang, Pavel Tokmakov, Martial Hebert
We present REM, a framework for segmenting a wide range of concepts in video that can be described through natural language.
no code implementations • 18 Oct 2024 • Yuxiang Lu, Shengcao Cao, Yu-Xiong Wang
Vision Foundation Models (VFMs) have demonstrated outstanding performance on numerous downstream tasks.
1 code implementation • 11 Oct 2024 • Xiuyu Yang, Yunze Man, Jun-Kun Chen, Yu-Xiong Wang
The creation of complex 3D scenes tailored to user specifications has been a tedious and challenging task with traditional 3D modeling tools.
no code implementations • 10 Oct 2024 • Shengcao Cao, Liang-Yan Gui, Yu-Xiong Wang
Contrary to the common practice that fine-tunes LMMs with additional grounding supervision, we find that the grounding ability can in fact emerge in LMMs trained without explicit grounding supervision.
1 code implementation • 9 Oct 2024 • Bowen Jin, Ziqi Pang, Bingjun Guo, Yu-Xiong Wang, Jiaxuan You, Jiawei Han
In this paper, we approach an overlooked yet critical task Graph2Image: generating images from multimodal attributed graphs (MMAGs).
1 code implementation • 5 Sep 2024 • Yunze Man, Shuhong Zheng, Zhipeng Bao, Martial Hebert, Liang-Yan Gui, Yu-Xiong Wang
To address this issue, we present a comprehensive study that probes various visual encoding models for 3D scene understanding, identifying the strengths and limitations of each model across different scenarios.
Ranked #3 on
Question Answering
on SQA3D
no code implementations • 26 Jul 2024 • Yunze Man, Yichen Sheng, Jianming Zhang, Liang-Yan Gui, Yu-Xiong Wang
Recent advancements in 3D object reconstruction from single images have primarily focused on improving the accuracy of object shapes.
no code implementations • CVPR 2024 • Jun-Kun Chen, Samuel Rota Bulò, Norman Müller, Lorenzo Porzi, Peter Kontschieder, Yu-Xiong Wang
This paper proposes ConsistDreamer - a novel framework that lifts 2D diffusion models with 3D awareness and 3D consistency, thus enabling high-fidelity instruction-guided scene editing.
no code implementations • CVPR 2024 • Linzhan Mou, Jun-Kun Chen, Yu-Xiong Wang
This paper proposes Instruct 4D-to-4D that achieves 4D awareness and spatial-temporal consistency for 2D diffusion models to generate high-quality instruction-guided dynamic scene editing results.
no code implementations • CVPR 2024 • Junbao Zhou, Ziqi Pang, Yu-Xiong Wang
With recent video object segmentation (VOS) benchmarks evolving to challenging scenarios, we revisit a simple but overlooked strategy: restricting the size of memory banks.
1 code implementation • CVPR 2024 • Yunze Man, Liang-Yan Gui, Yu-Xiong Wang
To address this challenge, we introduce SIG3D, an end-to-end Situation-Grounded model for 3D vision language reasoning.
Ranked #2 on
Question Answering
on SQA3D
no code implementations • 6 Jun 2024 • Qianlan Yang, Yu-Xiong Wang
Training autonomous agents with sparse rewards is a long-standing problem in online reinforcement learning (RL), due to low data efficiency.
no code implementations • 18 Apr 2024 • Shengcao Cao, Jiuxiang Gu, Jason Kuen, Hao Tan, Ruiyi Zhang, Handong Zhao, Ani Nenkova, Liang-Yan Gui, Tong Sun, Yu-Xiong Wang
Using raw images as the sole training data, our method achieves unprecedented performance in self-supervised open-world segmentation, marking a significant milestone towards high-quality open-world entity segmentation in the absence of human-annotated masks.
no code implementations • 28 Mar 2024 • Sirui Xu, Ziyin Wang, Yu-Xiong Wang, Liang-Yan Gui
However, extending such success to 3D dynamic human-object interaction (HOI) generation faces notable challenges, primarily due to the lack of large-scale interaction data and comprehensive descriptions that align with these interactions.
no code implementations • CVPR 2024 • Zhihao Zhang, Shengcao Cao, Yu-Xiong Wang
The limited scale of current 3D shape datasets hinders the advancements in 3D shape understanding, and motivates multi-modal learning approaches which transfer learned knowledge from data-abundant 2D image and language modalities to 3D shapes.
Ranked #1 on
Zero-shot 3D Point Cloud Classification
on ScanObjectNN (Pretrained on ShapeNet)
(using extra training data)
1 code implementation • NeurIPS 2023 • Shengcao Cao, Dhiraj Joshi, Liang-Yan Gui, Yu-Xiong Wang
The human visual perception system demonstrates exceptional capabilities in learning without explicit supervision and understanding the part-to-whole composition of objects.
1 code implementation • CVPR 2024 • Michal Shlapentokh-Rothman, Ansel Blume, Yao Xiao, Yuqun Wu, Sethuraman T V, Heyi Tao, Jae Yong Lee, Wilfredo Torres, Yu-Xiong Wang, Derek Hoiem
We investigate whether region-based representations are effective for recognition.
1 code implementation • NeurIPS 2023 • Jiahua Dong, Yu-Xiong Wang
In addition to the implicit neural radiance field (NeRF) modeling, our key insight is to exploit two sources of regularization that explicitly propagate the editing information across different views, thus ensuring multi-view consistency.
no code implementations • 10 Dec 2023 • Zhipeng Bao, Yijun Li, Krishna Kumar Singh, Yu-Xiong Wang, Martial Hebert
Despite recent significant strides achieved by diffusion-based Text-to-Image (T2I) models, current systems are still less capable of ensuring decent compositional generation aligned with text prompts, particularly for the multi-object generation.
1 code implementation • NeurIPS 2023 • Andy Zhou, Jindong Wang, Yu-Xiong Wang, Haohan Wang
We propose a conceptually simple and lightweight framework for improving the robustness of vision models through the combination of knowledge distillation and data augmentation.
Ranked #13 on
Domain Generalization
on ImageNet-Sketch
1 code implementation • 2 Nov 2023 • Kai Yan, Alexander G. Schwing, Yu-Xiong Wang
It minimizes the primal Wasserstein distance between the learner and expert state occupancies and leverages a contrastively learned distance metric.
2 code implementations • 19 Oct 2023 • Ziqi Pang, Ziyang Xie, Yunze Man, Yu-Xiong Wang
This paper reveals that large language models (LLMs), despite being trained solely on textual data, are surprisingly strong encoders for purely visual tasks in the absence of language.
Ranked #4 on
Question Answering
on SQA3D
2 code implementations • 6 Oct 2023 • Andy Zhou, Kai Yan, Michal Shlapentokh-Rothman, Haohan Wang, Yu-Xiong Wang
By leveraging the in-context learning ability of LMs, we integrate Monte Carlo Tree Search into LATS to enable LMs as agents, along with LM-powered value functions and self-reflections for proficient exploration and enhanced decision-making.
Ranked #15 on
Code Generation
on MBPP
1 code implementation • 2 Oct 2023 • Ziqi Pang, Deva Ramanan, Mengtian Li, Yu-Xiong Wang
Our benchmark inherently captures the disappearance and re-appearance of agents, presenting the emergent challenge of forecasting for occluded agents, which is a safety-critical problem yet overlooked by snapshot-based benchmarks.
1 code implementation • ICCV 2023 • Shuhong Zheng, Zhipeng Bao, Martial Hebert, Yu-Xiong Wang
To tackle the MTVS problem, we propose MuvieNeRF, a framework that incorporates both multi-task and cross-view knowledge to simultaneously synthesize multiple scene properties.
1 code implementation • ICCV 2023 • Yuanyi Zhong, Anand Bhattad, Yu-Xiong Wang, David Forsyth
Dense depth and surface normal predictors should possess the equivariant property to cropping-and-resizing -- cropping the input image should result in cropping the same output image.
no code implementations • 25 Sep 2023 • Zhiqing Sun, Sheng Shen, Shengcao Cao, Haotian Liu, Chunyuan Li, Yikang Shen, Chuang Gan, Liang-Yan Gui, Yu-Xiong Wang, Yiming Yang, Kurt Keutzer, Trevor Darrell
Large Multimodal Models (LMM) are built across modalities and the misalignment between two modalities can result in "hallucination", generating textual outputs that are not grounded by the multimodal information in context.
1 code implementation • ICCV 2023 • Sirui Xu, Zhengyuan Li, Yu-Xiong Wang, Liang-Yan Gui
This paper addresses a novel task of anticipating 3D human-object interactions (HOIs).
no code implementations • 8 Aug 2023 • Neehar Peri, Mengtian Li, Benjamin Wilson, Yu-Xiong Wang, James Hays, Deva Ramanan
LiDAR-based 3D detection plays a vital role in autonomous navigation.
2 code implementations • 3 Aug 2023 • Xinglong Sun, Jean Ponce, Yu-Xiong Wang
Our study reveals that, different from prior work, deformable convolution needs to be applied on an estimated depth map with a relatively high density for better performance.
no code implementations • 28 Jul 2023 • Garvita Allabadi, Ana Lucic, Siddarth Aananth, Tiffany Yang, Yu-Xiong Wang, Vikram Adve
Traditional semi-supervised object detection methods assume a fixed set of object classes (in-distribution or ID classes) during training and deployment, which limits performance in real-world scenarios where unseen classes (out-of-distribution or OOD classes) may appear.
no code implementations • 24 Jun 2023 • Brando Miranda, Patrick Yu, Saumya Goyal, Yu-Xiong Wang, Sanmi Koyejo
Using this analysis, we demonstrate the following: 1. when the formal diversity of a data set is low, PT beats MAML on average and 2. when the formal diversity is high, MAML beats PT on average.
1 code implementation • 8 Jun 2023 • Sirui Xu, Yu-Xiong Wang, Liang-Yan Gui
This paper aims to deal with the ignored real-world complexities in prior work on human motion forecasting, emphasizing the social properties of multi-person motion, the diversity of motion and social interactions, and the complexity of articulated motion.
1 code implementation • 23 May 2023 • Saba Ghaffari, Ehsan Saleh, Alexander G. Schwing, Yu-Xiong Wang, Martin D. Burke, Saurabh Sinha
Protein design, a grand challenge of the day, involves optimization on a fitness landscape, and leading methods adopt a model-based approach where a model is trained on a training set (protein sequences and fitness) and proposes candidates to explore next.
1 code implementation • ICCV 2023 • Ziyang Xie, Ziqi Pang, Yu-Xiong Wang
To further enhance multi-view consistency, we augment the uncertainty network with the global 3D structure optimized by a voxelized neural radiance field (Voxel-NeRF).
no code implementations • 5 May 2023 • Yunze Man, Liang-Yan Gui, Yu-Xiong Wang
In this work, we propose DualCross, a cross-modality cross-domain adaptation framework to facilitate the learning of a more robust monocular bird's-eye-view (BEV) perception model, which transfers the point cloud knowledge from a LiDAR sensor in one domain during the training phase to the camera-only testing scenario in a different domain.
1 code implementation • CVPR 2023 • Shengcao Cao, Dhiraj Joshi, Liang-Yan Gui, Yu-Xiong Wang
Object detectors often suffer from the domain gap between training (source domain) and real-world applications (target domain).
no code implementations • CVPR 2023 • Jun-Kun Chen, Jipeng Lyu, Yu-Xiong Wang
Our key insight is to exploit the explicit point cloud representation as the underlying structure to construct NeRFs, inspired by the intuitive interpretation of NeRF rendering as a process that projects or "plots" the associated 3D point cloud to a 2D image plane.
2 code implementations • CVPR 2023 • Zhipeng Bao, Pavel Tokmakov, Yu-Xiong Wang, Adrien Gaidon, Martial Hebert
Object discovery -- separating objects from the background without manual labels -- is a fundamental open challenge in computer vision.
1 code implementation • 9 Feb 2023 • Sirui Xu, Yu-Xiong Wang, Liang-Yan Gui
Predicting diverse human motions given a sequence of historical poses has received increasing attention.
Ranked #1 on
Human Pose Forecasting
on Human3.6M
(MMADE metric)
1 code implementation • CVPR 2023 • Ziqi Pang, Jie Li, Pavel Tokmakov, Dian Chen, Sergey Zagoruyko, Yu-Xiong Wang
It emphasizes spatio-temporal continuity and integrates both past and future reasoning for tracked objects.
1 code implementation • ICCV 2023 • Jiangwei Yu, Xiang Li, Xinran Zhao, Hongming Zhang, Yu-Xiong Wang
Learning about object state changes in Video Object Segmentation (VOS) is crucial for understanding and interacting with the visual world.
no code implementations • ICCV 2023 • Yuanyi Zhong, Haoran Tang, Jun-Kun Chen, Yu-Xiong Wang
Though self-supervised contrastive learning (CL) has shown its potential to achieve state-of-the-art accuracy without any supervision, its behavior still remains under investigated by academia.
no code implementations • CVPR 2023 • Yunze Man, Liang-Yan Gui, Yu-Xiong Wang
We design a BEV-guided multi-sensor attention block to take queries from BEV embeddings and learn the BEV representation from sensor-specific features.
1 code implementation • 27 Oct 2022 • Kuan-Ying Lee, Yuanyi Zhong, Yu-Xiong Wang
Existing work on continual learning (CL) is primarily devoted to developing algorithms for models trained from scratch.
1 code implementation • 18 Oct 2022 • Kai Yan, Alexander G. Schwing, Yu-Xiong Wang
To better benefit from available demonstrations, we develop a method to Combine Explicit and Implicit Priors (CEIP).
no code implementations • 10 Oct 2022 • Zhiqiu Lin, Deepak Pathak, Yu-Xiong Wang, Deva Ramanan, Shu Kong
LECO requires learning classifiers in distinct time periods (TPs); each TP introduces a new ontology of "fine" labels that refines old ontologies of "coarse" labels (e. g., dog breeds that refine the previous ${\tt dog}$).
1 code implementation • 11 Aug 2022 • Jun-Kun Chen, Yu-Xiong Wang
Being able to learn an effective semantic representation directly on raw point clouds has become a central topic in 3D understanding.
no code implementations • 2 Aug 2022 • Brando Miranda, Patrick Yu, Yu-Xiong Wang, Sanmi Koyejo
This novel insight contextualizes claims that transfer learning solutions are better than meta-learned solutions in the regime of low diversity under a fair comparison.
no code implementations • 10 Jun 2022 • Yuanyi Zhong, Haoran Tang, Junkun Chen, Jian Peng, Yu-Xiong Wang
Our insight has implications in improving the downstream robustness of supervised learning.
no code implementations • 9 Jun 2022 • Mingtong Zhang, Shuhong Zheng, Zhipeng Bao, Martial Hebert, Yu-Xiong Wang
Comprehensive 3D scene understanding, both geometrically and semantically, is important for real-world applications such as robot perception.
2 code implementations • CVPR 2022 • Shaden Alshammari, Yu-Xiong Wang, Deva Ramanan, Shu Kong
In contrast, weight decay penalizes larger weights more heavily and so learns small balanced weights; the MaxNorm constraint encourages growing small weights within a norm ball but caps all the weights by the radius.
Ranked #9 on
Long-tail Learning
on CIFAR-100-LT (ρ=10)
1 code implementation • CVPR 2022 • Zhipeng Bao, Pavel Tokmakov, Allan Jabri, Yu-Xiong Wang, Adrien Gaidon, Martial Hebert
Our experiments demonstrate that, despite only capturing a small subset of the objects that move, this signal is enough to generalize to segment both moving and static instances of dynamic objects.
1 code implementation • 24 Dec 2021 • Brando Miranda, Yu-Xiong Wang, Sanmi Koyejo
Recent work has suggested that a good embedding is all we need to solve many few-shot learning benchmarks.
no code implementations • 24 Dec 2021 • Brando Miranda, Yu-Xiong Wang, Sanmi Koyejo
We hypothesize that the diversity coefficient of the few-shot learning benchmark is predictive of whether meta-learning solutions will succeed or not.
2 code implementations • CVPR 2022 • Lue Fan, Ziqi Pang, Tianyuan Zhang, Yu-Xiong Wang, Hang Zhao, Feng Wang, Naiyan Wang, Zhaoxiang Zhang
In LiDAR-based 3D object detection for autonomous driving, the ratio of the object size to input scene size is significantly smaller compared to 2D detection cases.
Ranked #3 on
3D Object Detection
on waymo cyclist
3 code implementations • ICLR 2022 • Saba Ghaffari, Ehsan Saleh, David Forsyth, Yu-Xiong Wang
In this work, we demonstrate the effectiveness of Firth bias reduction in few-shot classification.
no code implementations • 29 Sep 2021 • Zhipeng Bao, Yu-Xiong Wang, Martial Hebert
Generative modeling has recently shown great promise in computer vision, but it has mostly focused on synthesizing visually realistic images.
1 code implementation • ICCV 2021 • Rajshekhar Das, Yu-Xiong Wang, JoséM. F. Moura
An effective approach to few-shot classification involves a prior model trained on a large-sample base domain, which is then finetuned over the novel few-shot task to yield generalizable representations.
no code implementations • ICCV 2021 • Yuanyi Zhong, Bodi Yuan, Hong Wu, Zhiqiang Yuan, Jian Peng, Yu-Xiong Wang
We leverage the pixel-level L2 loss and the pixel contrastive loss for the two purposes respectively.
no code implementations • 25 Jun 2021 • Zhipeng Bao, Martial Hebert, Yu-Xiong Wang
Generative modeling has recently shown great promise in computer vision, but it has mostly focused on synthesizing visually realistic images.
no code implementations • CVPR 2021 • Weilin Zhang, Yu-Xiong Wang
One critical factor in improving few-shot detection is to address the lack of variation in training data.
1 code implementation • 12 Apr 2021 • Nadine Chang, Zhiding Yu, Yu-Xiong Wang, Anima Anandkumar, Sanja Fidler, Jose M. Alvarez
As a result, image resampling alone is not enough to yield a sufficiently balanced distribution at the object level.
1 code implementation • CVPR 2021 • Yuanyi Zhong, JianFeng Wang, Lijuan Wang, Jian Peng, Yu-Xiong Wang, Lei Zhang
This paper presents a detection-aware pre-training (DAP) approach, which leverages only weakly-labeled classification-style datasets (e. g., ImageNet) for pre-training, but is specifically tailored to benefit object detection tasks.
no code implementations • ICCV 2021 • Liangke Gui, Adrien Bardes, Ruslan Salakhutdinov, Alexander Hauptmann, Martial Hebert, Yu-Xiong Wang
Learning to hallucinate additional examples has recently been shown as a promising direction to address few-shot learning tasks.
no code implementations • 19 Nov 2020 • Weilin Zhang, Yu-Xiong Wang, David A. Forsyth
Learning to detect an object in an image from very few training examples - few-shot object detection - is challenging, because the classifier that sees proposal boxes has very little training data.
no code implementations • 22 Aug 2020 • Vivek Roy, Yan Xu, Yu-Xiong Wang, Kris Kitani, Ruslan Salakhutdinov, Martial Hebert
Recent works have proposed to solve this task by augmenting the training data of the few-shot classes using generative models with the few-shot training samples as the seeds.
1 code implementation • 17 Aug 2020 • Nadine Chang, Jayanth Koushik, Aarti Singh, Martial Hebert, Yu-Xiong Wang, Michael J. Tarr
Methods in long-tail learning focus on improving performance for data-poor (rare) classes; however, performance for such classes remains much lower than performance for more data-rich (frequent) classes.
1 code implementation • ICLR 2021 • Zhipeng Bao, Yu-Xiong Wang, Martial Hebert
We propose a novel task of joint few-shot recognition and novel-view synthesis: given only one or few images of a novel object from arbitrary views with only category annotation, we aim to simultaneously learn an object classifier and generate images of that type of object from new viewpoints.
1 code implementation • ECCV 2020 • Mengtian Li, Yu-Xiong Wang, Deva Ramanan
While past work has studied the algorithmic trade-off between latency and accuracy, there has not been a clear metric to compare different methods along the Pareto optimal latency-accuracy curve.
Ranked #2 on
Real-Time Object Detection
on Argoverse-HD (Detection-Only, Val)
(using extra training data)
1 code implementation • 29 Nov 2019 • Ziqi Pang, Zhiyuan Hu, Pavel Tokmakov, Yu-Xiong Wang, Martial Hebert
Indeed, even the majority of few-shot learning methods rely on a large set of "base classes" for pretraining.
no code implementations • ICCV 2019 • Yu-Xiong Wang, Deva Ramanan, Martial Hebert
Few-shot learning, i. e., learning novel concepts from few examples, is fundamental to practical visual recognition systems.
Ranked #22 on
Few-Shot Object Detection
on MS-COCO (30-shot)
no code implementations • 25 Sep 2019 • Yu-Xiong Wang, Yuki Uchiyama, Martial Hebert, Karteek Alahari
Learning to hallucinate additional examples has recently been shown as a promising direction to address few-shot learning tasks, which aim to learn novel concepts from very few examples.
no code implementations • CVPR 2017 • Yu-Xiong Wang, Deva Ramanan, Martial Hebert
One of their remarkable properties is the ability to transfer knowledge from a large source dataset to a (typically smaller) target dataset.
1 code implementation • CVPR 2019 • Zitian Chen, Yanwei Fu, Yu-Xiong Wang, Lin Ma, Wei Liu, Martial Hebert
Humans can robustly learn novel visual concepts even when images undergo various deformations and lose certain information.
no code implementations • ICCV 2019 • Pavel Tokmakov, Yu-Xiong Wang, Martial Hebert
One of the key limitations of modern deep learning approaches lies in the amount of data required to train them.
no code implementations • ECCV 2018 • Liang-Yan Gui, Yu-Xiong Wang, Xiaodan Liang, Jose M. F. Moura
We explore an approach to forecasting human motion in a few milliseconds given an input 3D skeleton sequence based on a recurrent encoder-decoder framework.
no code implementations • ECCV 2018 • Liang-Yan Gui, Yu-Xiong Wang, Deva Ramanan, Jose M. F. Moura
This paper addresses the problem of few-shot human motion prediction, in the spirit of the recent progress on few-shot learning and meta-learning.
1 code implementation • CVPR 2018 • Yu-Xiong Wang, Ross Girshick, Martial Hebert, Bharath Hariharan
Humans can quickly learn new visual concepts, perhaps because they can easily visualize or imagine what novel objects look like from different views.
no code implementations • NeurIPS 2017 • Yu-Xiong Wang, Deva Ramanan, Martial Hebert
We cast this problem as transfer learning, where knowledge from the data-rich classes in the head of the distribution is transferred to the data-poor classes in the tail.
no code implementations • NeurIPS 2016 • Yu-Xiong Wang, Martial Hebert
Inspired by the transferability properties of CNNs, we introduce an additional unsupervised meta-training stage that exposes multiple top layer units to a large amount of unlabeled real-world images.
no code implementations • CVPR 2015 • Yu-Xiong Wang, Martial Hebert
In this paper, we explore an approach to generating detectors that is radically different from the conventional way of learning a detector from a large corpus of annotated positive and negative data samples.