no code implementations • 10 Feb 2025 • Amin Adibi, Xu Cao, Zongliang Ji, Jivat Neet Kaur, Winston Chen, Elizabeth Healey, Brighton Nuwagira, Wenqian Ye, Geoffrey Woollard, Maxwell A Xu, Hejie Cui, Johnny Xi, Trenton Chang, Vasiliki Bikia, Nicole Zhang, Ayush Noori, Yuan Xia, Md. Belal Hossain, Hanna A. Frank, Alina Peluso, Yuan Pu, Shannon Zejiang Shen, John Wu, Adibvafa Fallahpour, Sazan Mahbub, Ross Duncan, Yuwei Zhang, Yurui Cao, Zuheng Xu, Michael Craig, Rahul G. Krishnan, Rahmatollah Beheshti, James M. Rehg, Mohammad Ehsanul Karim, Megan Coffee, Leo Anthony Celi, Jason Alan Fries, Mohsen Sadatsafavi, Dennis Shung, Shannon McWeeney, Jessica Dafflon, Sarah Jabbour
The fourth Machine Learning for Health (ML4H) symposium was held in person on December 15th and 16th, 2024, in the traditional, ancestral, and unceded territories of the Musqueam, Squamish, and Tsleil-Waututh Nations in Vancouver, British Columbia, Canada.
no code implementations • 3 Feb 2025 • Mithun Saha, Maxwell A. Xu, Wanting Mao, Sameer Neupane, James M. Rehg, Santosh Kumar
Photoplethysmography (PPG)-based foundation models are gaining traction due to the widespread use of PPG in biosignal monitoring and their potential to generalize across diverse health applications.
no code implementations • 8 Jan 2025 • Zixuan Huang, Mark Boss, Aaryaman Vasishta, James M. Rehg, Varun Jampani
Generative methods handle uncertain regions better by modeling distributions, but are computationally expensive and the generation is often misaligned with visible surfaces.
1 code implementation • 12 Dec 2024 • Fiona Ryan, Ajay Bati, Sangmin Lee, Daniel Bolya, Judy Hoffman, James M. Rehg
We address the problem of gaze target estimation, which aims to predict where a person is looking in a scene.
1 code implementation • 9 Dec 2024 • Kevin Gao, Maxwell A. Xu, James M. Rehg, Alexander Moreno
We introduce PyPulse, a Python package for imputation of biosignals in both clinical and wearable sensor settings.
no code implementations • 2 Dec 2024 • Bolin Lai, Felix Juefei-Xu, Miao Liu, Xiaoliang Dai, Nikhil Mehta, Chenguang Zhu, Zeyi Huang, James M. Rehg, Sangmin Lee, Ning Zhang, Tong Xiao
We also introduce a relation regularization method to further disentangle image transformation features from irrelevant contents in exemplar images.
no code implementations • 27 Nov 2024 • Tarik Can Ozden, Ozgur Kara, Oguzhan Akcin, Kerem Zaman, Shashank Srivastava, Sandeep P. Chinchali, James M. Rehg
Extensive qualitative and quantitative results demonstrate that our model is scalable, optimization-free, adaptable to various diffusion-based editing tools, robust against counter-attacks, and, for the first time, effectively protects video content from editing.
no code implementations • 27 Nov 2024 • Maxwell A. Xu, Jaya Narain, Gregory Darnell, Haraldur Hallgrimsson, Hyewon Jeong, Darren Forde, Richard Fineman, Karthik J. Raghuram, James M. Rehg, Shirley Ren
We present RelCon, a novel self-supervised *Rel*ative *Con*trastive learning approach that uses a learnable distance measure in combination with a softened contrastive loss for training an motion foundation model from wearable sensors.
no code implementations • 26 Nov 2024 • Xiang Li, Zixuan Huang, Anh Thai, James M. Rehg
Symmetry is a ubiquitous and fundamental property in the visual world, serving as a critical cue for perception and structure interpretation.
no code implementations • 18 Nov 2024 • Xu Cao, Kaizhao Liang, Kuei-Da Liao, Tianren Gao, Wenqian Ye, Jintai Chen, Zhiguang Ding, Jianguo Cao, James M. Rehg, Jimeng Sun
To address this challenge, we propose the first Medical Video Generation (MVG) framework that enables controlled manipulation of disease-related image and video features, allowing precise, realistic, and personalized simulations of disease progression.
no code implementations • 17 Oct 2024 • Bolin Lai, Sam Toyer, Tushar Nagarajan, Rohit Girdhar, Shengxin Zha, James M. Rehg, Kris Kitani, Kristen Grauman, Ruta Desai, Miao Liu
Predicting future human behavior is an increasingly popular topic in computer vision, driven by the interest in applications such as autonomous vehicles, digital assistants and human-robot interactions.
1 code implementation • 9 Sep 2024 • Bikram Boote, Anh Thai, Wenqi Jia, Ozgur Kara, Stefan Stojanov, James M. Rehg, Sangmin Lee
Point tracking is a fundamental problem in computer vision with numerous applications in AR and robotics.
no code implementations • 12 Jul 2024 • Anh Thai, Weiyao Wang, Hao Tang, Stefan Stojanov, Matt Feiszli, James M. Rehg
While substantial progress has been made in 2D object part segmentation, the 3D counterpart has received less attention, in part due to the scarcity of annotated 3D datasets, which are expensive to collect.
1 code implementation • 27 Jun 2024 • Hui Wei, Maxwell A. Xu, Colin Samplawski, James M. Rehg, Santosh Kumar, Benjamin M. Marlin
Wearable sensors enable health researchers to continuously collect data pertaining to the physiological state of individuals in real-world settings.
no code implementations • 24 Jun 2024 • Wenqian Ye, Guangtao Zheng, Yunsheng Ma, Xu Cao, Bolin Lai, James M. Rehg, Aidong Zhang
Our findings illuminate the persistence of the reliance on spurious correlations from these models and underscore the urge for new methodologies to mitigate spurious biases.
1 code implementation • 14 Jun 2024 • Xu Cao, Bolin Lai, Wenqian Ye, Yunsheng Ma, Joerg Heintz, Jintai Chen, Jianguo Cao, James M. Rehg
Recently, Multimodal Large Language Models (MLLMs) have shown great promise in language-guided perceptual tasks such as recognition, segmentation, and object detection.
no code implementations • CVPR 2024 • Zixuan Huang, Justin Johnson, Shoubhik Debnath, James M. Rehg, Chao-yuan Wu
We present PointInfinity, an efficient family of point cloud diffusion models.
no code implementations • CVPR 2024 • Sangmin Lee, Bolin Lai, Fiona Ryan, Bikram Boote, James M. Rehg
Furthermore, we propose a novel multimodal baseline that leverages densely aligned language-visual representations by synchronizing visual features with their corresponding utterances.
1 code implementation • CVPR 2024 • Xu Cao, Tong Zhou, Yunsheng Ma, Wenqian Ye, Can Cui, Kun Tang, Zhipeng Cao, Kaizhao Liang, Ziran Wang, James M. Rehg, Chao Zheng
Specifically we annotate and leverage large-scale broad-coverage traffic and map data extracted from huge HD map annotations and use CLIP and LLaMA-2 / Vicuna to finetune a baseline model with instruction-following data.
1 code implementation • CVPR 2024 • Zixuan Huang, Stefan Stojanov, Anh Thai, Varun Jampani, James M. Rehg
In contrast, the traditional approach to this problem is regression-based, where deterministic models are trained to directly regress the object shape.
no code implementations • CVPR 2024 • Wenqi Jia, Miao Liu, Hao Jiang, Ishwarya Ananthabhotla, James M. Rehg, Vamsi Krishna Ithapu, Ruohan Gao
We propose a unified multi-modal framework -- Audio-Visual Conversational Attention (AV-CONV), for the joint prediction of conversation behaviors -- speaking and listening -- for both the camera wearer as well as all other social partners present in the egocentric video.
1 code implementation • CVPR 2024 • Yunsheng Ma, Can Cui, Xu Cao, Wenqian Ye, Peiran Liu, Juanwu Lu, Amr Abdelraouf, Rohit Gupta, Kyungtae Han, Aniket Bera, James M. Rehg, Ziran Wang
Autonomous driving (AD) has made significant strides in recent years.
3 code implementations • CVPR 2024 • Ozgur Kara, Bariscan Kurtkaya, Hidir Yesiltepe, James M. Rehg, Pinar Yanardag
Recent advancements in diffusion-based models have demonstrated significant success in generating images from text.
no code implementations • 6 Dec 2023 • Bolin Lai, Xiaoliang Dai, Lawrence Chen, Guan Pang, James M. Rehg, Miao Liu
Additionally, existing diffusion-based image manipulation models are sub-optimal in controlling the state transition of an action in egocentric image pixel space because of the domain gap.
1 code implementation • NeurIPS 2023 • Anh Thai, Ahmad Humayun, Stefan Stojanov, Zixuan Huang, Bikram Boote, James M. Rehg
This paper introduces Low-shot Object Learning with Mutual Exclusivity Bias (LSME), the first computational framing of mutual exclusivity bias, a phenomenon commonly observed in infants during word learning.
no code implementations • 30 Nov 2023 • Meera Hahn, Amit Raj, James M. Rehg
The challenging task of Vision-and-Language Navigation (VLN) requires embodied agents to follow natural language instructions to reach a goal location or object (e. g. `walk down the hallway and turn left at the piano').
2 code implementations • CVPR 2024 • Kristen Grauman, Andrew Westbury, Lorenzo Torresani, Kris Kitani, Jitendra Malik, Triantafyllos Afouras, Kumar Ashutosh, Vijay Baiyya, Siddhant Bansal, Bikram Boote, Eugene Byrne, Zach Chavis, Joya Chen, Feng Cheng, Fu-Jen Chu, Sean Crane, Avijit Dasgupta, Jing Dong, Maria Escobar, Cristhian Forigua, Abrham Gebreselasie, Sanjay Haresh, Jing Huang, Md Mohaiminul Islam, Suyog Jain, Rawal Khirodkar, Devansh Kukreja, Kevin J Liang, Jia-Wei Liu, Sagnik Majumder, Yongsen Mao, Miguel Martin, Effrosyni Mavroudi, Tushar Nagarajan, Francesco Ragusa, Santhosh Kumar Ramakrishnan, Luigi Seminara, Arjun Somayazulu, Yale Song, Shan Su, Zihui Xue, Edward Zhang, Jinxu Zhang, Angela Castillo, Changan Chen, Xinzhu Fu, Ryosuke Furuta, Cristina Gonzalez, Prince Gupta, Jiabo Hu, Yifei HUANG, Yiming Huang, Weslie Khoo, Anush Kumar, Robert Kuo, Sach Lakhavani, Miao Liu, Mi Luo, Zhengyi Luo, Brighid Meredith, Austin Miller, Oluwatumininu Oguntola, Xiaqing Pan, Penny Peng, Shraman Pramanick, Merey Ramazanova, Fiona Ryan, Wei Shan, Kiran Somasundaram, Chenan Song, Audrey Southerland, Masatoshi Tateno, Huiyu Wang, Yuchen Wang, Takuma Yagi, Mingfei Yan, Xitong Yang, Zecheng Yu, Shengxin Cindy Zha, Chen Zhao, Ziwei Zhao, Zhifan Zhu, Jeff Zhuo, Pablo Arbelaez, Gedas Bertasius, David Crandall, Dima Damen, Jakob Engel, Giovanni Maria Farinella, Antonino Furnari, Bernard Ghanem, Judy Hoffman, C. V. Jawahar, Richard Newcombe, Hyun Soo Park, James M. Rehg, Yoichi Sato, Manolis Savva, Jianbo Shi, Mike Zheng Shou, Michael Wray
We present Ego-Exo4D, a diverse, large-scale multimodal multiview video dataset and benchmark challenge.
1 code implementation • 1 Nov 2023 • Maxwell A. Xu, Alexander Moreno, Hui Wei, Benjamin M. Marlin, James M. Rehg
The success of self-supervised contrastive learning hinges on identifying positive data pairs, such that when they are pushed together in embedding space, the space encodes useful information for subsequent downstream tasks.
1 code implementation • 10 Jun 2023 • Supriya Nagesh, Nina Mishra, Yonatan Naamad, James M. Rehg, Mehul A. Shah, Alexei Wagner
Machine learning models perform well on several healthcare tasks and can help reduce the burden on the healthcare system.
no code implementations • 6 May 2023 • Bolin Lai, Fiona Ryan, Wenqi Jia, Miao Liu, James M. Rehg
Motivated by this observation, we introduce the first model that leverages both the video and audio modalities for egocentric gaze anticipation.
no code implementations • CVPR 2023 • Zixuan Huang, Varun Jampani, Anh Thai, Yuanzhen Li, Stefan Stojanov, James M. Rehg
We present ShapeClipper, a novel method that reconstructs 3D object shapes from real-world single-view RGB images.
no code implementations • CVPR 2023 • Fiona Ryan, Hao Jiang, Abhinav Shukla, James M. Rehg, Vamsi Krishna Ithapu
In a noisy conversation environment such as a dinner party, people often exhibit selective auditory attention, or the ability to focus on a particular speaker while tuning out others.
no code implementations • 16 Dec 2022 • Bolin Lai, Hongxin Zhang, Miao Liu, Aryan Pariani, Fiona Ryan, Wenqi Jia, Shirley Anugrah Hayati, James M. Rehg, Diyi Yang
We also explore the generalization ability of language models for persuasion modeling and the role of persuasion strategies in predicting social deduction game outcomes.
1 code implementation • 14 Dec 2022 • Maxwell A. Xu, Alexander Moreno, Supriya Nagesh, V. Burak Aydemir, David W. Wetter, Santosh Kumar, James M. Rehg
The promise of Mobile Health (mHealth) is the ability to use wearable sensors to monitor participant physiology at high frequencies during daily life to enable temporally-precise health interventions.
1 code implementation • 28 Nov 2022 • Stefan Stojanov, Anh Thai, Zixuan Huang, James M. Rehg
A hallmark of the deep learning era for computer vision is the successful use of large-scale labeled datasets to train feature representations for tasks ranging from object recognition and semantic segmentation to optical flow estimation and novel view synthesis of 3D scenes.
no code implementations • 10 Oct 2022 • Meera Hahn, James M. Rehg
We address the challenging task of Localization via Embodied Dialog (LED).
no code implementations • 8 Aug 2022 • Bolin Lai, Miao Liu, Fiona Ryan, James M. Rehg
To this end, we design the transformer encoder to embed the global context as one additional visual token and further propose a novel Global-Local Correlation (GLC) module to explicitly model the correlation of the global token and each local token.
no code implementations • 21 Apr 2022 • Zixuan Huang, Stefan Stojanov, Anh Thai, Varun Jampani, James M. Rehg
We present a novel 3D shape reconstruction method which learns to predict an implicit 3D shape representation from a single RGB image.
1 code implementation • 21 Mar 2022 • Wenqi Jia, Miao Liu, James M. Rehg
We introduce the novel problem of anticipating a time series of future hand masks from egocentric video.
no code implementations • 1 Nov 2021 • Alexander Moreno, Supriya Nagesh, Zhenke Wu, Walter Dempsey, James M. Rehg
Theoretically, we show new existence results for both kernel exponential and deformed exponential families, and that the deformed case has similar approximation capabilities to kernel exponential families.
no code implementations • 1 Nov 2021 • Supriya Nagesh, Alexander Moreno, Stephanie M. Carpenter, Jamie Yap, Soujanya Chatterjee, Steven Lloyd Lizotte, Neng Wan, Santosh Kumar, Cho Lam, David W. Wetter, Inbal Nahum-Shani, James M. Rehg
The transformer model achieves a non-response prediction AUC of 0. 77 and is significantly better than classical ML and LSTM-based deep learning models.
1 code implementation • 26 Oct 2021 • Yu-Ying Liu, Alexander Moreno, Maxwell A. Xu, Shuang Li, Jena C. McDaniel, Nancy C. Brady, Agata Rozga, Fuxin Li, Le Song, James M. Rehg
We solve the first challenge by reformulating the estimation problem as an equivalent discrete time-inhomogeneous hidden Markov model.
1 code implementation • NeurIPS 2021 • Meera Hahn, Devendra Chaplot, Shubham Tulsiani, Mustafa Mukadam, James M. Rehg, Abhinav Gupta
Most prior methods for learning navigation policies require access to simulation environments, as they need online policy interaction and rely on ground-truth maps for rewards.
8 code implementations • CVPR 2022 • Kristen Grauman, Andrew Westbury, Eugene Byrne, Zachary Chavis, Antonino Furnari, Rohit Girdhar, Jackson Hamburger, Hao Jiang, Miao Liu, Xingyu Liu, Miguel Martin, Tushar Nagarajan, Ilija Radosavovic, Santhosh Kumar Ramakrishnan, Fiona Ryan, Jayant Sharma, Michael Wray, Mengmeng Xu, Eric Zhongcong Xu, Chen Zhao, Siddhant Bansal, Dhruv Batra, Vincent Cartillier, Sean Crane, Tien Do, Morrie Doulaty, Akshay Erapalli, Christoph Feichtenhofer, Adriano Fragomeni, Qichen Fu, Abrham Gebreselasie, Cristina Gonzalez, James Hillis, Xuhua Huang, Yifei HUANG, Wenqi Jia, Weslie Khoo, Jachym Kolar, Satwik Kottur, Anurag Kumar, Federico Landini, Chao Li, Yanghao Li, Zhenqiang Li, Karttikeya Mangalam, Raghava Modhugu, Jonathan Munro, Tullie Murrell, Takumi Nishiyasu, Will Price, Paola Ruiz Puentes, Merey Ramazanova, Leda Sari, Kiran Somasundaram, Audrey Southerland, Yusuke Sugano, Ruijie Tao, Minh Vo, Yuchen Wang, Xindi Wu, Takuma Yagi, Ziwei Zhao, Yunyi Zhu, Pablo Arbelaez, David Crandall, Dima Damen, Giovanni Maria Farinella, Christian Fuegen, Bernard Ghanem, Vamsi Krishna Ithapu, C. V. Jawahar, Hanbyul Joo, Kris Kitani, Haizhou Li, Richard Newcombe, Aude Oliva, Hyun Soo Park, James M. Rehg, Yoichi Sato, Jianbo Shi, Mike Zheng Shou, Antonio Torralba, Lorenzo Torresani, Mingfei Yan, Jitendra Malik
We introduce Ego4D, a massive-scale egocentric video dataset and benchmark suite.
no code implementations • 20 May 2021 • Miao Liu, Lingni Ma, Kiran Somasundaram, Yin Li, Kristen Grauman, James M. Rehg, Chao Li
Given a video captured from a first person perspective and the environment context of where the video is recorded, can we recognize what the person is doing and identify where the action occurs in the 3D space?
1 code implementation • CVPR 2021 • Chanho Kim, Li Fuxin, Mazen Alotaibi, James M. Rehg
Many approaches model each target in isolation and lack the ability to use all the targets in the scene to jointly update the memory.
3 code implementations • 18 Jan 2021 • Anh Thai, Stefan Stojanov, Zixuan Huang, Isaac Rehg, James M. Rehg
Continual learning has been extensively studied for classification tasks with methods developed to primarily avoid catastrophic forgetting, a phenomenon where earlier learned concepts are forgotten at the expense of more recent samples.
1 code implementation • CVPR 2021 • Stefan Stojanov, Anh Thai, James M. Rehg
It is widely accepted that reasoning about object shape is important for object recognition.
no code implementations • 26 Nov 2020 • Miao Liu, Dexin Yang, Yan Zhang, Zhaopeng Cui, James M. Rehg, Siyu Tang
We introduce a novel task of reconstructing a time series of second-person 3D human body meshes from monocular egocentric videos.
2 code implementations • EMNLP 2020 • Meera Hahn, Jacob Krantz, Dhruv Batra, Devi Parikh, James M. Rehg, Stefan Lee, Peter Anderson
In this paper, we focus on the LED task -- providing a strong baseline model with detailed ablations characterizing both dataset biases and the importance of various modeling choices.
2 code implementations • 14 Jun 2020 • Anh Thai, Stefan Stojanov, Vijay Upadhya, James M. Rehg
This is challenging as it requires a model to learn a representation that can infer both the visible and occluded portions of any object using a limited training set.
no code implementations • 31 May 2020 • Yin Li, Miao Liu, James M. Rehg
Moving beyond the dataset, we propose a novel deep model for joint gaze estimation and action recognition in FPV.
no code implementations • 17 Apr 2020 • Keuntaek Lee, Bogdan Vlahov, Jason Gibson, James M. Rehg, Evangelos A. Theodorou
In this work, we present a method for obtaining an implicit objective function for vision-based navigation.
1 code implementation • CVPR 2021 • Weiyang Liu, Rongmei Lin, Zhen Liu, James M. Rehg, Liam Paull, Li Xiong, Le Song, Adrian Weller
The inductive bias of a neural network is largely determined by the architecture and the training algorithm.
1 code implementation • CVPR 2020 • Eunji Chong, Yongxin Wang, Nataniel Ruiz, James M. Rehg
We address the problem of detecting attention targets in video.
no code implementations • NeurIPS 2020 • Alexander Moreno, Zhenke Wu, Jamie Yap, David Wetter, Cho Lam, Inbal Nahum-Shani, Walter Dempsey, James M. Rehg
Panel count data describes aggregated counts of recurrent events observed at discrete time points.
1 code implementation • NeurIPS 2019 • Weiyang Liu, Zhen Liu, James M. Rehg, Le Song
By generalizing inner product with a bilinear matrix, we propose the neural similarity which serves as a learnable parametric similarity measure for CNNs.
1 code implementation • CVPR 2020 • Rongmei Lin, Weiyang Liu, Zhen Liu, Chen Feng, Zhiding Yu, James M. Rehg, Li Xiong, Le Song
Inspired by the Thomson problem in physics where the distribution of multiple propelling electrons on a unit sphere can be modeled via minimizing some potential energy, hyperspherical energy minimization has demonstrated its potential in regularizing neural networks and improving their generalization power.
no code implementations • 13 May 2019 • Grady Williams, Brian Goldfain, James M. Rehg, Evangelos A. Theodorou
We consider the problem of online adaptation of a neural network designed to represent vehicle dynamics.
no code implementations • 22 Apr 2019 • Meera Hahn, Asim Kadav, James M. Rehg, Hans Peter Graf
Localizing moments in untrimmed videos via language queries is a new and interesting task that requires the ability to accurately ground language into video.
1 code implementation • CVPR 2019 • Shashank Tripathi, Siddhartha Chandra, Amit Agrawal, Ambrish Tyagi, James M. Rehg, Visesh Chari
The synthesizer and target networks are trained in an adversarial manner wherein each network is updated with a goal to outdo the other.
no code implementations • CVPR 2019 • Ching-Hang Chen, Ambrish Tyagi, Amit Agrawal, Dylan Drover, Rohith MV, Stefan Stojanov, James M. Rehg
Additionally, to learn from 2D poses "in the wild", we train an unsupervised 2D domain adapter network to allow for an expansion of 2D data.
Ranked #78 on
3D Human Pose Estimation
on MPI-INF-3DHP
(AUC metric)
no code implementations • 5 Apr 2019 • Miao Liu, Xin Chen, Yun Zhang, Yin Li, James M. Rehg
To this end, we make use of attention modules that learn to highlight regions in the video and aggregate features for recognition.
Ranked #41 on
Action Recognition
on UCF101
no code implementations • 2 Jan 2019 • Meera Hahn, Andrew Silva, James M. Rehg
We describe a novel cross-modal embedding space for actions, named Action2Vec, which combines linguistic cues from class labels with spatio-temporal features derived from video clips.
1 code implementation • CVPR 2019 • Zhaoyang Lv, Frank Dellaert, James M. Rehg, Andreas Geiger
In this paper, we provide a modern synthesis of the classic inverse compositional algorithm for dense image alignment.
no code implementations • 22 Sep 2018 • Meera Hahn, Nataniel Ruiz, Jean-Baptiste Alayrac, Ivan Laptev, James M. Rehg
Automatic generation of textual video descriptions that are time-aligned with video content is a long-standing goal in computer vision.
no code implementations • ECCV 2018 • Chanho Kim, Fuxin Li, James M. Rehg
We also propose novel data augmentation approaches to efficiently train recurrent models that score object tracks on both appearance and motion.
no code implementations • ECCV 2018 • Yin Li, Miao Liu, James M. Rehg
We address the task of jointly determining what a person is doing and where they are looking based on the analysis of video captured by a headworn camera.
no code implementations • CVPR 2018 • Abhijit Kundu, Yin Li, James M. Rehg
Our method produces a compact 3D representation of the scene, which can be readily used for applications like autonomous driving.
Ranked #3 on
Vehicle Pose Estimation
on KITTI Cars Hard
(using extra training data)
1 code implementation • CVPR 2018 • Weiyang Liu, Zhen Liu, Zhiding Yu, Bo Dai, Rongmei Lin, Yisen Wang, James M. Rehg, Le Song
Inner product-based convolution has been a central component of convolutional neural networks (CNNs) and the key to learning visual representations.
1 code implementation • ECCV 2018 • Zhaoyang Lv, Kihwan Kim, Alejandro Troccoli, Deqing Sun, James M. Rehg, Jan Kautz
Estimation of 3D motion in a dynamic scene from a temporal pair of images is a core task in many scene understanding problems.
no code implementations • ICML 2018 • Weiyang Liu, Bo Dai, Xingguo Li, Zhen Liu, James M. Rehg, Le Song
We propose an active teacher model that can actively query the learner (i. e., make the learner take exams) for estimating the learner's status and provably guide the learner to achieve faster convergence.
14 code implementations • 2 Oct 2017 • Nataniel Ruiz, Eunji Chong, James M. Rehg
Estimating the head pose of a person is a crucial problem that has a large amount of applications such as aiding in gaze estimation, modeling attention, fitting 3D models to video and performing face alignment.
Ranked #5 on
Head Pose Estimation
on AFLW
1 code implementation • 15 Aug 2017 • Nataniel Ruiz, James M. Rehg
Face detection is a very important task and a necessary pre-processing step for many applications such as facial landmark detection, pose estimation, sentiment analysis and face recognition.
no code implementations • ICML 2017 • Walter H. Dempsey, Alexander Moreno, Christy K. Scott, Michael L. Dennis, David H. Gustafson, Susan A. Murphy, James M. Rehg
We present a parameter learning method for GLM emissions and survival model fitting, and present promising results on both synthetic data and an mHealth drug use dataset.
4 code implementations • 7 Jul 2017 • Grady Williams, Paul Drews, Brian Goldfain, James M. Rehg, Evangelos A. Theodorou
We present an information theoretic approach to stochastic optimal control problems that can be used to derive general sampling based optimization schemes.
Robotics
2 code implementations • ICML 2017 • Weiyang Liu, Bo Dai, Ahmad Humayun, Charlene Tay, Chen Yu, Linda B. Smith, James M. Rehg, Le Song
Different from traditional machine teaching which views the learners as batch algorithms, we study a new paradigm where the learner uses an iterative algorithm and a teacher can feed examples sequentially and intelligently based on the current performance of the learner.
no code implementations • 28 Jun 2016 • Alexander Moreno, Tameem Adel, Edward Meeds, James M. Rehg, Max Welling
Approximate Bayesian Computation (ABC) is a framework for performing likelihood-free posterior inference for simulation models.
no code implementations • ICCV 2015 • Ahmad Humayun, Fuxin Li, James M. Rehg
We propose a new energy minimization framework incorporating geodesic distances between segments which solves this problem.
no code implementations • NeurIPS 2015 • Yu-Ying Liu, Shuang Li, Fuxin Li, Le Song, James M. Rehg
The Continuous-Time Hidden Markov Model (CT-HMM) is an attractive approach to modeling disease progression due to its ability to describe noisy observations arriving irregularly in time.
no code implementations • ICCV 2015 • Arridhana Ciptadi, James M. Rehg
We address the problem of minimizing human effort in interactive tracking by learning sequence-specific model parameters.
no code implementations • ICCV 2015 • Chanho Kim, Fuxin Li, Arridhana Ciptadi, James M. Rehg
This paper revisits the classical multiple hypotheses tracking (MHT) algorithm in a tracking-by-detection framework.
no code implementations • CVPR 2016 • Yin Li, Manohar Paluri, James M. Rehg, Piotr Dollár
In this work we present a simple yet effective approach for training edge detectors without human supervision.
no code implementations • CVPR 2015 • Jia Xu, Lopamudra Mukherjee, Yin Li, Jamieson Warner, James M. Rehg, Vikas Singh
Motivated by these applications, this paper focuses on the problem of egocentric video summarization.
no code implementations • CVPR 2015 • Yin Li, Zhefan Ye, James M. Rehg
We propose to utilize these mid-level egocentric cues for egocentric action recognition.
no code implementations • CVPR 2015 • Zhengyang Wu, Fuxin Li, Rahul Sukthankar, James M. Rehg
We propose a robust algorithm to generate video segment proposals.
1 code implementation • CVPR 2014 • Yin Li, Xiaodi Hou, Christof Koch, James M. Rehg, Alan L. Yuille
The dataset design bias does not only create the discomforting disconnection between fixations and salient object segmentation, but also misleads the algorithm designing.
no code implementations • CVPR 2014 • Ahmad Humayun, Fuxin Li, James M. Rehg
By precomputing a graph which can be used for parametric min-cuts over different seeds, we speed up the generation of the segment pool.
no code implementations • CVPR 2013 • Alireza Fathi, James M. Rehg
The key to differentiating these actions is the ability to identify how they change the state of objects and materials in the environment.