no code implementations • 17 May 2023 • Yunzhe Zhang, Yao Lu, Lei Xu, Kunlin Yang, Hui Tang, Shuyuan Ye, Qi Xuan
This paper investigates the differences in data organization between contrastive and supervised learning methods, focusing on the concept of locally dense clusters.
no code implementations • 4 Apr 2023 • Yao Lu, Pengyuan Zhou, Yong Liao, Haiyong Xie
Urban anomaly predictions, such as traffic accident prediction and crime prediction, are of vital importance to smart city security and maintenance.
1 code implementation • 14 Mar 2023 • Hui Tang, Yao Lu, Qi Xuan
Our SR-init method is inspired by the discovery that the accuracy drop due to stochastic re-initialization of layer parameters differs in various layers.
no code implementations • 2 Mar 2023 • Austin Stone, Ted Xiao, Yao Lu, Keerthana Gopalakrishnan, Kuang-Huei Lee, Quan Vuong, Paul Wohlhart, Brianna Zitkovich, Fei Xia, Chelsea Finn, Karol Hausman
This brings up a notably difficult challenge for robots: while robot learning approaches allow robots to learn many different behaviors from first-hand experience, it is impractical for robots to have first-hand experiences that span all of this semantic information.
no code implementations • 1 Mar 2023 • Wenlong Huang, Fei Xia, Dhruv Shah, Danny Driess, Andy Zeng, Yao Lu, Pete Florence, Igor Mordatch, Sergey Levine, Karol Hausman, Brian Ichter
Recent progress in large language models (LLMs) has demonstrated the ability to learn and leverage Internet-scale knowledge through pre-training with autoregressive models.
1 code implementation • 21 Feb 2023 • Langtian Qin, Hancheng Lu, Yao Lu, Chenwu Zhang, Feng Wu
To address single-base station (BS) transmission limitation and serious edge effect in traditional cellular-based edge service caching networks, in this paper, we proposed a novel user-centric edge service caching framework where each user is jointly provided with edge caching and wireless transmission services by a specific BS cluster instead of a single BS.
1 code implementation • 13 Dec 2022 • Anthony Brohan, Noah Brown, Justice Carbajal, Yevgen Chebotar, Joseph Dabis, Chelsea Finn, Keerthana Gopalakrishnan, Karol Hausman, Alex Herzog, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Tomas Jackson, Sally Jesmonth, Nikhil J Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Isabel Leal, Kuang-Huei Lee, Sergey Levine, Yao Lu, Utsav Malla, Deeksha Manjunath, Igor Mordatch, Ofir Nachum, Carolina Parada, Jodilyn Peralta, Emily Perez, Karl Pertsch, Jornell Quiambao, Kanishka Rao, Michael Ryoo, Grecia Salazar, Pannag Sanketi, Kevin Sayed, Jaspiar Singh, Sumedh Sontakke, Austin Stone, Clayton Tan, Huong Tran, Vincent Vanhoucke, Steve Vega, Quan Vuong, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Tianhe Yu, Brianna Zitkovich
By transferring knowledge from large, diverse, task-agnostic datasets, modern machine learning models can solve specific downstream tasks either zero-shot or with small task-specific datasets to a high level of performance.
no code implementations • 16 Nov 2022 • Yao Lu, Yanan Liu
In this paper, we propose a method to jointly determine the status of hand-object interaction.
1 code implementation • CVPR 2023 • Michael S. Ryoo, Keerthana Gopalakrishnan, Kumara Kahatapitiya, Ted Xiao, Kanishka Rao, Austin Stone, Yao Lu, Julian Ibarz, Anurag Arnab
The model's memory module ensures that a new observation will only be processed with the contents of the memory (and not the entire history), meaning that it can efficiently process long sequences with a bounded computational cost at each step.
Ranked #1 on
Action Detection
on Charades
no code implementations • 19 Oct 2022 • Thomas Lew, Sumeet Singh, Mario Prats, Jeffrey Bingham, Jonathan Weisz, Benjie Holson, Xiaohan Zhang, Vikas Sindhwani, Yao Lu, Fei Xia, Peng Xu, Tingnan Zhang, Jie Tan, Montserrat Gonzalez
This problem is challenging, as it requires planning wiping actions while reasoning over uncertain latent dynamics of crumbs and spills captured via high-dimensional visual observations.
no code implementations • 15 Oct 2022 • Kuang-Huei Lee, Ted Xiao, Adrian Li, Paul Wohlhart, Ian Fischer, Yao Lu
The predictive information, the mutual information between the past and future, has been shown to be a useful representation learning auxiliary loss for training reinforcement learning agents, as the ability to model what will happen next is critical to success on many control tasks.
no code implementations • 8 Oct 2022 • Yao Lu, Jide Zhang, Su Zheng, Zhen Li, Lingli Wang
In this paper, two approximate 3*3 multipliers are proposed and the synthesis results of the ASAP-7nm process library justify that they can reduce the area by 31. 38% and 36. 17%, and the power consumption by 36. 73% and 35. 66% compared with the exact multiplier, respectively.
no code implementations • 13 Sep 2022 • Anna Gottardi, Osman Ipek, Giuseppe Castellucci, Shui Hu, Lavina Vaz, Yao Lu, Anju Khatri, Anjali Chadha, Desheng Zhang, Sattvik Sahai, Prerna Dwivedi, Hangjie Shi, Lucy Hu, Andy Huang, Luke Dai, Bofei Yang, Varun Somani, Pankaj Rajan, Ron Rezac, Michael Johnston, Savanna Stiff, Leslie Ball, David Carmel, Yang Liu, Dilek Hakkani-Tur, Oleg Rokhlenko, Kate Bland, Eugene Agichtein, Reza Ghanadan, Yoelle Maarek
Since its inception in 2016, the Alexa Prize program has enabled hundreds of university students to explore and compete to develop conversational agents through the SocialBot Grand Challenge.
no code implementations • 16 Aug 2022 • Jiayan Gu, Ashiq Anjum, Yan Wu, Lu Liu, John Panneerselvam, Yao Lu, Bo Yuan
The experimental results show that the proposed least-used key selection method improves the service retrieval efficiency significantly compared with the designated key selection method in the case of the unequal appearing probability of parameters in service retrieval requests under three indexing models.
no code implementations • ICIAP 2022 • Mang Ning, Xiaoliang Ma, Yao Lu, Simone Calderara, Rita Cucchiara
In this paper, we introduce SeeFar to achieve vehicle speed estimation and traffic flow analysis based on YOLOv5 and DeepSORT from a moving drone.
no code implementations • 10 May 2022 • Yongji Wu, Matthew Lentz, Danyang Zhuo, Yao Lu
With the advent of ubiquitous deployment of smart devices and the Internet of Things, data sources for machine learning inference have increasingly moved to the edge of the network.
no code implementations • 5 Apr 2022 • Ikechukwu Uchendu, Ted Xiao, Yao Lu, Banghua Zhu, Mengyuan Yan, Joséphine Simon, Matthew Bennice, Chuyuan Fu, Cong Ma, Jiantao Jiao, Sergey Levine, Karol Hausman
In addition, we provide an upper bound on the sample complexity of JSRL and show that with the help of a guide-policy, one can improve the sample complexity for non-optimism exploration methods from exponential in horizon to polynomial.
2 code implementations • 4 Apr 2022 • Michael Ahn, Anthony Brohan, Noah Brown, Yevgen Chebotar, Omar Cortes, Byron David, Chelsea Finn, Chuyuan Fu, Keerthana Gopalakrishnan, Karol Hausman, Alex Herzog, Daniel Ho, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Eric Jang, Rosario Jauregui Ruano, Kyle Jeffrey, Sally Jesmonth, Nikhil J Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Kuang-Huei Lee, Sergey Levine, Yao Lu, Linda Luu, Carolina Parada, Peter Pastor, Jornell Quiambao, Kanishka Rao, Jarek Rettinghouse, Diego Reyes, Pierre Sermanet, Nicolas Sievers, Clayton Tan, Alexander Toshev, Vincent Vanhoucke, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Mengyuan Yan, Andy Zeng
We show how low-level skills can be combined with large language models so that the language model provides high-level knowledge about the procedures for performing complex and temporally-extended instructions, while value functions associated with these skills provide the grounding necessary to connect this knowledge to a particular physical environment.
no code implementations • 2 Feb 2022 • Yanan Liu, Laurie Bose, Yao Lu, Piotr Dudek, Walterio Mayol-Cuevas
This work presents a method to implement fully convolutional neural networks (FCNs) on Pixel Processor Array (PPA) sensors, and demonstrates coarse segmentation and object localisation tasks.
2 code implementations • 20 Jan 2022 • Su Zheng, Zhen Li, Yao Lu, Jingbo Gao, Jide Zhang, Lingli Wang
We propose an optimization method for the automatic design of approximate multipliers, which minimizes the average error according to the operand distributions.
1 code implementation • 2 Jan 2022 • Shunzhou Wang, Tianfei Zhou, Yao Lu, Huijun Di
DPT consists of two branches, with each associated with a Transformer for learning from an original or gradient image sequence.
no code implementations • NeurIPS 2021 • Ligeng Zhu, Hongzhou Lin, Yao Lu, Yujun Lin, Song Han
Federated Learning is an emerging direction in distributed machine learning that en-ables jointly training a model without sharing the data.
1 code implementation • 24 Nov 2021 • Yao Lu, Wen Yang, Yunzhe Zhang, Zuohui Chen, Jinyin Chen, Qi Xuan, Zhen Wang, Xiaoniu Yang
Specifically, we model the process of class separation of intermediate representations in pre-trained DNNs as the evolution of communities in dynamic graphs.
1 code implementation • 22 Nov 2021 • Zuohui Chen, Yao Lu, Jinxuan Hu, Wen Yang, Qi Xuan, Zhen Wang, Xiaoniu Yang
Understanding the black-box representations in Deep Neural Networks (DNN) is an essential problem in deep learning.
no code implementations • ICLR 2022 • Dhruv Shah, Peng Xu, Yao Lu, Ted Xiao, Alexander Toshev, Sergey Levine, Brian Ichter
Hierarchical reinforcement learning aims to enable this by providing a bank of low-level skills as action abstractions.
Hierarchical Reinforcement Learning
reinforcement-learning
+1
no code implementations • 28 Oct 2021 • Zhuangzhi Chen, Jingyang Xiang, Yao Lu, Qi Xuan, Xiaoniu Yang
In this paper, we study the graph structure of the neural network, and propose regular graph based pruning (RGP) to perform a one-shot neural network pruning.
no code implementations • 29 Sep 2021 • Yao Lu, Walterio W. Mayol-Cuevas
In this paper, we concern with the problem of how to automatically extract the steps that compose real-life hand activities.
no code implementations • 29 Sep 2021 • Yao Lu, Walterio W. Mayol-Cuevas
To show the ability to preserve the semantic information of our method, we also report the performance of grasp type classification on GUN-71 dataset and outperforms the benchmark by only using the predicted 3-d hand pose.
no code implementations • 29 Sep 2021 • Yao Lu, Walterio W. Mayol-Cuevas
We compare our method with the most recent work from Shan et al. \cite{Shan20} on selected images from EPIC-KITCHENS \cite{damen2018scaling} dataset and achieve $89\%$ accuracy on HOI (hand-object interaction) detection which is comparative to Shan's ($92\%$).
no code implementations • 22 Jul 2021 • Wanqing Xie, Lizhong Liang, Yao Lu, Hui Luo, Xiaofeng Liu
The superior performance of our system shows the validity of combining facial video recording with the SDS score for more accurate self-diagnose.
1 code implementation • 20 Jul 2021 • Li Shen, Yao Lu, Hao Chen, Hao Wei, Donghai Xie, Jiabao Yue, Rui Chen, Shouye Lv, Bitao Jiang
This paper therefore introduces S2Looking, a building-change-detection dataset that contains large-scale side-looking satellite images captured at various off-nadir angles.
no code implementations • 25 Jun 2021 • Wanqing Xie, Lizhong Liang, Yao Lu, Chen Wang, Jihong Shen, Hui Luo, Xiaofeng Liu
To automatically interpret depression from the SDS evaluation and the paired video, we propose an end-to-end hierarchical framework for the long-term variable-length video, which is also conditioned on the questionnaire results and the answering time.
no code implementations • ICML Workshop AML 2021 • Zuohui Chen, Renxuan Wang, Yao Lu, Jingyang Xiang, Qi Xuan
Experiments on CIFAR10 and SVHN show that the FLOPs and size of our generated model are only 24. 46\% and 4. 86\% of the original model.
no code implementations • 8 Jun 2021 • Jiayan Gu, Yan Wu, Ashiq Anjum, John Panneerselvam, Yao Lu, Bo Yuan
With the development of Edge Computing and Artificial Intelligence (AI) technologies, edge devices are witnessed to generate data at unprecedented volume.
1 code implementation • 3 May 2021 • Jie Hu, Liujuan Cao, Yao Lu, Shengchuan Zhang, Yan Wang, Ke Li, Feiyue Huang, Ling Shao, Rongrong Ji
However, such an upgrade is not applicable to instance segmentation, due to its significantly higher output dimensions compared to object detection.
Ranked #19 on
Instance Segmentation
on COCO test-dev
1 code implementation • ACL 2022 • Yao Lu, Max Bartolo, Alastair Moore, Sebastian Riedel, Pontus Stenetorp
When primed with only a handful of training samples, very large, pretrained language models such as GPT-3 have shown competitive results when compared to fully-supervised, fine-tuned, large, pretrained language models.
no code implementations • 15 Apr 2021 • Yevgen Chebotar, Karol Hausman, Yao Lu, Ted Xiao, Dmitry Kalashnikov, Jake Varley, Alex Irpan, Benjamin Eysenbach, Ryan Julian, Chelsea Finn, Sergey Levine
We consider the problem of learning useful robotic skills from previously collected offline data without access to manually specified rewards or additional online exploration, a setting that is becoming increasingly important for scaling robot learning by reusing past robotic data.
no code implementations • 6 Apr 2021 • Pramod Chunduri, Jaeho Bang, Yao Lu, Joy Arulraj
ZEUS trains a reinforcement learning agent that learns to adaptively modify the input video segments that are subsequently sent to an action classification network.
2 code implementations • ICCV 2021 • Shihao Jiang, Dylan Campbell, Yao Lu, Hongdong Li, Richard Hartley
We demonstrate that the optical flow estimates in the occluded regions can be significantly improved without damaging the performance in non-occluded regions.
Ranked #3 on
Optical Flow Estimation
on Sintel-clean
1 code implementation • CVPR 2021 • Shihao Jiang, Yao Lu, Hongdong Li, Richard Hartley
In this paper, we show that the dense correlation volume representation is redundant and accurate flow estimation can be achieved with only a fraction of elements in it.
Ranked #8 on
Optical Flow Estimation
on KITTI 2015 (train)
no code implementations • 26 Mar 2021 • Iretiayo Akinola, Anelia Angelova, Yao Lu, Yevgen Chebotar, Dmitry Kalashnikov, Jacob Varley, Julian Ibarz, Michael S. Ryoo
We propose a vision-based architecture search algorithm for robot manipulation learning, which discovers interactions between low dimension action inputs and high dimensional visual inputs.
1 code implementation • EMNLP 2020 • Yao Lu, Yue Dong, Laurent Charlin
Multi-document summarization is a challenging task for which there exists little large-scale datasets.
no code implementations • 24 Jun 2020 • Kangfu Mei, Yao Lu, Qiaosi Yi, Hao-Yu Wu, Juncheng Li, Rui Huang
Perceptual learning approaches like perceptual loss are empirically powerful for such tasks but they usually rely on the pre-trained classification network to provide features, which are not necessarily optimal in terms of visual perception of image transformation.
no code implementations • 22 Jun 2020 • Yao Lu, Stephen Gould, Thalaiyasingam Ajanthan
The problem of vanishing and exploding gradients has been a long-standing obstacle that hinders the effective training of neural networks.
no code implementations • CVPR 2021 • Yao Lu, Sören Pirk, Jan Dlabal, Anthony Brohan, Ankita Pasad, Zhao Chen, Vincent Casser, Anelia Angelova, Ariel Gordon
Many computer vision tasks address the problem of scene understanding and are naturally interrelated e. g. object classification, detection, scene segmentation, depth estimation, etc.
1 code implementation • Springer 2020 • Yuchong Gu, Zitao Zen, Haibin Chen, Jun Wei, Yaqin Zhang, Binghui Chen, Yingqin Li, Yujuan Qin, Qing Xie, Zhuoren Jiang, Yao Lu
Super-resolution (SR) in medical imaging is an emerging application in medical imaging due to the needs of high quality images acquired with limited radiation dose, such as low dose Computer Tomography (CT), low field magnetic resonance imaging (MRI).
2 code implementations • ACL 2020 • Raphael Schumann, Lili Mou, Yao Lu, Olga Vechtomova, Katja Markert
Automatic sentence summarization produces a shorter version of a sentence, while preserving its most important information.
1 code implementation • WS 2019 • Raphael Tang, Yao Lu, Jimmy Lin
Knowledge distillation can effectively transfer knowledge from BERT, a deep language representation model, to traditional, shallow word embedding-based neural networks, helping them approach or exceed the quality of other heavyweight language representation models.
no code implementations • 1 Nov 2019 • Lingling Yang, Leanne Lai Hang Chan, Yao Lu
Here, we proposed a CNNLSTM based neural network architecture termed EEG_CNNLSTMNet for the classification of EEG signals in response to grating stimuli with different spatial frequencies.
no code implementations • 1 Nov 2019 • Lingling Yang, Dongyang Li, Yao Lu
In this paper, we propose a new T+2 churn customer prediction model, in which the churn customers in two months are recognized and the one-month window T+1 is reserved to carry out churn management strategies.
no code implementations • 25 Sep 2019 • Ligeng Zhu, Yao Lu, Yujun Lin, Song Han
Traditional synchronous distributed training is performed inside a cluster, since it requires high bandwidth and low latency network (e. g. 25Gb Ethernet or Infini-band).
1 code implementation • 30 Aug 2019 • Zhuoren Jiang, Jian Wang, Lujun Zhao, Changlong Sun, Yao Lu, Xiaozhong Liu
Aspect category detection is an essential task for sentiment analysis and opinion mining.
no code implementations • 27 Aug 2019 • Yuqi Huo, Xiaoli Xu, Yao Lu, Yulei Niu, Zhiwu Lu, Ji-Rong Wen
In addition to motion vectors, we also provide a temporal fusion method to explicitly induce the temporal context.
no code implementations • 12 May 2019 • Danlu Chen, Xu-Yao Zhang, Wei zhang, Yao Lu, Xiuli Li, Tao Mei
Taking scene text detection as the application, where no suitable ensemble learning strategy exists, PEL can significantly improve the performance, compared to either individual state-of-the-art models, or the fusion of multiple models by non-maximum suppression.
4 code implementations • 28 Mar 2019 • Raphael Tang, Yao Lu, Linqing Liu, Lili Mou, Olga Vechtomova, Jimmy Lin
In the natural language processing literature, neural networks are becoming increasingly deeper and complex.
Ranked #56 on
Sentiment Analysis
on SST-2 Binary classification
1 code implementation • 31 Dec 2018 • Zhuoren Jiang, Yue Yin, Liangcai Gao, Yao Lu, Xiaozhong Liu
While the volume of scholarly publications has increased at a frenetic pace, accessing and consuming the useful candidate papers, in very large digital libraries, is becoming an essential and challenging task for scholars.
no code implementations • 26 Jun 2018 • Li Wang, Weiyuan Shao, Yao Lu, Hao Ye, Jian Pu, Yingbin Zheng
Crowd counting is one of the core tasks in various surveillance applications.
no code implementations • 16 Apr 2018 • Yao Lu, Mehrtash Harandi, Richard Hartley, Razvan Pascanu
Advanced optimization algorithms such as Newton method and AdaGrad benefit from second order derivative or second order statistics to achieve better descent directions and faster convergence rates.
no code implementations • 14 Apr 2018 • Tian Guo, Tao Lin, Yao Lu
In this paper, we propose an interpretable LSTM recurrent neural network, i. e., multi-variable LSTM for time series with exogenous variables.
no code implementations • 13 Apr 2018 • Haonan Qiu, Yingbin Zheng, Hao Ye, Yao Lu, Feng Wang, Liang He
The performances of existing action localization approaches remain unsatisfactory in precisely determining the beginning and the end of an action.
no code implementations • 20 Feb 2018 • Yao Lu, Jack Valmadre, Heng Wang, Juho Kannala, Mehrtash Harandi, Philip H. S. Torr
State-of-the-art neural network models estimate large displacement optical flow in multi-resolution and use warping to propagate the estimation between two resolutions.
1 code implementation • 26 Nov 2017 • Linqing Liu, Yao Lu, Min Yang, Qiang Qu, Jia Zhu, Hongyan Li
In this paper, we propose an adversarial process for abstractive text summarization, in which we simultaneously train a generative model G and a discriminative model D. In particular, we build the generator G as an agent of reinforcement learning, which takes the raw text as input and predicts the abstractive summarization.
Ranked #5 on
Text Summarization
on CNN / Daily Mail (Anonymized)
no code implementations • 16 May 2017 • Yao Lu, Zhirong Yang, Juho Kannala, Samuel Kaski
A key to the problem is learning a representation of relations.
no code implementations • 1 Feb 2017 • Li Wang, Yao Lu, Hong Wang, Yingbin Zheng, Hao Ye, xiangyang xue
We perform fast vehicle detection from traffic surveillance cameras.
1 code implementation • 7 Sep 2016 • Yao Lu, Jukka Corander, Zhirong Yang
To solve this problem, we introduce a fast normalization method and normalize the similarity matrix to be doubly stochastic such that all the data points have equal total similarities.
no code implementations • CVPR 2016 • Yao Lu, Xue Bai, Linda Shapiro, Jue Wang
Interactive video segmentation systems aim at producing sub-pixel-level object boundaries for visual effect applications.
Interactive Video Object Segmentation
Semantic Segmentation
+2
no code implementations • HLT 2016 • Linqing Liu, Yao Lu, Ye Luo, Renxian Zhang, Laurent Itti, Jianwei Lu
Spammer detection on social network is a challenging problem.
no code implementations • ICCV 2015 • Huijun Di, Qingxuan Shi, Feng Lv, Ming Qin, Yao Lu
Our goal is to estimate contour flow (the contour pairs with consistent point correspondence) from inconsistent contours extracted independently in two video frames.
1 code implementation • 2 Jun 2015 • Yao Lu
The outputs of a trained neural network contain much richer information than just an one-hot classifier.
no code implementations • 19 Jan 2015 • Jinwu Liu, Yao Lu, Tianfei Zhou
Multiple Instance Learning (MIL) recently provides an appealing way to alleviate the drifting problem in visual tracking.
no code implementations • 28 Oct 2014 • Tianfei Zhou, Yao Lu, Feng Lv, Huijun Di, Qingjie Zhao, Jian Zhang
Stochastic sampling based trackers have shown good performance for abrupt motion tracking so that they have gained popularity in recent years.
1 code implementation • 3 Jul 2012 • Yao Lu, Kai-Zhu Huang, Cheng-Lin Liu
In particular, with high accuracy, our algorithm takes only a few seconds (in a PC) to match two graphs of 1, 000 nodes.
no code implementations • IEEE International Conference on Computer Vision 2011 • Xiangyang Xue, Wei zhang, Jie Zhang, Bin Wu, Jianping Fan, Yao Lu
The cross-level label coherence en-codes the consistency between the labels at the image leveland the labels at the region level.