no code implementations • JEP/TALN/RECITAL 2021 • Xiaoou Wang, Xingyu Liu, Yimei Yue
Cet article décrit la participation de l’équipe Nantalco à la tâche 2 du Défi Fouille de Textes 2021 (DEFT) : évaluation automatique de copies d’après une référence existante.
no code implementations • JEP/TALN/RECITAL 2022 • Iris Eshkol-Taravella, Angèle Barbedette, Xingyu Liu, Valentin-Gabriel Soumah
Ce travail a pour objectif de développer un modèle linguistique pour classifier automatiquement des questions issues de transcriptions d’enregistrements provenant des corpus ESLO2 et ACSYNT en deux catégories “spontané” et “préparé”.
no code implementations • 25 Dec 2024 • Ruiqi Liu, Xingyu Liu, Xiaohao Xu, Yixuan Zhang, Yongxin Ge, Lubin Weng
Group Re-identification (G-ReID) faces greater complexity than individual Re-identification (ReID) due to challenges like mutual occlusion, dynamic member interactions, and evolving group structures.
no code implementations • 7 Dec 2024 • Ruida Zhang, Chengxi Li, Chenyangguang Zhang, Xingyu Liu, Haili Yuan, Yanyan Li, Xiangyang Ji, Gim Hee Lee
To eliminate the reliance on 3D trackers and enhance robustness across diverse environments, we propose a stable object tracking module by leveraging associations from 2D deep trackers within a 3D object fusion strategy.
no code implementations • 2 Dec 2024 • Xingyu Liu, Yingyue Li, Chengxi Li, Gu Wang, Chenyangguang Zhang, Ziqin Huang, Xiangyang Ji
In this report, we provide the technical details of the submitted method GFreeDet, which exploits Gaussian splatting and vision Foundation models for the model-free unseen object Detection track in the BOP 2024 Challenge.
no code implementations • 25 Nov 2024 • Xingyu Liu, Gu Wang, Ruida Zhang, Chenyangguang Zhang, Federico Tombari, Xiangyang Ji
Unseen object pose estimation methods often rely on CAD models or multiple reference views, making the onboarding stage costly.
no code implementations • 9 Oct 2024 • Qianqian Xue, Peng Zhang, Xingyu Liu, Wenjian Wang, Guanglei Zhang
Additionally, we introduce the fluorescence signals as conditional information in the model training to sample a reconstructed image that is highly consistent with the input fluorescence signals from the noisy images.
no code implementations • 1 Jul 2024 • Jiehui Zhou, Linxiao Yang, Xingyu Liu, Xinyue Gu, Liang Sun, Wei Chen
In this paper, we propose CURLS, a novel rule learning method leveraging HTE, which can effectively describe subgroups with significant treatment effects.
1 code implementation • 13 May 2024 • Jian Liu, Wei Sun, Hui Yang, Zhiwen Zeng, Chongpei Liu, Jin Zheng, Xingyu Liu, Hossein Rahmani, Nicu Sebe, Ajmal Mian
Object pose estimation is a fundamental computer vision problem with broad applications in augmented reality and robotics.
no code implementations • 6 May 2024 • Xingyu Liu, Deepak Pathak, Ding Zhao
We investigate the problem of transferring an expert policy from a source robot to multiple different robots.
no code implementations • 5 Apr 2024 • Xingyu Liu, Chenyangguang Zhang, Gu Wang, Ruida Zhang, Xiangyang Ji
In robotic vision, a de-facto paradigm is to learn in simulated environments and then transfer to real-world applications, which poses an essential challenge in bridging the sim-to-real domain gap.
1 code implementation • CVPR 2024 • Ruida Zhang, Chenyangguang Zhang, Yan Di, Fabian Manhardt, Xingyu Liu, Federico Tombari, Xiangyang Ji
Objects that are close in the embedding space are considered similar in geometry.
2 code implementations • 9 Nov 2023 • Meiling Fang, Marco Huber, Julian Fierrez, Raghavendra Ramachandra, Naser Damer, Alhasan Alkhaddour, Maksim Kasantcev, Vasiliy Pryadchenko, Ziyuan Yang, Huijie Huangfu, Yingyu Chen, Yi Zhang, Yuchen Pan, Junjun Jiang, Xianming Liu, Xianyun Sun, Caiyong Wang, Xingyu Liu, Zhaohua Chang, Guangzhe Zhao, Juan Tapia, Lazaro Gonzalez-Soler, Carlos Aravena, Daniel Schulz
This paper presents a summary of the Competition on Face Presentation Attack Detection Based on Privacy-aware Synthetic Training Data (SynFacePAD 2023) held at the 2023 International Joint Conference on Biometrics (IJCB 2023).
no code implementations • 8 Oct 2023 • Zudi Lin, Donglai Wei, Aarush Gupta, Xingyu Liu, Deqing Sun, Hanspeter Pfister
Objects with complex structures pose significant challenges to existing instance segmentation methods that rely on boundary or affinity maps, which are vulnerable to small errors around contacting pixels that cause noticeable connectivity change.
no code implementations • 7 Oct 2023 • Yuqi Xiang, Feitong Chen, Qinsi Wang, Yang Gang, Xiang Zhang, Xinghao Zhu, Xingyu Liu, Lin Shao
In this work, we introduce $\textit{Diff-Transfer}$, a novel framework leveraging differentiable physics simulation to efficiently transfer robotic skills.
no code implementations • 6 Oct 2023 • Patrick Tinsley, Sandip Purnapatra, Mahsa Mitcheff, Aidan Boyd, Colton Crum, Kevin Bowyer, Patrick Flynn, Stephanie Schuckers, Adam Czajka, Meiling Fang, Naser Damer, Xingyu Liu, Caiyong Wang, Xianyun Sun, Zhaohua Chang, Xinyue Li, Guangzhe Zhao, Juan Tapia, Christoph Busch, Carlos Aravena, Daniel Schulz
New elements in this fifth competition include (1) GAN-generated iris images as a category of presentation attack instruments (PAI), and (2) an evaluation of human accuracy at detecting PAI as a reference benchmark.
no code implementations • 2 Oct 2023 • Yuyou Zhang, Yaru Niu, Xingyu Liu, Ding Zhao
Instead of perceiving the hyper-redundancy and flexibility of snake robots as mere challenges, there lies an unexplored potential in leveraging these traits to enhance robustness and generalizability at the control policy level.
no code implementations • 3 Jul 2023 • Xingyu Liu, Juan Chen, Quan Wen
Traditional convolutional neural networks are limited to handling Euclidean space data, overlooking the vast realm of real-life scenarios represented as graph data, including transportation networks, social networks, and reference networks.
no code implementations • 11 Mar 2023 • Xingyu Liu, Alex Leonardi, Lu Yu, Chris Gilmer-Hill, Matthew Leavitt, Jonathan Frankle
We find that augmenting future runs with KD from previous runs dramatically reduces the time necessary to train these models, even taking into account the overhead of KD.
no code implementations • ICCV 2023 • Qichen Fu, Xingyu Liu, ran Xu, Juan Carlos Niebles, Kris M. Kitani
Accurately estimating 3D hand pose is crucial for understanding how humans interact with the world.
no code implementations • ICCV 2023 • Xingyu Liu, Sanping Zhou, Le Wang, Gang Hua
Learning discriminative features from very few labeled samples to identify novel classes has received increasing attention in skeleton-based action recognition.
Action Recognition Few-Shot Skeleton-Based Action Recognition +1
no code implementations • 8 Dec 2022 • Xingyu Liu, Deepak Pathak, Kris M. Kitani
The ability to learn from human demonstration endows robots with the ability to automate various tasks.
1 code implementation • 17 Jul 2022 • Xingyu Liu, Gu Wang, Yi Li, Xiangyang Ji
While category-level 9DoF object pose estimation has emerged recently, previous correspondence-based or direct regression methods are both limited in accuracy due to the huge intra-category variances in object shape and color, etc.
1 code implementation • 17 Jul 2022 • Yansong Tang, Xingyu Liu, Xumin Yu, Danyang Zhang, Jiwen Lu, Jie zhou
Different from the conventional adversarial learning-based approaches for UDA, we utilize a self-supervision scheme to reduce the domain shift between two skeleton-based action datasets.
1 code implementation • 19 Mar 2022 • Gu Wang, Fabian Manhardt, Xingyu Liu, Xiangyang Ji, Federico Tombari
6D object pose estimation is a fundamental yet challenging problem in computer vision.
no code implementations • 19 Feb 2022 • Xuefeng Liang, Longshan Yao, Xingyu Liu, Ying Zhou
Instead, we propose a Tripartite solution to partition training data more precisely into three subsets: hard, noisy, and clean.
1 code implementation • 10 Feb 2022 • Xingyu Liu, Deepak Pathak, Kris M. Kitani
We interpolate between the source robot and the target robot by finding a continuous evolutionary change of robot parameters.
no code implementations • 7 Nov 2021 • Xingyu Liu, Kris M. Kitani
Manipulating articulated objects requires multiple robot arms in general.
1 code implementation • CVPR 2022 • Qichen Fu, Xingyu Liu, Kris M. Kitani
While our voting function is able to improve the bounding box of the active object, one round of voting is typically not enough to accurately localize the active object.
8 code implementations • CVPR 2022 • Kristen Grauman, Andrew Westbury, Eugene Byrne, Zachary Chavis, Antonino Furnari, Rohit Girdhar, Jackson Hamburger, Hao Jiang, Miao Liu, Xingyu Liu, Miguel Martin, Tushar Nagarajan, Ilija Radosavovic, Santhosh Kumar Ramakrishnan, Fiona Ryan, Jayant Sharma, Michael Wray, Mengmeng Xu, Eric Zhongcong Xu, Chen Zhao, Siddhant Bansal, Dhruv Batra, Vincent Cartillier, Sean Crane, Tien Do, Morrie Doulaty, Akshay Erapalli, Christoph Feichtenhofer, Adriano Fragomeni, Qichen Fu, Abrham Gebreselasie, Cristina Gonzalez, James Hillis, Xuhua Huang, Yifei HUANG, Wenqi Jia, Weslie Khoo, Jachym Kolar, Satwik Kottur, Anurag Kumar, Federico Landini, Chao Li, Yanghao Li, Zhenqiang Li, Karttikeya Mangalam, Raghava Modhugu, Jonathan Munro, Tullie Murrell, Takumi Nishiyasu, Will Price, Paola Ruiz Puentes, Merey Ramazanova, Leda Sari, Kiran Somasundaram, Audrey Southerland, Yusuke Sugano, Ruijie Tao, Minh Vo, Yuchen Wang, Xindi Wu, Takuma Yagi, Ziwei Zhao, Yunyi Zhu, Pablo Arbelaez, David Crandall, Dima Damen, Giovanni Maria Farinella, Christian Fuegen, Bernard Ghanem, Vamsi Krishna Ithapu, C. V. Jawahar, Hanbyul Joo, Kris Kitani, Haizhou Li, Richard Newcombe, Aude Oliva, Hyun Soo Park, James M. Rehg, Yoichi Sato, Jianbo Shi, Mike Zheng Shou, Antonio Torralba, Lorenzo Torresani, Mingfei Yan, Jitendra Malik
We introduce Ego4D, a massive-scale egocentric video dataset and benchmark suite.
no code implementations • ICCV 2021 • Xingyu Liu, Shun Iwase, Kris M. Kitani
We present a large-scale stereo RGB image object pose estimation dataset named the $\textbf{StereOBJ-1M}$ dataset.
no code implementations • 21 Sep 2021 • Xingyu Liu, Shun Iwase, Kris M. Kitani
To address this problem, we propose a novel continuous representation called Keypoint Distance Field (KDF) for projected 2D keypoint locations.
1 code implementation • ICCV 2021 • Shun Iwase, Xingyu Liu, Rawal Khirodkar, Rio Yokota, Kris M. Kitani
Furthermore, we utilize differentiable Levenberg-Marquardt (LM) optimization to refine a pose fast and accurately by minimizing the feature-metric error between the input and rendered image representations without the need of zooming in.
Ranked #5 on 6D Pose Estimation using RGB on LineMOD
no code implementations • 16 Nov 2020 • Yuan Chang, Chao Yan, Xingyu Liu, Xiangke Wang, Han Zhou, Xiaojia Xiang, Dengqing Tang
This paper presents a time-efficient scheme for Mars exploration by the cooperation of multiple drones and a rover.
1 code implementation • Medical Image Computing and Computer Assisted Intervention 2020 • Donglai Wei, Zudi Lin, Daniel Franco-Barranco, Nils Wendt, Xingyu Liu, Wenjie Yin, Xin Huang, Aarush Gupta, Won-Dong Jang, Xueying Wang, Ignacio Arganda-Carreras, Jeff Lichtman, Hanspeter Pfister
On MitoEM, we find existing instance segmentation methods often fail to correctly segment mitochondria with complex shapes or close contacts with other instances.
Ranked #2 on 3D Instance Segmentation on MitoEM (AP75-R-Test metric)
1 code implementation • CVPR 2020 • Xingyu Liu, Rico Jonschkowski, Anelia Angelova, Kurt Konolige
We address two problems: first, we establish an easy method for capturing and labeling 3D keypoints on desktop objects with an RGB camera; and second, we develop a deep neural network, called $KeyPose$, that learns to accurately predict object poses using 3D keypoints, from stereo input, and works even for transparent objects.
2 code implementations • ICCV 2019 • Xingyu Liu, Mengyuan Yan, Jeannette Bohg
Understanding dynamic 3D environment is crucial for robotic agents and many other applications.
2 code implementations • CVPR 2019 • Xingyu Liu, Joon-Young Lee, Hailin Jin
In particular, it can effectively learn representations for videos by mixing appearance and long-range motion with an RGB-only input.
10 code implementations • CVPR 2019 • Xingyu Liu, Charles R. Qi, Leonidas J. Guibas
In this work, we propose a novel deep neural network named $FlowNet3D$ that learns scene flow from point clouds in an end-to-end fashion.
1 code implementation • ICLR 2018 • Xingyu Liu, Jeff Pool, Song Han, William J. Dally
First, we move the ReLU operation into the Winograd domain to increase the sparsity of the transformed activations.
no code implementations • 24 May 2017 • Huizi Mao, Song Han, Jeff Pool, Wenshuo Li, Xingyu Liu, Yu Wang, William J. Dally
Since memory reference is more than two orders of magnitude more expensive than arithmetic operations, the regularity of sparse structure leads to more efficient hardware design.
4 code implementations • 4 Feb 2016 • Song Han, Xingyu Liu, Huizi Mao, Jing Pu, Ardavan Pedram, Mark A. Horowitz, William J. Dally
EIE has a processing power of 102GOPS/s working directly on a compressed network, corresponding to 3TOPS/s on an uncompressed network, and processes FC layers of AlexNet at 1. 88x10^4 frames/sec with a power dissipation of only 600mW.