Search Results for author: Xiaotong Chen

Found 14 papers, 5 papers with code

Tabletop Transparent Scene Reconstruction via Epipolar-Guided Optical Flow with Monocular Depth Completion Prior

no code implementations15 Oct 2023 Xiaotong Chen, Zheming Zhou, Zhuo Deng, Omid Ghasemalizadeh, Min Sun, Cheng-Hao Kuo, Arnie Sen

Reconstructing transparent objects using affordable RGB-D cameras is a persistent challenge in robotic perception due to inconsistent appearances across views in the RGB domain and inaccurate depth readings in each single-view.

3D Reconstruction Depth Completion +3

TransNet: Transparent Object Manipulation Through Category-Level Pose Estimation

no code implementations23 Jul 2023 Huijie Zhang, Anthony Opipari, Xiaotong Chen, Jiyue Zhu, Zeren Yu, Odest Chadwicke Jenkins

TransNet is evaluated in terms of pose estimation accuracy on a large-scale transparent object dataset and compared to a state-of-the-art category-level pose estimation approach.

Depth Completion Object +3

Pair DETR: Contrastive Learning Speeds Up DETR Training

no code implementations29 Oct 2022 Seyed Mehdi Iranmanesh, Xiaotong Chen, Kuo-Chin Lien

In this approach, we detect an object bounding box as a pair of keypoints, the top-left corner and the center, using two decoders.

Contrastive Learning object-detection +3

TransNet: Category-Level Transparent Object Pose Estimation

no code implementations22 Aug 2022 Huijie Zhang, Anthony Opipari, Xiaotong Chen, Jiyue Zhu, Zeren Yu, Odest Chadwicke Jenkins

TransNet is evaluated in terms of pose estimation accuracy on a recent, large-scale transparent object dataset and compared to a state-of-the-art category-level pose estimation approach.

Depth Completion Object +3

TE2Rules: Explaining Tree Ensembles using Rules

2 code implementations29 Jun 2022 G Roshan Lal, Xiaotong Chen, Varun Mithal

Tree Ensemble (TE) models, such as Gradient Boosted Trees, often achieve optimal performance on tabular datasets, yet their lack of transparency poses challenges for comprehending their decision logic.

Binary Classification Decision Making +2

VLMbench: A Compositional Benchmark for Vision-and-Language Manipulation

no code implementations17 Jun 2022 Kaizhi Zheng, Xiaotong Chen, Odest Chadwicke Jenkins, Xin Eric Wang

We hope the new simulator and benchmark will facilitate future research on language-guided robotic manipulation.

Object

ClearPose: Large-scale Transparent Object Dataset and Benchmark

1 code implementation8 Mar 2022 Xiaotong Chen, Huijie Zhang, Zeren Yu, Anthony Opipari, Odest Chadwicke Jenkins

Transparent objects are ubiquitous in household settings and pose distinct challenges for visual sensing and perception systems.

Benchmarking Depth Completion +3

ProgressLabeller: Visual Data Stream Annotation for Training Object-Centric 3D Perception

1 code implementation1 Mar 2022 Xiaotong Chen, Huijie Zhang, Zeren Yu, Stanley Lewis, Odest Chadwicke Jenkins

We demonstrate the effectiveness of ProgressLabeller by rapidly create a dataset of over 1M samples with which we fine-tune a state-of-the-art pose estimation network in order to markedly improve the downstream robotic grasp success rates.

Pose Estimation

PatchTrack: Multiple Object Tracking Using Frame Patches

no code implementations1 Jan 2022 Xiaotong Chen, Seyed Mehdi Iranmanesh, Kuo-Chin Lien

In this paper, we present PatchTrack, a Transformer-based joint-detection-and-tracking system that predicts tracks using patches of the current frame of interest.

Multiple Object Tracking Object

LIT: Light-field Inference of Transparency for Refractive Object Localization

no code implementations2 Oct 2019 Zheming Zhou, Xiaotong Chen, Odest Chadwicke Jenkins

With respect to this ProLIT dataset, we demonstrate that LIT can outperform both state-of-the-art end-to-end pose estimation methods and a generative pose estimator on transparent objects.

Object Object Localization +2

DigiVoice: Voice Biomarker Featurization and Analysis Pipeline

1 code implementation17 Jun 2019 Larry Zhang, Xiaotong Chen, Abbad Vakil, Ali Byott, Reza Hosseini Ghomi

Simultaneously, voice has shown potential for analysis in precision medicine as a biomarker for screening illnesses.

Data Visualization feature selection

GRIP: Generative Robust Inference and Perception for Semantic Robot Manipulation in Adversarial Environments

no code implementations20 Mar 2019 Xiaotong Chen, Rui Chen, Zhiqiang Sui, Zhefan Ye, Yanqi Liu, R. Iris Bahar, Odest Chadwicke Jenkins

In this work, we propose Generative Robust Inference and Perception (GRIP) as a two-stage object detection and pose estimation system that aims to combine relative strengths of discriminative CNNs and generative inference methods to achieve robust estimation.

Object object-detection +3

Cannot find the paper you are looking for? You can Submit a new open access paper.