Search Results for author: Qi Wu

Found 172 papers, 81 papers with code

Vision-and-Language Navigation Today and Tomorrow: A Survey in the Era of Foundation Models

no code implementations9 Jul 2024 Yue Zhang, Ziqiao Ma, Jialu Li, Yanyuan Qiao, Zun Wang, Joyce Chai, Qi Wu, Mohit Bansal, Parisa Kordjamshidi

Vision-and-Language Navigation (VLN) has gained increasing attention over recent years and many approaches have emerged to advance their development.

Vision and Language Navigation

HumanPlus: Humanoid Shadowing and Imitation from Humans

no code implementations15 Jun 2024 Zipeng Fu, Qingqing Zhao, Qi Wu, Gordon Wetzstein, Chelsea Finn

This policy transfers to the real world and allows humanoid robots to follow human body and hand motion in real time using only a RGB camera, i. e. shadowing.

SMART: Scene-motion-aware human action recognition framework for mental disorder group

1 code implementation7 Jun 2024 Zengyuan Lai, Jiarui Yang, Songpengcheng Xia, Qi Wu, Zhen Sun, Wenxian Yu, Ling Pei

In this paper, we innovatively propose to build a vision-based HAR dataset including abnormal actions often occurring in the mental disorder group and then introduce a novel Scene-Motion-aware Action Recognition Technology framework, named SMART, consisting of two technical modules.

Action Recognition Temporal Action Localization

Augmented Commonsense Knowledge for Remote Object Grounding

1 code implementation3 Jun 2024 Bahram Mohammadi, Yicong Hong, Yuankai Qi, Qi Wu, Shirui Pan, Javen Qinfeng Shi

To address enhancing representation, we propose an augmented commonsense knowledge model (ACK) to leverage commonsense information as a spatio-temporal knowledge graph for improving agent navigation.

Decision Making Object +1

Streaming Video Diffusion: Online Video Editing with Diffusion Models

1 code implementation30 May 2024 Feng Chen, Zhen Yang, Bohan Zhuang, Qi Wu

We present a novel task called online video editing, which is designed to edit \textbf{streaming} frames while maintaining temporal consistency.

Video Editing

Risk-Neutral Generative Networks

no code implementations28 May 2024 Zhonghao Xian, Xing Yan, Cheuk Hang Leung, Qi Wu

It is also generative in that the functional form of the stochastic curve, although parameterized by neural nets, is an explicit and deterministic function of the standard normal.


MotionLLM: Multimodal Motion-Language Learning with Large Language Models

no code implementations27 May 2024 Qi Wu, Yubo Zhao, Yifan Wang, Yu-Wing Tai, Chi-Keung Tang

Recent advancements in Multimodal Large Language Models (MM-LLMs) have demonstrated promising potential in terms of generalization and robustness when applied to different modalities.

Language Modelling Motion Captioning

BotDGT: Dynamicity-aware Social Bot Detection with Dynamic Graph Transformers

1 code implementation23 Apr 2024 Buyun He, Yingguang Yang, Qi Wu, Hao liu, Renyu Yang, Hao Peng, Xiang Wang, Yong Liao, Pengyuan Zhou

To tackle these challenges, we propose BotDGT, a novel framework that not only considers the topological structure, but also effectively incorporates dynamic nature of social network.


VL-Mamba: Exploring State Space Models for Multimodal Learning

no code implementations20 Mar 2024 Yanyuan Qiao, Zheng Yu, Longteng Guo, Sihan Chen, Zijia Zhao, Mingzhen Sun, Qi Wu, Jing Liu

The extensive experiments on diverse multimodal benchmarks with competitive performance show the effectiveness of our proposed VL-Mamba and demonstrate the great potential of applying state space models for multimodal learning tasks.

Language Modelling Large Language Model +2

Thermal-NeRF: Neural Radiance Fields from an Infrared Camera

no code implementations15 Mar 2024 Tianxiang Ye, Qi Wu, Junyuan Deng, Guoqing Liu, Liu Liu, Songpengcheng Xia, Liang Pang, Wenxian Yu, Ling Pei

In recent years, Neural Radiance Fields (NeRFs) have demonstrated significant potential in encoding highly-detailed 3D geometry and environmental appearance, positioning themselves as a promising alternative to traditional explicit representation for 3D scene reconstruction.

3D Scene Reconstruction

Decomposing Disease Descriptions for Enhanced Pathology Detection: A Multi-Aspect Vision-Language Pre-training Framework

1 code implementation CVPR 2024 Vu Minh Hieu Phan, Yutong Xie, Yuankai Qi, Lingqiao Liu, Liyang Liu, BoWen Zhang, Zhibin Liao, Qi Wu, Minh-Son To, Johan W. Verjans

Medical vision language pre-training (VLP) has emerged as a frontier of research, enabling zero-shot pathological recognition by comparing the query image with the textual descriptions for each disease.

Language Modelling Large Language Model

Unveiling the Potential of Robustness in Evaluating Causal Inference Models

no code implementations28 Feb 2024 Yiyan Huang, Cheuk Hang Leung, Siyi Wang, Yijun Li, Qi Wu

The growing demand for personalized decision-making has led to a surge of interest in estimating the Conditional Average Treatment Effect (CATE).

Causal Inference counterfactual +2

Explicit Interaction for Fusion-Based Place Recognition

2 code implementations27 Feb 2024 Jingyi Xu, Junyi Ma, Qi Wu, Zijie Zhou, Yue Wang, Xieyuanli Chen, Ling Pei

Fusion-based place recognition is an emerging technique jointly utilizing multi-modal perception data, to recognize previously visited places in GPS-denied scenarios for robots and autonomous vehicles.

Autonomous Vehicles

NaVid: Video-based VLM Plans the Next Step for Vision-and-Language Navigation

no code implementations24 Feb 2024 Jiazhao Zhang, Kunyu Wang, Rongtao Xu, Gengze Zhou, Yicong Hong, Xiaomeng Fang, Qi Wu, Zhizheng Zhang, He Wang

Vision-and-language navigation (VLN) stands as a key research problem of Embodied AI, aiming at enabling agents to navigate in unseen environments following linguistic instructions.

Decision Making Instruction Following +3

ModaVerse: Efficiently Transforming Modalities with LLMs

1 code implementation CVPR 2024 Xinyu Wang, Bohan Zhuang, Qi Wu

This alignment process, which synchronizes a language model trained on textual data with encoders and decoders trained on multi-modal data, often necessitates extensive training of several projection layers in multiple stages.

Language Modelling Large Language Model

SpecFormer: Guarding Vision Transformer Robustness via Maximum Singular Value Penalization

1 code implementation2 Jan 2024 Xixu Hu, Runkai Zheng, Jindong Wang, Cheuk Hang Leung, Qi Wu, Xing Xie

Vision Transformers (ViTs) are increasingly used in computer vision due to their high performance, but their vulnerability to adversarial attacks is a concern.

Computational Efficiency

WebVLN: Vision-and-Language Navigation on Websites

1 code implementation25 Dec 2023 Qi Chen, Dileepa Pitawela, Chongyang Zhao, Gengze Zhou, Hsiang-Ting Chen, Qi Wu

Vision-and-Language Navigation (VLN) task aims to enable AI agents to accurately understand and follow natural language instructions to navigate through real-world environments, ultimately reaching specific target locations.

Navigate Vision and Language Navigation

Subject-Oriented Video Captioning

no code implementations20 Dec 2023 Yunchuan Ma, Chang Teng, Yuankai Qi, Guorong Li, Laiyu Qing, Qi Wu, Qingming Huang

To address this problem, we propose a new video captioning task, subject-oriented video captioning, which allows users to specify the describing target via a bounding box.

Video Captioning

MMBaT: A Multi-task Framework for mmWave-based Human Body Reconstruction and Translation Prediction

no code implementations16 Dec 2023 Jiarui Yang, Songpengcheng Xia, YiFan Song, Qi Wu, Ling Pei

Human body reconstruction with Millimeter Wave (mmWave) radar point clouds has gained significant interest due to its ability to work in adverse environments and its capacity to mitigate privacy concerns associated with traditional camera-based solutions.

The Causal Impact of Credit Lines on Spending Distributions

1 code implementation16 Dec 2023 Yijun Li, Cheuk Hang Leung, Xiangqian Sun, Chaoqun Wang, Yiyan Huang, Xing Yan, Qi Wu, Dongdong Wang, Zhixiang Huang

Consumer credit services offered by e-commerce platforms provide customers with convenient loan access during shopping and have the potential to stimulate sales.

Invariant Random Forest: Tree-Based Model Solution for OOD Generalization

no code implementations7 Dec 2023 Yufan Liao, Qi Wu, Xing Yan

This paper introduces a novel and effective solution for OOD generalization of decision tree models, named Invariant Decision Tree (IDT).

Dynamic Inertial Poser (DynaIP): Part-Based Motion Dynamics Learning for Enhanced Human Pose Estimation with Sparse Inertial Sensors

1 code implementation CVPR 2024 Yu Zhang, Songpengcheng Xia, Lei Chu, Jiarui Yang, Qi Wu, Ling Pei

This paper introduces a novel human pose estimation approach using sparse inertial sensors, addressing the shortcomings of previous methods reliant on synthetic data.

Diversity Pose Estimation

Continual Self-supervised Learning: Towards Universal Multi-modal Medical Data Representation Learning

1 code implementation CVPR 2024 Yiwen Ye, Yutong Xie, Jianpeng Zhang, Ziyang Chen, Qi Wu, Yong Xia

In this paper, we reconsider versatile self-supervised learning from the perspective of continual learning and propose MedCoSS, a continuous self-supervised learning approach for multi-modal medical data.

Continual Learning Representation Learning +1

Watermarking Vision-Language Pre-trained Models for Multi-modal Embedding as a Service

1 code implementation10 Nov 2023 Yuanmin Tang, Jing Yu, Keke Gai, Xiangyan Qu, Yue Hu, Gang Xiong, Qi Wu

Our extensive experiments on various datasets indicate that the proposed watermarking approach is effective and safe for verifying the copyright of VLPs for multi-modal EaaS and robust against model extraction attacks.

Model extraction

Context-I2W: Mapping Images to Context-dependent Words for Accurate Zero-Shot Composed Image Retrieval

1 code implementation28 Sep 2023 Yuanmin Tang, Jing Yu, Keke Gai, Jiamin Zhuang, Gang Xiong, Yue Hu, Qi Wu

Different from Composed Image Retrieval task that requires expensive labels for training task-specific models, Zero-Shot Composed Image Retrieval (ZS-CIR) involves diverse tasks with a broad range of visual content manipulation intent that could be related to domain, scene, object, and attribute.

Attribute Image Retrieval +4

Align before Search: Aligning Ads Image to Text for Accurate Cross-Modal Sponsored Search

1 code implementation28 Sep 2023 Yuanmin Tang, Jing Yu, Keke Gai, Yujing Wang, Yue Hu, Gang Xiong, Qi Wu

Conventional research mainly studies from the view of modeling the implicit correlations between images and texts for query-ads matching, ignoring the alignment of detailed product information and resulting in suboptimal search performance. In this work, we propose a simple alignment network for explicitly mapping fine-grained visual parts in ads images to the corresponding text, which leverages the co-occurrence structure consistency between vision and language spaces without requiring expensive labeled training data.

cross-modal alignment Image-text matching +1

SwitchGPT: Adapting Large Language Models for Non-Text Outputs

no code implementations14 Sep 2023 Xinyu Wang, Bohan Zhuang, Qi Wu

To bridge this gap, we propose a novel approach, \methodname, from a modality conversion perspective that evolves a text-based LLM into a multi-modal one.

DeLELSTM: Decomposition-based Linear Explainable LSTM to Capture Instantaneous and Long-term Effects in Time Series

no code implementations26 Aug 2023 Chaoqun Wang, Yijun Li, Xiangqian Sun, Qi Wu, Dongdong Wang, Zhixiang Huang

The tensorized LSTM assigns each variable with a unique hidden state making up a matrix $\mathbf{h}_t$, and the standard LSTM models all the variables with a shared hidden state $\mathbf{H}_t$.

Time Series Time Series Forecasting

BHSD: A 3D Multi-Class Brain Hemorrhage Segmentation Dataset

1 code implementation22 Aug 2023 Biao Wu, Yutong Xie, Zeyu Zhang, Jinchao Ge, Kaspar Yaxley, Suzan Bahadir, Qi Wu, Yifan Liu, Minh-Son To

Intracranial hemorrhage (ICH) is a pathological condition characterized by bleeding inside the skull or brain, which can be attributed to various factors.

Image Segmentation Medical Image Segmentation +2

March in Chat: Interactive Prompting for Remote Embodied Referring Expression

1 code implementation ICCV 2023 Yanyuan Qiao, Yuankai Qi, Zheng Yu, Jing Liu, Qi Wu

Nevertheless, this poses more challenges than other VLN tasks since it requires agents to infer a navigation plan only based on a short instruction.

Referring Expression Vision and Language Navigation

VLN-PETL: Parameter-Efficient Transfer Learning for Vision-and-Language Navigation

1 code implementation ICCV 2023 Yanyuan Qiao, Zheng Yu, Qi Wu

The performance of the Vision-and-Language Navigation~(VLN) tasks has witnessed rapid progress recently thanks to the use of large pre-trained vision-and-language models.

Transfer Learning Vision and Language Navigation +1

Likelihood-Based Text-to-Image Evaluation with Patch-Level Perceptual and Semantic Credit Assignment

1 code implementation16 Aug 2023 Qi Chen, Chaorui Deng, Zixiong Huang, BoWen Zhang, Mingkui Tan, Qi Wu

In this paper, we propose to evaluate text-to-image generation performance by directly estimating the likelihood of the generated images using a pre-trained likelihood-based text-to-image generative model, i. e., a higher likelihood indicates better perceptual quality and better text-image alignment.

Text-to-Image Generation

Prompt Switch: Efficient CLIP Adaptation for Text-Video Retrieval

1 code implementation ICCV 2023 Chaorui Deng, Qi Chen, Pengda Qin, Da Chen, Qi Wu

In text-video retrieval, recent works have benefited from the powerful learning capabilities of pre-trained text-image foundation models (e. g., CLIP) by adapting them to the video domain.

Retrieval Video Captioning +1

Self-Prompting Large Vision Models for Few-Shot Medical Image Segmentation

1 code implementation15 Aug 2023 Qi Wu, Yuyao Zhang, Marawan Elbatel

Recent advancements in large foundation models have shown promising potential in the medical industry due to their flexible prompting capability.

Decoder Image Segmentation +3

Identity-Consistent Aggregation for Video Object Detection

1 code implementation ICCV 2023 Chaorui Deng, Da Chen, Qi Wu

In Video Object Detection (VID), a common practice is to leverage the rich temporal contexts from the video to enhance the object representations in each frame.

Object object-detection +1

AerialVLN: Vision-and-Language Navigation for UAVs

1 code implementation ICCV 2023 Shubo Liu, Hongsheng Zhang, Yuankai Qi, Peng Wang, Yaning Zhang, Qi Wu

Navigating in the sky is more complicated than on the ground because agents need to consider the flying height and more complex spatial relationship reasoning.

cross-modal alignment Navigate +1

Scaling Data Generation in Vision-and-Language Navigation

1 code implementation ICCV 2023 Zun Wang, Jialu Li, Yicong Hong, Yi Wang, Qi Wu, Mohit Bansal, Stephen Gould, Hao Tan, Yu Qiao

Recent research in language-guided visual navigation has demonstrated a significant demand for the diversity of traversable environments and the quantity of supervision for training generalizable agents.

Imitation Learning Vision and Language Navigation +1

Probabilistic Learning of Multivariate Time Series with Temporal Irregularity

1 code implementation15 Jun 2023 Yijun Li, Cheuk Hang Leung, Qi Wu

Multivariate sequential data collected in practice often exhibit temporal irregularities, including nonuniform time intervals and component misalignment.

Imputation Time Series

Deep into The Domain Shift: Transfer Learning through Dependence Regularization

1 code implementation31 May 2023 Shumin Ma, Zhiri Yuan, Qi Wu, Yiyan Huang, Xixu Hu, Cheuk Hang Leung, Dongdong Wang, Zhixiang Huang

This paper proposes a new domain adaptation approach in which one can measure the differences in the internal dependence structure separately from those in the marginals.

Domain Adaptation Transfer Learning

Attention Mechanisms in Medical Image Segmentation: A Survey

no code implementations29 May 2023 Yutong Xie, Bing Yang, Qingbiao Guan, Jianpeng Zhang, Qi Wu, Yong Xia

This paper systematically reviews the basic principles of attention mechanisms and their applications in medical image segmentation.

Image Segmentation Medical Image Segmentation +3

NavGPT: Explicit Reasoning in Vision-and-Language Navigation with Large Language Models

1 code implementation26 May 2023 Gengze Zhou, Yicong Hong, Qi Wu

Trained with an unprecedented scale of data, large language models (LLMs) like ChatGPT and GPT-4 exhibit the emergence of significant reasoning abilities from model scaling.

Instruction Following Vision and Language Navigation +1

S4M: Generating Radiology Reports by A Single Model for Multiple Body Parts

1 code implementation26 May 2023 Qi Chen, Yutong Xie, Biao Wu, Minh-Son To, James Ang, Qi Wu

In this paper, we seek to design a report generation model that is able to generate reasonable reports even given different images of various body parts.

 Ranked #1 on Medical Report Generation on IU X-Ray (using extra training data)

Medical Report Generation

Realistic Noise Synthesis with Diffusion Models

no code implementations23 May 2023 Qi Wu, Mingyan Han, Ting Jiang, Haoqiang Fan, Bing Zeng, Shuaicheng Liu

Deep image denoising models often rely on large amount of training data for the high quality performance.

Image Denoising Noise Estimation

DIPNet: Efficiency Distillation and Iterative Pruning for Image Super-Resolution

no code implementations14 Apr 2023 Lei Yu, Xinpeng Li, Youwei Li, Ting Jiang, Qi Wu, Haoqiang Fan, Shuaicheng Liu

To address this issue, we propose a novel multi-stage lightweight network boosting method, which can enable lightweight networks to achieve outstanding performance.

Image Super-Resolution Network Pruning

Photon Field Networks for Dynamic Real-Time Volumetric Global Illumination

no code implementations14 Apr 2023 David Bauer, Qi Wu, Kwan-Liu Ma

In this paper, we present a novel method to enable real-time global illumination for volume data visualization.

Data Visualization

HyperINR: A Fast and Predictive Hypernetwork for Implicit Neural Representations via Knowledge Distillation

no code implementations9 Apr 2023 Qi Wu, David Bauer, Yuyang Chen, Kwan-Liu Ma

Implicit Neural Representations (INRs) have recently exhibited immense potential in the field of scientific visualization for both data generation and visualization tasks.

Knowledge Distillation Novel View Synthesis +1

Distributed Neural Representation for Reactive in situ Visualization

no code implementations28 Mar 2023 Qi Wu, Joseph A. Insley, Victor A. Mateevitsi, Silvio Rizzi, Michael E. Papka, Kwan-Liu Ma

In this work, we develop an implicit neural representation for distributed volume data and incorporate it into the DIVA reactive programming system.

Program Generation from Diverse Video Demonstrations

no code implementations1 Feb 2023 Anthony Manchin, Jamie Sherrah, Qi Wu, Anton Van Den Hengel

The ability to use inductive reasoning to extract general rules from multiple observations is a vital indicator of intelligence.

Dynamic CVaR Portfolio Construction with Attention-Powered Generative Factor Learning

no code implementations18 Jan 2023 Chuting Sun, Qi Wu, Xing Yan

The dynamic portfolio construction problem requires dynamic modeling of the joint distribution of multivariate stock returns.

Portfolio Optimization

Learning to Dub Movies via Hierarchical Prosody Models

1 code implementation CVPR 2023 Gaoxiang Cong, Liang Li, Yuankai Qi, ZhengJun Zha, Qi Wu, Wenyu Wang, Bin Jiang, Ming-Hsuan Yang, Qingming Huang

Given a piece of text, a video clip and a reference audio, the movie dubbing (also known as visual voice clone V2C) task aims to generate speeches that match the speaker's emotion presented in the video using the desired speaker voice as reference.

Decorr: Environment Partitioning for Invariant Learning and OOD Generalization

no code implementations18 Nov 2022 Yufan Liao, Qi Wu, Xing Yan

Invariant learning methods try to find an invariant predictor across several environments and have become popular in OOD generalization.

Efficient and Accurate Quantized Image Super-Resolution on Mobile NPUs, Mobile AI & AIM 2022 challenge: Report

2 code implementations7 Nov 2022 Andrey Ignatov, Radu Timofte, Maurizio Denna, Abdel Younes, Ganzorig Gankhuyag, Jingang Huh, Myeong Kyun Kim, Kihwan Yoon, Hyeon-Cheol Moon, Seungho Lee, Yoonsik Choe, Jinwoo Jeong, Sungjei Kim, Maciej Smyl, Tomasz Latkowski, Pawel Kubik, Michal Sokolski, Yujie Ma, Jiahao Chao, Zhou Zhou, Hongfan Gao, Zhengfeng Yang, Zhenbing Zeng, Zhengyang Zhuge, Chenghua Li, Dan Zhu, Mengdi Sun, Ran Duan, Yan Gao, Lingshun Kong, Long Sun, Xiang Li, Xingdong Zhang, Jiawei Zhang, Yaqi Wu, Jinshan Pan, Gaocheng Yu, Jin Zhang, Feng Zhang, Zhe Ma, Hongbin Wang, Hojin Cho, Steve Kim, Huaen Li, Yanbo Ma, Ziwei Luo, Youwei Li, Lei Yu, Zhihong Wen, Qi Wu, Haoqiang Fan, Shuaicheng Liu, Lize Zhang, Zhikai Zong, Jeremy Kwon, Junxi Zhang, Mengyuan Li, Nianxiang Fu, Guanchen Ding, Han Zhu, Zhenzhong Chen, Gen Li, Yuanfan Zhang, Lei Sun, Dafeng Zhang, Neo Yang, Fitz Liu, Jerry Zhao, Mustafa Ayazoglu, Bahri Batuhan Bilecen, Shota Hirose, Kasidis Arunruangsirilert, Luo Ao, Ho Chun Leung, Andrew Wei, Jie Liu, Qiang Liu, Dahai Yu, Ao Li, Lei Luo, Ce Zhu, Seongmin Hong, Dongwon Park, Joonhee Lee, Byeong Hyun Lee, Seunggyu Lee, Se Young Chun, Ruiyuan He, Xuhao Jiang, Haihang Ruan, Xinjian Zhang, Jing Liu, Garas Gendy, Nabil Sabor, Jingchao Hou, Guanghui He

While numerous solutions have been proposed for this problem in the past, they are usually not compatible with low-power mobile NPUs having many computational and memory constraints.

Image Super-Resolution

Toward 3D Spatial Reasoning for Human-like Text-based Visual Question Answering

no code implementations21 Sep 2022 Hao Li, Jinfa Huang, Peng Jin, Guoli Song, Qi Wu, Jie Chen

Under this setting, these 2D spatial reasoning approaches cannot distinguish the fine-grain spatial relations between visual objects and scene texts on the same image plane, thereby impairing the interpretability and performance of TextVQA models.

Image Captioning Optical Character Recognition (OCR) +2

FoVolNet: Fast Volume Rendering using Foveated Deep Neural Networks

no code implementations20 Sep 2022 David Bauer, Qi Wu, Kwan-Liu Ma

We introduce FoVolNet -- a method to significantly increase the performance of volume data visualization.

Data Visualization Image Reconstruction +1

Learning Distinct and Representative Styles for Image Captioning

1 code implementation17 Sep 2022 Qi Chen, Chaorui Deng, Qi Wu

Our innovative idea is to explore the rich modes in the training caption corpus to learn a set of "mode embeddings", and further use them to control the mode of the generated captions for existing image captioning models.

Diversity Image Captioning +1

Robust Causal Learning for the Estimation of Average Treatment Effects

no code implementations5 Sep 2022 Yiyan Huang, Cheuk Hang Leung, Xing Yan, Qi Wu, Shumin Ma, Zhiri Yuan, Dongdong Wang, Zhixiang Huang

Theoretically, the RCL estimators i) are as consistent and doubly robust as the DML estimators, and ii) can get rid of the error-compounding issue.

Decision Making

Moderately-Balanced Representation Learning for Treatment Effects with Orthogonality Information

no code implementations5 Sep 2022 Yiyan Huang, Cheuk Hang Leung, Shumin Ma, Qi Wu, Dongdong Wang, Zhixiang Huang

In this paper, we propose a moderately-balanced representation learning (MBRL) framework based on recent covariates balanced representation learning methods and orthogonal machine learning theory.

Learning Theory Multi-Task Learning +2

ClusTR: Exploring Efficient Self-attention via Clustering for Vision Transformers

no code implementations28 Aug 2022 Yutong Xie, Jianpeng Zhang, Yong Xia, Anton Van Den Hengel, Qi Wu

Besides, we further extend the clustering-guided attention from single-scale to multi-scale, which is conducive to dense prediction tasks.

Clustering Diversity +1

Fast Nearest Convolution for Real-Time Efficient Image Super-Resolution

2 code implementations24 Aug 2022 Ziwei Luo, Youwei Li, Lei Yu, Qi Wu, Zhihong Wen, Haoqiang Fan, Shuaicheng Liu

The proposed nearest convolution has the same performance as the nearest upsampling but is much faster and more suitable for Android NNAPI.

Image Super-Resolution Quantization

Interactive Volume Visualization via Multi-Resolution Hash Encoding based Neural Representation

1 code implementation23 Jul 2022 Qi Wu, David Bauer, Michael J. Doyle, Kwan-Liu Ma

Neural networks have shown great potential in compressing volume data for visualization.

Optical Field Recovery in Jones Space

no code implementations22 Jun 2022 Qi Wu, Yixiao Zhu, Hexun Jiang, Qunbi Zhuge, Weisheng Hu

For cost-sensitive short-reach optical networks, some advanced single-polarization (SP) optical field recovery schemes are recently proposed to avoid chromatic dispersion-induced power fading effect, and improve the spectral efficiency for larger potential capacity.


Attract me to Buy: Advertisement Copywriting Generation with Multimodal Multi-structured Information

no code implementations7 May 2022 Zhipeng Zhang, Xinglin Hou, Kai Niu, Zhongzhen Huang, Tiezheng Ge, Yuning Jiang, Qi Wu, Peng Wang

Therefore, we present a dataset, E-MMAD (e-commercial multimodal multi-structured advertisement copywriting), which requires, and supports much more detailed information in text generation.

Text Generation Video Captioning

BSRT: Improving Burst Super-Resolution with Swin Transformer and Flow-Guided Deformable Alignment

1 code implementation18 Apr 2022 Ziwei Luo, Youwei Li, Shen Cheng, Lei Yu, Qi Wu, Zhihong Wen, Haoqiang Fan, Jian Sun, Shuaicheng Liu

To overcome the challenges in BurstSR, we propose a Burst Super-Resolution Transformer (BSRT), which can significantly improve the capability of extracting inter-frame information and reconstruction.

Burst Image Reconstruction Burst Image Super-Resolution +2

Vision-and-Language Navigation: A Survey of Tasks, Methods, and Future Directions

1 code implementation ACL 2022 Jing Gu, Eliana Stefani, Qi Wu, Jesse Thomason, Xin Eric Wang

A long-term goal of AI research is to build intelligent agents that can communicate with humans in natural language, perceive the environment, and perform real-world tasks.

Vision and Language Navigation

MuKEA: Multimodal Knowledge Extraction and Accumulation for Knowledge-based Visual Question Answering

1 code implementation CVPR 2022 Yang Ding, Jing Yu, Bang Liu, Yue Hu, Mingxin Cui, Qi Wu

Knowledge-based visual question answering requires the ability of associating external knowledge for open-ended cross-modal scene understanding.

Implicit Relations Question Answering +2

Bridging the Gap Between Learning in Discrete and Continuous Environments for Vision-and-Language Navigation

1 code implementation CVPR 2022 Yicong Hong, Zun Wang, Qi Wu, Stephen Gould

To bridge the discrete-to-continuous gap, we propose a predictor to generate a set of candidate waypoints during navigation, so that agents designed with high-level actions can be transferred to and trained in continuous environments.

Imitation Learning Vision and Language Navigation

Maintaining Reasoning Consistency in Compositional Visual Question Answering

1 code implementation CVPR 2022 Chenchen Jing, Yunde Jia, Yuwei Wu, Xinyu Liu, Qi Wu

Existing VQA models can answer a compositional question well, but cannot work well in terms of reasoning consistency in answering the compositional question and its sub-questions.

Question Answering Visual Question Answering

LocFormer: Enabling Transformers to Perform Temporal Moment Localization on Long Untrimmed Videos With a Feature Sampling Approach

no code implementations19 Dec 2021 Cristian Rodriguez-Opazo, Edison Marrese-Taylor, Basura Fernando, Hiroya Takamura, Qi Wu

We propose LocFormer, a Transformer-based model for video grounding which operates at a constant memory footprint regardless of the video length, i. e. number of frames.

Inductive Bias Video Grounding

UniMiSS: Universal Medical Self-Supervised Learning via Breaking Dimensionality Barrier

1 code implementation17 Dec 2021 Yutong Xie, Jianpeng Zhang, Yong Xia, Qi Wu

In this paper, we advocate bringing a wealth of 2D images like chest X-rays as compensation for the lack of 3D data, aiming to build a universal medical self-supervised representation learning framework, called UniMiSS.

Image Classification Medical Image Classification +2

Debiased Visual Question Answering from Feature and Sample Perspectives

1 code implementation NeurIPS 2021 Zhiquan Wen, Guanghui Xu, Mingkui Tan, Qingyao Wu, Qi Wu

From the sample perspective, we construct two types of negative samples to assist the training of the models, without introducing additional annotations.

Bias Detection Question Answering +1

V2C: Visual Voice Cloning

no code implementations CVPR 2022 Qi Chen, Yuanqing Li, Yuankai Qi, Jiaqiu Zhou, Mingkui Tan, Qi Wu

Existing Voice Cloning (VC) tasks aim to convert a paragraph text to a speech with desired voice specified by a reference audio.

Voice Cloning

Medical Visual Question Answering: A Survey

no code implementations19 Nov 2021 Zhihong Lin, Donghao Zhang, Qingyi Tao, Danli Shi, Gholamreza Haffari, Qi Wu, Mingguang He, ZongYuan Ge

Medical Visual Question Answering~(VQA) is a combination of medical artificial intelligence and popular VQA challenges.

Medical Visual Question Answering Question Answering +1

Risk and return prediction for pricing portfolios of non-performing consumer credit

no code implementations28 Oct 2021 Siyi Wang, Xing Yan, Bangqi Zheng, Hu Wang, Wangli Xu, Nanbo Peng, Qi Wu

We design a system for risk-analyzing and pricing portfolios of non-performing consumer credit loans.

quantile regression

Memory Regulation and Alignment toward Generalizer RGB-Infrared Person

1 code implementation18 Sep 2021 Feng Chen, Fei Wu, Qi Wu, Zhiguo Wan

The domain shift, coming from unneglectable modality gap and non-overlapped identity classes between training and test sets, is a major issue of RGB-Infrared person re-identification.

Attribute Metric Learning +1

Data-driven advice for interpreting local and global model predictions in bioinformatics problems

no code implementations13 Aug 2021 Markus Loecher, Qi Wu

For random forests, we find extremely high similarities and correlations of both local and global SHAP values and CFC scores, leading to very similar rankings and interpretations.

Feature Importance

Communicative Learning with Natural Gestures for Embodied Navigation Agents with Human-in-the-Scene

1 code implementation5 Aug 2021 Qi Wu, Cheng-Ju Wu, Yixin Zhu, Jungseock Joo

In a series of experiments, we demonstrate that human gesture cues, even without predefined semantics, improve the object-goal navigation for an embodied agent, outperforming various state-of-the-art methods.

Data Hiding with Deep Learning: A Survey Unifying Digital Watermarking and Steganography

no code implementations20 Jul 2021 Zihan Wang, Olivia Byrnes, Hu Wang, Ruoxi Sun, Congbo Ma, Huaming Chen, Qi Wu, Minhui Xue

The advancement of secure communication and identity verification fields has significantly increased through the use of deep learning techniques for data hiding.

Sketch, Ground, and Refine: Top-Down Dense Video Captioning

no code implementations CVPR 2021 Chaorui Deng, ShiZhe Chen, Da Chen, Yuan He, Qi Wu

The dense video captioning task aims to detect and describe a sequence of events in a video for detailed and coherent storytelling.

Dense Video Captioning Sentence

Room-and-Object Aware Knowledge Reasoning for Remote Embodied Referring Expression

1 code implementation CVPR 2021 Chen Gao, Jinyu Chen, Si Liu, Luting Wang, Qiong Zhang, Qi Wu

The Remote Embodied Referring Expression (REVERIE) is a recently raised task that requires an agent to navigate to and localise a referred remote object according to a high-level language instruction.

Instruction Following Navigate +2

Proposal-free One-stage Referring Expression via Grid-Word Cross-Attention

no code implementations5 May 2021 Wei Suo, Mengyang Sun, Peng Wang, Qi Wu

Referring Expression Comprehension (REC) has become one of the most important tasks in visual reasoning, since it is an essential step for many vision-and-language tasks such as visual question answering.

Question Answering Referring Expression +3

Chop Chop BERT: Visual Question Answering by Chopping VisualBERT's Heads

no code implementations30 Apr 2021 Chenyu Gao, Qi Zhu, Peng Wang, Qi Wu

Based on this observation, we design a dynamic chopping module that can automatically remove heads and layers of the VisualBERT at an instance level when dealing with different questions.

Question Answering Visual Question Answering +1

Towards Accurate Text-based Image Captioning with Content Diversity Exploration

1 code implementation CVPR 2021 Guanghui Xu, Shuaicheng Niu, Mingkui Tan, Yucheng Luo, Qing Du, Qi Wu

This task, however, is very challenging because an image often contains complex texts and visual information that is hard to be described comprehensively.

Caption Generation Diversity +1

The Road to Know-Where: An Object-and-Room Informed Sequential BERT for Indoor Vision-Language Navigation

1 code implementation ICCV 2021 Yuankai Qi, Zizheng Pan, Yicong Hong, Ming-Hsuan Yang, Anton Van Den Hengel, Qi Wu

Vision-and-Language Navigation (VLN) requires an agent to find a path to a remote location on the basis of natural-language instructions and a set of photo-realistic panoramas.

Vision and Language Navigation Vision-Language Navigation

Jo-SRC: A Contrastive Approach for Combating Noisy Labels

no code implementations CVPR 2021 Yazhou Yao, Zeren Sun, Chuanyi Zhang, Fumin Shen, Qi Wu, Jian Zhang, Zhenmin Tang

Due to the memorization effect in Deep Neural Networks (DNNs), training with noisy labels usually results in inferior model performance.

Contrastive Learning Memorization

Robust Orthogonal Machine Learning of Treatment Effects

no code implementations22 Mar 2021 Yiyan Huang, Cheuk Hang Leung, Qi Wu, Xing Yan

Theoretically, the RCL estimators i) satisfy the (higher-order) orthogonal condition and are as \textit{consistent and doubly robust} as the DML estimators, and ii) get rid of the error-compounding issue.

BIG-bench Machine Learning

Learning for Visual Navigation by Imagining the Success

no code implementations28 Feb 2021 Mahdi Kazemi Moghaddam, Ehsan Abbasnejad, Qi Wu, Javen Shi, Anton Van Den Hengel

ForeSIT is trained to imagine the recurrent latent representation of a future state that leads to success, e. g. either a sub-goal state that is important to reach before the target, or the goal state itself.

Navigate Reinforcement Learning (RL) +1

Multi-intersection Traffic Optimisation: A Benchmark Dataset and a Strong Baseline

no code implementations24 Jan 2021 Hu Wang, Hao Chen, Qi Wu, Congbo Ma, Yidong Li, Chunhua Shen

To address these issues, in this work we carefully design our settings and propose a new dataset including both synthetic and real traffic data in more complex scenarios.


How to Train Your Agent to Read and Write

1 code implementation4 Jan 2021 Li Liu, Mengge He, Guanghui Xu, Mingkui Tan, Qi Wu

Typically, this requires an agent to fully understand the knowledge from the given text materials and generate correct and fluent novel paragraphs, which is very challenging in practice.

KG-to-Text Generation Knowledge Graphs

Semantics for Robotic Mapping, Perception and Interaction: A Survey

no code implementations2 Jan 2021 Sourav Garg, Niko Sünderhauf, Feras Dayoub, Douglas Morrison, Akansel Cosgun, Gustavo Carneiro, Qi Wu, Tat-Jun Chin, Ian Reid, Stephen Gould, Peter Corke, Michael Milford

In robotics and related research fields, the study of understanding is often referred to as semantics, which dictates what does the world "mean" to a robot, and is strongly tied to the question of how to represent that meaning.

Autonomous Driving Navigate

Memory-Gated Recurrent Networks

1 code implementation24 Dec 2020 Yaquan Zhang, Qi Wu, Nanbo Peng, Min Dai, Jing Zhang, Hu Wang

The essence of multivariate sequential learning is all about how to extract dependencies in data.

Time Series Time Series Analysis

The Causal Learning of Retail Delinquency

no code implementations17 Dec 2020 Yiyan Huang, Cheuk Hang Leung, Xing Yan, Qi Wu, Nanbo Peng, Dongdong Wang, Zhixiang Huang

Classical estimators overlook the confounding effects and hence the estimation error can be magnificent.

Simple is not Easy: A Simple Strong Baseline for TextVQA and TextCaps

1 code implementation9 Dec 2020 Qi Zhu, Chenyu Gao, Peng Wang, Qi Wu

Texts appearing in daily scenes that can be recognized by OCR (Optical Character Recognition) tools contain significant information, such as street name, product brand and prices.

Decoder Image Captioning +4

Confidence-aware Non-repetitive Multimodal Transformers for TextCaps

1 code implementation7 Dec 2020 Zhaokai Wang, Renda Bao, Qi Wu, Si Liu

Our CNMT consists of a reading, a reasoning and a generation modules, in which Reading Module employs better OCR systems to enhance text reading ability and a confidence embedding to select the most noteworthy tokens.

Image Captioning Optical Character Recognition +1

Generative Learning of Heterogeneous Tail Dependence

no code implementations26 Nov 2020 Xiangqian Sun, Xing Yan, Qi Wu

We propose a multivariate generative model to capture the complex dependence structure often encountered in business and financial data.

Modular Graph Attention Network for Complex Visual Relational Reasoning

no code implementations22 Nov 2020 Yihan Zheng, Zhiquan Wen, Mingkui Tan, Runhao Zeng, Qi Chen, YaoWei Wang, Qi Wu

Moreover, to capture the complex logic in a query, we construct a relational graph to represent the visual objects and their relationships, and propose a multi-step reasoning method to progressively understand the complex logic.

Graph Attention Question Answering +5

Language-guided Navigation via Cross-Modal Grounding and Alternate Adversarial Learning

no code implementations22 Nov 2020 Weixia Zhang, Chao Ma, Qi Wu, Xiaokang Yang

We then propose to recursively alternate the learning schemes of imitation and exploration to narrow the discrepancy between training and inference.

Imitation Learning Navigate +1

Language and Visual Entity Relationship Graph for Agent Navigation

1 code implementation NeurIPS 2020 Yicong Hong, Cristian Rodriguez-Opazo, Yuankai Qi, Qi Wu, Stephen Gould

From both the textual and visual perspectives, we find that the relationships among the scene, its objects, and directional clues are essential for the agent to interpret complex instructions and correctly perceive the environment.

Dynamic Time Warping Navigate +2

MARS: Mixed Virtual and Real Wearable Sensors for Human Activity Recognition with Multi-Domain Deep Learning Model

no code implementations20 Sep 2020 Ling Pei, Songpengcheng Xia, Lei Chu, Fanyi Xiao, Qi Wu, Wenxian Yu, Robert Qiu

Together with the rapid development of the Internet of Things (IoT), human activity recognition (HAR) using wearable Inertial Measurement Units (IMUs) becomes a promising technology for many research areas.

Human Activity Recognition Transfer Learning

CogTree: Cognition Tree Loss for Unbiased Scene Graph Generation

1 code implementation16 Sep 2020 Jing Yu, Yuan Chai, Yujing Wang, Yue Hu, Qi Wu

We first build a cognitive structure CogTree to organize the relationships based on the prediction of a biased SGG model.

Ranked #2 on Scene Graph Generation on Visual Genome (mean Recall @20 metric)

Graph Generation Unbiased Scene Graph Generation

Attention-SLAM: A Visual Monocular SLAM Learning from Human Gaze

1 code implementation15 Sep 2020 Jinquan Li, Ling Pei, Danping Zou, Songpengcheng Xia, Qi Wu, Tao Li, Zhen Sun, Wenxian Yu

This paper proposes a novel simultaneous localization and mapping (SLAM) approach, namely Attention-SLAM, which simulates human navigation mode by combining a visual saliency model (SalNavNet) with traditional monocular visual SLAM.

Simultaneous Localization and Mapping

Data-driven Meta-set Based Fine-Grained Visual Classification

1 code implementation6 Aug 2020 Chuanyi Zhang, Yazhou Yao, Xiangbo Shu, Zechao Li, Zhenmin Tang, Qi Wu

To this end, we propose a data-driven meta-set based approach to deal with noisy web images for fine-grained recognition.

Classification Fine-Grained Image Classification +3

Object-and-Action Aware Model for Visual Language Navigation

no code implementations ECCV 2020 Yuankai Qi, Zizheng Pan, Shengping Zhang, Anton Van Den Hengel, Qi Wu

The first is object description (e. g., 'table', 'door'), each presenting as a tip for the agent to determine the next action by finding the item visible in the environment, and the second is action specification (e. g., 'go straight', 'turn left') which allows the robot to directly predict the next movements without relying on visual perceptions.

Object Vision and Language Navigation

Soft Expert Reward Learning for Vision-and-Language Navigation

no code implementations ECCV 2020 Hu Wang, Qi Wu, Chunhua Shen

In this paper, we introduce a Soft Expert Reward Learning (SERL) model to overcome the reward engineering designing and generalisation problems of the VLN task.

Reinforcement Learning (RL) Vision and Language Navigation

Length-Controllable Image Captioning

1 code implementation ECCV 2020 Chaorui Deng, Ning Ding, Mingkui Tan, Qi Wu

We verify the merit of the proposed length level embedding on three models: two state-of-the-art (SOTA) autoregressive models with different types of decoder, as well as our proposed non-autoregressive model, to show its generalization ability.

controllable image captioning Decoder +1

Semantic Equivalent Adversarial Data Augmentation for Visual Question Answering

1 code implementation ECCV 2020 Ruixue Tang, Chao Ma, Wei Emma Zhang, Qi Wu, Xiaokang Yang

However, there are few works studying the data augmentation problem for VQA and none of the existing image based augmentation schemes (such as rotation and flipping) can be directly applied to VQA due to its semantic structure -- an $\langle image, question, answer\rangle$ triplet needs to be maintained correctly.

Adversarial Attack Data Augmentation +2

Referring Expression Comprehension: A Survey of Methods and Datasets

no code implementations19 Jul 2020 Yanyuan Qiao, Chaorui Deng, Qi Wu

In this survey, we first examine the state of the art by comparing modern approaches to the problem.

object-detection Object Detection +2

DAM: Deliberation, Abandon and Memory Networks for Generating Detailed and Non-repetitive Responses in Visual Dialogue

4 code implementations7 Jul 2020 Xiaoze Jiang, Jing Yu, Yajing Sun, Zengchang Qin, Zihao Zhu, Yue Hu, Qi Wu

The ability of generating detailed and non-repetitive responses is crucial for the agent to achieve human-like conversation.

Mucko: Multi-Layer Cross-Modal Knowledge Reasoning for Fact-based Visual Question Answering

no code implementations16 Jun 2020 Zihao Zhu, Jing Yu, Yujing Wang, Yajing Sun, Yue Hu, Qi Wu

In this paper, we depict an image by a multi-modal heterogeneous graph, which contains multiple layers of information corresponding to the visual, semantic and factual features.

Question Answering Visual Question Answering

Foreground-Background Imbalance Problem in Deep Object Detectors: A Review

no code implementations16 Jun 2020 Joya Chen, Qi Wu, Dong Liu, Tong Xu

Recent years have witnessed the remarkable developments made by deep learning techniques for object detection, a fundamentally challenging problem of computer vision.

Object object-detection +1

Structured Multimodal Attentions for TextVQA

2 code implementations1 Jun 2020 Chenyu Gao, Qi Zhu, Peng Wang, Hui Li, Yuliang Liu, Anton Van Den Hengel, Qi Wu

In this paper, we propose an end-to-end structured multimodal attention (SMA) neural network to mainly solve the first two issues above.

Graph Attention Optical Character Recognition (OCR) +3

Optimistic Agent: Accurate Graph-Based Value Estimation for More Successful Visual Navigation

no code implementations7 Apr 2020 Mahdi Kazemi Moghaddam, Qi Wu, Ehsan Abbasnejad, Javen Qinfeng Shi

Through empirical studies, we show that our agent, dubbed as the optimistic agent, has a more realistic estimate of the state value during a navigation episode which leads to a higher success rate.

Reinforcement Learning (RL) Visual Navigation

Sub-Instruction Aware Vision-and-Language Navigation

1 code implementation EMNLP 2020 Yicong Hong, Cristian Rodriguez-Opazo, Qi Wu, Stephen Gould

Vision-and-language navigation requires an agent to navigate through a real 3D environment following natural language instructions.

Navigate Vision and Language Navigation

Fine-grained Video-Text Retrieval with Hierarchical Graph Reasoning

4 code implementations CVPR 2020 Shizhe Chen, Yida Zhao, Qin Jin, Qi Wu

To improve fine-grained video-text retrieval, we propose a Hierarchical Graph Reasoning (HGR) model, which decomposes video-text matching into global-to-local levels.

Cross-Modal Retrieval Text Matching +2

Say As You Wish: Fine-grained Control of Image Caption Generation with Abstract Scene Graphs

1 code implementation CVPR 2020 Shizhe Chen, Qin Jin, Peng Wang, Qi Wu

From the ASG, we propose a novel ASG2Caption model, which is able to recognise user intentions and semantics in the graph, and therefore generate desired captions according to the graph structure.

Attribute Caption Generation +2

Intelligent Home 3D: Automatic 3D-House Design from Linguistic Descriptions Only

1 code implementation CVPR 2020 Qi Chen, Qi Wu, Rui Tang, Yu-Han Wang, Shuai Wang, Mingkui Tan

To this end, we propose a House Plan Generative Model (HPGM) that first translates the language input to a structural graph representation and then predicts the layout of rooms with a Graph Conditioned Layout Prediction Network (GC LPN) and generates the interior texture with a Language Conditioned Texture GAN (LCT-GAN).

Text to 3D

DualVD: An Adaptive Dual Encoding Model for Deep Visual Understanding in Visual Dialogue

1 code implementation17 Nov 2019 Xiaoze Jiang, Jing Yu, Zengchang Qin, Yingying Zhuang, Xingxing Zhang, Yue Hu, Qi Wu

More importantly, we can tell which modality (visual or semantic) has more contribution in answering the current question by visualizing the gate values.

feature selection Question Answering +2

Integrating Temporal and Spatial Attentions for VATEX Video Captioning Challenge 2019

no code implementations15 Oct 2019 Shizhe Chen, Yida Zhao, Yuqing Song, Qin Jin, Qi Wu

This notebook paper presents our model in the VATEX video captioning challenge.

Video Captioning

BERT-based Financial Sentiment Index and LSTM-based Stock Return Predictability

no code implementations21 Jun 2019 Joshua Zoen Git Hiew, Xin Huang, Hao Mou, Duan Li, Qi Wu, Yabo Xu

On the other hand, by combining with the other two commonly-used methods when it comes to building the sentiment index in the financial literature, i. e., the option-implied and the market-implied approaches, we propose a more general and comprehensive framework for the financial sentiment analysis, and further provide convincing outcomes for the predictability of individual stock return by combining LSTM (with a feature of a nonlinear mapping).

Sentiment Analysis

Neural Learning of Online Consumer Credit Risk

no code implementations5 Jun 2019 Di Wang, Qi Wu, Wen Zhang

This paper takes a deep learning approach to understand consumer credit risk when e-commerce platforms issue unsecured credit to finance customers' purchase.

Time Series Time Series Analysis

Understanding Distributional Ambiguity via Non-robust Chance Constraint

no code implementations3 Jun 2019 Qi Wu, Shumin Ma, Cheuk Hang Leung, Wei Liu, Nanbo Peng

Without the boundedness constraint, the CCO problem is shown to perform uniformly better than the DRO problem, irrespective of the radius of the ambiguity set, the choice of the divergence measure, or the tail heaviness of the center distribution.

Portfolio Optimization

Show, Price and Negotiate: A Negotiator with Online Value Look-Ahead

no code implementations7 May 2019 Amin Parvaneh, Ehsan Abbasnejad, Qi Wu, Javen Qinfeng Shi, Anton Van Den Hengel

Negotiation, as an essential and complicated aspect of online shopping, is still challenging for an intelligent agent.

REVERIE: Remote Embodied Visual Referring Expression in Real Indoor Environments

1 code implementation CVPR 2020 Yuankai Qi, Qi Wu, Peter Anderson, Xin Wang, William Yang Wang, Chunhua Shen, Anton Van Den Hengel

One of the long-term challenges of robotics is to enable robots to interact with humans in the visual world via natural language, as humans are visual animals that communicate through language.

Referring Expression Vision and Language Navigation

You Only Look & Listen Once: Towards Fast and Accurate Visual Grounding

no code implementations12 Feb 2019 Chaorui Deng, Qi Wu, Guanghui Xu, Zhuliang Yu, Yanwu Xu, Kui Jia, Mingkui Tan

Most state-of-the-art methods in VG operate in a two-stage manner, wherein the first stage an object detector is adopted to generate a set of object proposals from the input image and the second stage is simply formulated as a cross-modal matching problem that finds the best match between the language query and all region proposals.

object-detection Object Detection +2

Gold Seeker: Information Gain from Policy Distributions for Goal-oriented Vision-and-Langauge Reasoning

no code implementations CVPR 2020 Ehsan Abbasnejad, Iman Abbasnejad, Qi Wu, Javen Shi, Anton Van Den Hengel

For each potential action a distribution of the expected outcomes is calculated, and the value of the potential information gain assessed.

Visual Dialog

What's to know? Uncertainty as a Guide to Asking Goal-oriented Questions

no code implementations CVPR 2019 Ehsan Abbasnejad, Qi Wu, Javen Shi, Anton Van Den Hengel

We propose a solution to this problem based on a Bayesian model of the uncertainty in the implicit model maintained by the visual dialogue agent, and in the function used to select an appropriate output.

Visual Dialog

Neighbourhood Watch: Referring Expression Comprehension via Language-guided Graph Attention Networks

no code implementations CVPR 2019 Peng Wang, Qi Wu, Jiewei Cao, Chunhua Shen, Lianli Gao, Anton Van Den Hengel

Being composed of node attention component and edge attention component, the proposed graph attention mechanism explicitly represents inter-object relationships, and properties with a flexibility and power impossible with competing approaches.

Graph Attention Object +2

Deep Template Matching for Offline Handwritten Chinese Character Recognition

no code implementations15 Nov 2018 Zhiyuan Li, Min Jin, Qi Wu, Huaxiang Lu

Just like its remarkable achievements in many computer vision tasks, the convolutional neural networks (CNN) provide an end-to-end solution in handwritten Chinese character recognition (HCCR) with great success.

Binary Classification Offline Handwritten Chinese Character Recognition +1

Goal-Oriented Visual Question Generation via Intermediate Rewards

no code implementations ECCV 2018 Jun-Jie Zhang, Qi Wu, Chunhua Shen, Jian Zhang, Jianfeng Lu, Anton Van Den Hengel

Despite significant progress in a variety of vision-and-language problems, developing a method capable of asking intelligent, goal-oriented questions about images is proven to be an inscrutable challenge.

Informativeness Question Generation +2

Connecting Language and Vision to Actions

no code implementations ACL 2018 Peter Anderson, Abhishek Das, Qi Wu

A long-term goal of AI research is to build intelligent agents that can see the rich visual environment around us, communicate this understanding in natural language to humans and other agents, and act in a physical or embodied environment.

Image Captioning Language Modelling +3

Topological Data Analysis Made Easy with the Topology ToolKit

no code implementations21 Jun 2018 Guillaume Favelier, Charles Gueunet, Attila Gyulassy, Julien Kitware, Joshua Levine, Jonas Lukasczyk, Daisuke Sakurai, Maxime Soler, Julien Tierny, Will Usher, Qi Wu

This tutorial presents topological methods for the analysis and visualization of scientific data from a user's perspective, with the Topology ToolKit (TTK), a recently released open-source library for topological data analysis.

Topological Data Analysis

Visual Grounding via Accumulated Attention

no code implementations CVPR 2018 Chaorui Deng, Qi Wu, Qingyao Wu, Fuyuan Hu, Fan Lyu, Mingkui Tan

There are three main challenges in VG: 1) what is the main focus in a query; 2) how to understand an image; 3) how to locate an object.