no code implementations • 22 Feb 2023 • Yikai Wang, Jianan Wang, Guansong Lu, Hang Xu, Zhenguo Li, Wei zhang, Yanwei Fu
In the image manipulation phase, SeMani adopts a generative model to synthesize new images conditioned on the entity-irrelevant regions and target text descriptions.
no code implementations • 2 Jan 2023 • Yikai Wang, Yanwei Fu, Xinwei Sun
A noisy training set usually leads to the degradation of the generalization and robustness of neural networks.
no code implementations • CVPR 2022 • Yikai Wang, TengQi Ye, Lele Cao, Wenbing Huang, Fuchun Sun, Fengxiang He, DaCheng Tao
Recently, there is a trend of leveraging multiple sources of input data, such as complementing the 3D point cloud with 2D images that often have richer color and fewer noises.
no code implementations • 13 Sep 2022 • ZiHao Wang, Qihao Liang, Kejun Zhang, Yuxing Wang, Chen Zhang, Pengfei Yu, Yongsheng Feng, Wenbo Liu, Yikai Wang, Yuntai Bao, Yiheng Yang
In this paper, we propose SongDriver, a real-time music accompaniment generation system without logical latency nor exposure bias.
no code implementations • 17 Jul 2022 • Ke Fan, Yikai Wang, Qian Yu, Da Li, Yanwei Fu
In contrast, this paper proposes a simple Test-time Linear Training (ETLT) method for OOD detection.
Out-of-Distribution Detection
Out of Distribution (OOD) Detection
4 code implementations • CVPR 2022 • Yikai Wang, Xinghao Chen, Lele Cao, Wenbing Huang, Fuchun Sun, Yunhe Wang
Many adaptations of transformers have emerged to address the single-modal vision tasks, where self-attention modules are stacked to handle input sources like images.
Ranked #1 on
Semantic Segmentation
on SUN-RGBD
1 code implementation • CVPR 2022 • Yikai Wang, Xinwei Sun, Yanwei Fu
Noisy training set usually leads to the degradation of generalization and robustness of neural networks.
1 code implementation • ICLR 2022 • Yinfeng Yu, Wenbing Huang, Fuchun Sun, Changan Chen, Yikai Wang, Xiaohong Liu
In this work, we design an acoustically complex environment in which, besides the target sound, there exists a sound attacker playing a zero-sum game with the agent.
1 code implementation • 4 Dec 2021 • Yikai Wang, Fuchun Sun, Wenbing Huang, Fengxiang He, DaCheng Tao
For the application of dense image prediction, the validity of CEN is tested by four different scenarios: multimodal fusion, cycle multimodal fusion, multitask learning, and multimodal multitask learning.
Ranked #15 on
Semantic Segmentation
on NYU Depth v2
1 code implementation • ICCV 2021 • Yikai Wang, Yi Yang, Fuchun Sun, Anbang Yao
In the low-bit quantization field, training Binary Neural Networks (BNNs) is the extreme solution to ease the deployment of deep models on resource-constrained devices, having the lowest storage cost and significantly cheaper bit-wise operations compared to 32-bit floating-point counterparts.
no code implementations • 29 Sep 2021 • Yikai Wang, Xinwei Sun, Yanwei Fu
Specifically, we re-purpose a sparse linear model with incidental parameters as a unified Relative Instance Credibility Inference (RICI) framework, which will detect and remove outliers in the forward pass of each mini-batch and use the remaining instances to train the network.
1 code implementation • 11 Aug 2021 • Yikai Wang, Fuchun Sun, Ming Lu, Anbang Yao
We propose a compact and effective framework to fuse multimodal features at multiple layers in a single network.
Ranked #22 on
Semantic Segmentation
on NYU Depth v2
1 code implementation • 11 Aug 2021 • Yikai Wang, Wenbing Huang, Bin Fang, Fuchun Sun, Chang Li
By contrast, EIP models the tactile sensor as a group of coordinated particles, and the elastic property is applied to regulate the deformation of particles during contact.
no code implementations • 1 Jan 2021 • Lujun Li, Yikai Wang, Anbang Yao, Yi Qian, Xiao Zhou, Ke He
In this paper, we present Explicit Connection Distillation (ECD), a new KD framework, which addresses the knowledge distillation problem in a novel perspective of bridging dense intermediate feature connections between a student network and its corresponding teacher generated automatically in the training, achieving knowledge transfer goal via direct cross-network layer-to-layer gradients propagation, without need to define complex distillation losses and assume a pre-trained teacher model to be available.
no code implementations • 30 Nov 2020 • Yikai Wang, Weijian Li
We found that by mapping different word embeddings into the joint component, sentiment performance can be greatly improved for the original word embeddings with lower performance.
no code implementations • 23 Nov 2020 • Yikai Wang, Wenbing Huang, Bin Fang, Fuchun Sun
At its core, EIP models the tactile sensor as a group of coordinated particles, and the elastic theory is applied to regulate the deformation of particles during the contact process.
1 code implementation • NeurIPS 2020 • Yikai Wang, Wenbing Huang, Fuchun Sun, Tingyang Xu, Yu Rong, Junzhou Huang
Deep multimodal fusion by using multiple sources of data for classification or regression has exhibited a clear advantage over the unimodal counterpart on various applications.
no code implementations • 19 Aug 2020 • Yikai Wang, Ying Guo
In this paper, we propose a novel blind source separation method with low-rank structure and uniform sparsity (LOCUS) as a fully data-driven decomposition method for network measures.
1 code implementation • ECCV 2020 • Yikai Wang, Fuchun Sun, Duo Li, Anbang Yao
We propose a general method to train a single convolutional neural network which is capable of switching image resolutions at inference.
2 code implementations • 15 Jul 2020 • Yikai Wang, Li Zhang, Yuan YAO, Yanwei Fu
We rank the credibility of pseudo-labeled instances along the regularization path of their corresponding incidental parameters, and the most trustworthy pseudo-labeled examples are preserved as the augmented labeled instances.
1 code implementation • CVPR 2020 • Yikai Wang, Chengming Xu, Chen Liu, Li Zhang, Yanwei Fu
To measure the credibility of each pseudo-labeled instance, we then propose to solve another linear regression hypothesis by increasing the sparsity of the incidental parameters and rank the pseudo-labeled instances with their sparsity degree.
no code implementations • 3 Nov 2019 • Yikai Wang, Liang Zhang, Quanyu Dai, Fuchun Sun, Bo Zhang, Yang He, Weipeng Yan, Yongjun Bao
In deep CTR models, exploiting users' historical data is essential for learning users' behaviors and interests.