no code implementations • 17 Apr 2025 • Shizhe Diao, Yu Yang, Yonggan Fu, Xin Dong, Dan Su, Markus Kliegl, Zijia Chen, Peter Belcak, Yoshi Suhara, Hongxu Yin, Mostofa Patwary, Yingyan, Lin, Jan Kautz, Pavlo Molchanov
We analyze the final data mixture, elucidating the characteristics of an optimal data mixture.
no code implementations • 12 Apr 2025 • Lingyou Zhou, Xin Dong, Kehai Qiu, Gang Yu, Jie Zhang, Jiliang Zhang
In this paper, we characterize the adaptive multiple path loss exponent (AMPLE) radio propagation model under urban macrocell (UMa) and urban microcell (UMi) scenarios from 0. 85-5 GHz using Ranplan Professional.
no code implementations • 17 Mar 2025 • Xin Dong, Rui Miao, Suyan Zhang, Shuaibing Jia, Leifeng Zhang, Yong Liang, Jianhua Zhang, Yi Zhun Zhu
Furthermore, most repositioning models are only used to complete the relationship matrix, and their practicality is poor when dealing with drug cold start problems.
no code implementations • 11 Feb 2025 • Penghao Lu, Xin Dong, Yuansheng Zhou, Lei Cheng, Chuan Yuan, Linjian Mo
To address the above problem, we propose a novel and general generative retrieval framework, namely Leverag- ing Document-Oriented Contrastive Learning in Generative Retrieval (DOGR), which leverages contrastive learning to improve generative retrieval tasks.
no code implementations • 13 Jan 2025 • Sujia Wang, Xiangwei Shen, Yansong Tang, Xin Dong, Wenjia Geng, Lei Chen
Repetitive action counting (RAC) aims to estimate the number of class-agnostic action occurrences in a video without exemplars.
no code implementations • 23 Dec 2024 • Ente Lin, Xujie Zhang, Fuwei Zhao, Yuxuan Luo, Xin Dong, Long Zeng, Xiaodan Liang
However, existing methods often face a dilemma: lightweight approaches, such as adapters, are prone to generate inconsistent textures; while finetune-based methods involve high training costs and struggle to maintain the generalization capabilities of pretrained diffusion models, limiting their performance across diverse scenarios.
no code implementations • 11 Dec 2024 • Xin Dong, Sen Jia, Hongyu Xiong
In this paper, we propose COEF-VQ, a novel cascaded MLLM framework for better video quality understanding on TikTok.
no code implementations • 20 Nov 2024 • Xin Dong, Yonggan Fu, Shizhe Diao, Wonmin Byeon, Zijia Chen, Ameya Sunil Mahabaleshwarkar, Shih-Yang Liu, Matthijs Van Keirsbilck, Min-Hung Chen, Yoshi Suhara, Yingyan Lin, Jan Kautz, Pavlo Molchanov
We propose Hymba, a family of small language models featuring a hybrid-head parallel architecture that integrates transformer attention mechanisms with state space models (SSMs) for enhanced efficiency.
no code implementations • 30 Sep 2024 • Jing Liu, Tianyi Zeng, Abdelkhalick Mohammad, Xin Dong, Dragos Axinte
This paper introduces a simple-structured, model-less fuzzy logic controller for the closed-loop control of continuum robots.
no code implementations • 1 Aug 2024 • Yuhang Li, Xin Dong, Chen Chen, Weiming Zhuang, Lingjuan Lyu
In computer vision, it is well-known that a lack of data diversity will impair model performance.
1 code implementation • 23 Jul 2024 • Shoaib Ahmed Siddiqui, Xin Dong, Greg Heinrich, Thomas Breuel, Jan Kautz, David Krueger, Pavlo Molchanov
Large Language Models (LLMs) are not only resource-intensive to train but even more costly to deploy in production.
no code implementations • 15 Jul 2024 • Xingzhi Zhou, Xin Dong, Chunhao Li, Yuning Bai, Yulong Xu, Ka Chun Cheung, Simon See, Xinpeng Song, Runshun Zhang, Xuezhong Zhou, Nevin L. Zhang
However, this task faces limitations due to the scarcity of high-quality clinical datasets and the complex relationship between symptoms and herbs.
no code implementations • 14 Jun 2024 • Hui Liu, Wenya Wang, Hao Sun, Chris Xing Tian, Chenqi Kong, Xin Dong, Haoliang Li
Large Language Models (LLMs) have demonstrated impressive in-context learning (ICL) capabilities from few-shot demonstration exemplars.
no code implementations • 28 May 2024 • Jun Zheng, Fuwei Zhao, Youjiang Xu, Xin Dong, Xiaodan Liang
To faithfully recover the clothing details, the extracted garment features are fused with the self-attention outputs of the denoising DiT and the ControlNet.
no code implementations • 1 May 2024 • Xujie Zhang, Ente Lin, Xiu Li, Yuxuan Luo, Michael Kampffmeyer, Xin Dong, Xiaodan Liang
Besides, to remove the segmentation dependency, MMTryon uses a parsing-free garment encoder and leverages a novel scalable data generation pipeline to convert existing VITON datasets to a form that allows MMTryon to be trained without requiring any explicit segmentation.
1 code implementation • CVPR 2024 • Lihua Jing, Rui Wang, Wenqi Ren, Xin Dong, Cong Zou
Adversarial patch attacks present a significant threat to real-world object detectors due to their practical feasibility.
no code implementations • 28 Mar 2024 • Yuhang Li, Xin Dong, Chen Chen, Jingtao Li, Yuxin Wen, Michael Spranger, Lingjuan Lyu
Synthetic image data generation represents a promising avenue for training deep learning models, particularly in the realm of transfer learning, where obtaining real images within a specific domain can be prohibitively expensive due to privacy and intellectual property considerations.
no code implementations • 23 Mar 2024 • Minzhou Pan, Zhenting Wang, Xin Dong, Vikash Sehwag, Lingjuan Lyu, Xue Lin
In this paper, we propose WaterMark Detection (WMD), the first invisible watermark detection method under a black-box and annotation-free setting.
no code implementations • 6 Dec 2023 • Xujie Zhang, Xiu Li, Michael Kampffmeyer, Xin Dong, Zhenyu Xie, Feida Zhu, Haoye Dong, Xiaodan Liang
Image-based Virtual Try-On (VITON) aims to transfer an in-shop garment image onto a target person.
1 code implementation • CVPR 2024 • Wenjie Zhao, Jia Li, Xin Dong, Yu Xiang, Yunhui Guo
Semantic segmentation models, while effective for in-distribution categories, face challenges in real-world deployment due to encountering out-of-distribution (OoD) objects.
no code implementations • 7 Oct 2023 • Tian Jin, Nolan Clement, Xin Dong, Vaishnavh Nagarajan, Michael Carbin, Jonathan Ragan-Kelley, Gintare Karolina Dziugaite
We study two natural scaling techniques -- weight pruning and simply training a smaller or larger model, which we refer to as dense scaling -- and their effects on two core capabilities of LLMs: (a) recalling facts presented during pre-training and (b) processing information presented in-context during inference.
no code implementations • 11 Aug 2023 • Xin Dong, Rui Wang, Siyuan Liang, Aishan Liu, Lihua Jing
As for the weak black-box scenario feasibility, we obverse that representations of the average feature in multiple face recognition models are similar, thus we propose to utilize the average feature via the crawled dataset from the Internet as the target to guide the generation, which is also agnostic to identities of unknown face recognition systems; in nature, the low-frequency perturbations are more visually perceptible by the human vision system.
2 code implementations • CVPR 2023 • Zhenyu Xie, Zaiyu Huang, Xin Dong, Fuwei Zhao, Haoye Dong, Xijin Zhang, Feida Zhu, Xiaodan Liang
Specifically, compared with the previous global warping mechanism, LFGP employs local flows to warp garments parts individually, and assembles the local warped results via the global garment parsing, resulting in reasonable warped parts and a semantic-correct intact garment even with challenging inputs. On the other hand, our DGT training strategy dynamically truncates the gradient in the overlap area and the warped garment is no more required to meet the boundary constraint, which effectively avoids the texture squeezing problem.
no code implementations • 31 Jan 2023 • Xin Dong, Ruize Wu, Chao Xiong, Hai Li, Lei Cheng, Yong He, Shiyou Qian, Jian Cao, Linjian Mo
GDOD decomposes gradients into task-shared and task-conflict components explicitly and adopts a general update rule for avoiding interference across all task gradients.
no code implementations • 27 Jul 2022 • Zhenyu Xie, Zaiyu Huang, Fuwei Zhao, Haoye Dong, Michael Kampffmeyer, Xin Dong, Feida Zhu, Xiaodan Liang
In this work, we take a step forwards to explore versatile virtual try-on solutions, which we argue should possess three main properties, namely, they should support unsupervised training, arbitrary garment categories, and controllable garment editing.
no code implementations • 19 Jul 2022 • Xin Dong, Sai Qian Zhang, Ang Li, H. T. Kung
Federated Learning aims at training a global model from multiple decentralized devices (i. e. clients) without exchanging their private local data.
no code implementations • 1 Jul 2022 • Zhongnan Qu, Syed Shakib Sarwar, Xin Dong, Yuecheng Li, Ekin Sumbul, Barbara De Salvo
The limited and dynamically varied resources on edge devices motivate us to deploy an optimized deep neural network that can adapt its sub-networks to fit in different resource constraints.
no code implementations • 25 May 2022 • Zhi Chen, Jijia Bao, Lu Chen, Yuncong Liu, Da Ma, Bei Chen, Mengyue Wu, Su Zhu, Xin Dong, Fujiang Ge, Qingliang Miao, Jian-Guang Lou, Kai Yu
In this work, we aim to build a unified dialogue foundation model (DFM) which can be used to solve massive diverse dialogue tasks.
1 code implementation • 6 May 2022 • Yuhang Li, Shikuang Deng, Xin Dong, Shi Gu
We demonstrate that our method can handle the SNN conversion with batch normalization layers and effectively preserve the high accuracy even in 32 time steps.
no code implementations • CVPR 2022 • Xin Dong, Barbara De Salvo, Meng Li, Chiao Liu, Zhongnan Qu, H. T. Kung, Ziyun Li
We design deep neural networks (DNNs) and corresponding networks' splittings to distribute DNNs' workload to camera sensors and a centralized aggregator on head mounted devices to meet system performance targets in inference accuracy and latency under the given hardware resource constraints.
no code implementations • CVPR 2022 • Xin Dong, Fuwei Zhao, Zhenyu Xie, Xijin Zhang, Daniel K. Du, Min Zheng, Xiang Long, Xiaodan Liang, Jianchao Yang
While significant progress has been made in garment transfer, one of the most applicable directions of human-centric image generation, existing works overlook the in-the-wild imagery, presenting severe garment-person misalignment as well as noticeable degradation in fine texture details.
1 code implementation • 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) 2021 • Xin Dong, Yi Zheng, Zixin Shu, Kai Chang, Dengying Yan, Jianan Xia, Qiang Zhu, Kunyu Zhong, Xinyan Wang, Kuo Yang, Xuezhong Zhou
In addition, the comprehensive experiments of TCMPR with different hyper parameters (i. e., feature embedding, feature dimension and feature fusion) that demonstrates that our method has high performance on TCM prescription recommendation and potentially promote clinical diagnosis and treatment of TCM precision medicine.
no code implementations • 18 Sep 2021 • Yuguang Yang, Yu Pan, Xin Dong, Minqiang Xu
Second, we design a novel model inference scheme based on RepVGG which can efficiently improve the QbE search quality.
no code implementations • ACL 2021 • Xin Dong, Yaxin Zhu, Zuohui Fu, Dongkuan Xu, Gerard de Melo
Due to recent pretrained multilingual representation models, it has become feasible to exploit labeled data from one language to train a cross-lingual model that can then be applied to multiple new languages.
Cross-Lingual Natural Language Inference
Data Augmentation
+1
no code implementations • 13 Jul 2021 • Xin Dong, Hongxu Yin, Jose M. Alvarez, Jan Kautz, Pavlo Molchanov, H. T. Kung
Prior works usually assume that SC offers privacy benefits as only intermediate features, instead of private data, are shared from devices to the cloud.
1 code implementation • 13 Jun 2021 • Yuhang Li, Shikuang Deng, Xin Dong, Ruihao Gong, Shi Gu
Moreover, our calibration algorithm can produce SNN with state-of-the-art architecture on the large-scale ImageNet dataset, including MobileNet and RegNet.
no code implementations • CVPR 2022 • Xin Dong, Junfeng Guo, Ang Li, Wei-Te Ting, Cong Liu, H. T. Kung
Based upon this observation, we propose a novel metric called Neural Mean Discrepancy (NMD), which compares neural means of the input examples and training data.
no code implementations • 16 Feb 2021 • John Arrington, Reynier Cruz-Torres, Winston DeGraw, Xin Dong, Leo Greiner, Samuel Heppelmann, Barbara Jacak, Yuanjing Ji, Matthew Kelsey, Spencer R. Klein, Yue Shi Lai, Grazyna Odyniec, Sooraj Radhakrishnan, Ernst Sichtermann, Youqi Son, Fernando Torales Acosta, Lei Xia, Nu Xu, Feng Yuan, Yuxiang Zhao
The proposed electron-ion collider has a rich physics program to study the internal structure of protons and heavy nuclei.
Nuclear Experiment Nuclear Theory
no code implementations • COLING 2020 • Sm Mazharul Islam, Xin Dong, Gerard de Melo
Sentiment analysis is an area of substantial relevance both in industry and in academia, including for instance in social studies.
no code implementations • ICCV 2021 • Yuhang Li, Feng Zhu, Ruihao Gong, Mingzhu Shen, Xin Dong, Fengwei Yu, Shaoqing Lu, Shi Gu
However, the inversion process only utilizes biased feature statistics stored in one model and is from low-dimension to high-dimension.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Wen Tai, H. T. Kung, Xin Dong, Marcus Comiter, Chang-Fu Kuo
We introduce exBERT, a training method to extend BERT pre-trained models from a general domain to a new pre-trained model for a specific domain with a new additive vocabulary under constrained training resources (i. e., constrained computation and data).
no code implementations • 29 Jul 2020 • Xin Dong, Yaxin Zhu, Yupeng Zhang, Zuohui Fu, Dongkuan Xu, Sen yang, Gerard de Melo
The resulting model then serves as a teacher to induce labels for unlabeled target language samples that can be used during further adversarial training, allowing us to gradually adapt our model to the target language.
no code implementations • 17 Mar 2020 • Yuhang Li, Wei Wang, Haoli Bai, Ruihao Gong, Xin Dong, Fengwei Yu
Network quantization has rapidly become one of the most widely used methods to compress and accelerate deep neural networks.
no code implementations • 29 Jan 2020 • Zuohui Fu, Yikun Xian, Shijie Geng, Yingqiang Ge, Yuting Wang, Xin Dong, Guang Wang, Gerard de Melo
A number of cross-lingual transfer learning approaches based on neural networks have been proposed for the case when large amounts of parallel text are at our disposal.
no code implementations • 18 Dec 2019 • Xin Dong, Jingchao Ni, Wei Cheng, Zhengzhang Chen, Bo Zong, Dongjin Song, Yanchi Liu, Haifeng Chen, Gerard de Melo
In practice, however, these two sets of reviews are notably different: users' reviews reflect a variety of items that they have bought and are hence very heterogeneous in their topics, while an item's reviews pertain only to that single item and are thus topically homogeneous.
no code implementations • 4 Dec 2019 • Yuhang Li, Xin Dong, Sai Qian Zhang, Haoli Bai, Yuanpeng Chen, Wei Wang
We first bring up three omitted issues in extremely low-bit networks: the squashing range of quantized values; the gradient vanishing during backpropagation and the unexploited hardware acceleration of ternary networks.
no code implementations • IJCNLP 2019 • Xin Dong, Gerard de Melo
Based on massive amounts of data, recent pretrained contextual representation models have made significant strides in advancing a number of different English NLP tasks.
1 code implementation • ICLR 2020 • Yuhang Li, Xin Dong, Wei Wang
We propose Additive Powers-of-Two~(APoT) quantization, an efficient non-uniform quantization scheme for the bell-shaped and long-tailed distribution of weights and activations in neural networks.
no code implementations • 1 May 2019 • Bradley McDanel, Sai Qian Zhang, H. T. Kung, Xin Dong
A highlight of our full-stack approach which attributes to the achieved high energy efficiency is an efficient Selector-Accumulator (SAC) architecture for implementing the multiplier-accumulator (MAC) operation present in any digital CNN hardware.
no code implementations • 11 Dec 2018 • Yinghao Xu, Xin Dong, Yudian Li, Hao Su
To reduce memory footprint and run-time latency, techniques such as neural network pruning and binarization have been explored separately.
no code implementations • 24 Nov 2018 • Song-Hai Zhang, Zhengping Zhou, Bin Liu, Xin Dong, Dun Liang, Peter Hall, Shi-Min Hu
In this work, we propose a novel topic consisting of two dual tasks: 1) given a scene, recommend objects to insert, 2) given an object category, retrieve suitable background scenes.
no code implementations • ACL 2018 • Xin Dong, Gerard de Melo
Deep convolutional neural networks excel at sentiment polarity classification, but tend to require substantial amounts of training data, which moreover differs quite significantly between domains.
Ranked #75 on
Sentiment Analysis
on SST-2 Binary classification
1 code implementation • CVPR 2019 • Shilin Zhu, Xin Dong, Hao Su
Binary neural networks (BNN) have been studied extensively since they run dramatically faster at lower memory and power consumption than floating-point networks, thanks to the efficiency of bit operations.
7 code implementations • CVPR 2019 • Song-Hai Zhang, Rui-Long Li, Xin Dong, Paul L. Rosin, Zixi Cai, Han Xi, Dingcheng Yang, Hao-Zhi Huang, Shi-Min Hu
We demonstrate that our pose-based framework can achieve better accuracy than the state-of-art detection-based approach on the human instance segmentation problem, and can moreover better handle occlusion.
Ranked #1 on
Pose-Based Human Instance Segmentation
on OCHuman
2 code implementations • NeurIPS 2017 • Xin Dong, Shangyu Chen, Sinno Jialin Pan
How to develop slim and accurate deep neural networks has become crucial for real- world applications, especially for those employed in embedded systems.