no code implementations • ICML 2018 • Jingwei Zhuo, Chang Liu, Jiaxin Shi, Jun Zhu, Ning Chen, Bo Zhang
Stein variational gradient descent (SVGD) is a recently proposed particle-based Bayesian inference method, which has attracted a lot of interest due to its remarkable approximation ability and particle efficiency compared to traditional variational inference and Markov Chain Monte Carlo methods.
no code implementations • 16 May 2018 • Chang Liu, Xiangrui Zeng, Kaiwen Wang, Qiang Guo, Min Xu
Cellular Electron Cryo-Tomography (CECT) is a powerful 3D imaging tool for studying the native structure and organization of macromolecules inside single cells.
no code implementations • CVPR 2018 • Xiaojun Xu, Xinyun Chen, Chang Liu, Anna Rohrbach, Trevor Darrell, Dawn Song
Our work sheds new light on understanding adversarial attacks on vision systems which have a language component and shows that attention, bounding box localization, and compositional internal structures are vulnerable to adversarial attacks.
no code implementations • 5 Mar 2018 • Chang Liu, Xiaolin Wu, Xiao Shu
All existing image enhancement methods, such as HDR tone mapping, cannot recover A/D quantization losses due to insufficient or excessive lighting, (underflow and overflow problems).
no code implementations • ICLR 2018 • Xinyun Chen, Chang Liu, Dawn Song
In our evaluation, we show that using our novel approach, neural parsing programs can be learned to achieve 100% test accuracy on test inputs that are 500x longer than the training samples.
no code implementations • 22 Feb 2018 • Nicholas Carlini, Chang Liu, Úlfar Erlingsson, Jernej Kos, Dawn Song
This paper describes a testing methodology for quantitatively assessing the risk that rare or unique training-data sequences are unintentionally memorized by generative sequence models---a common type of machine-learning model.
no code implementations • 20 Feb 2018 • Qiuyuan Huang, Li Deng, Dapeng Wu, Chang Liu, Xiaodong He
This paper proposes a new architecture - Attentive Tensor Product Learning (ATPL) - to represent grammatical structures in deep learning models.
no code implementations • 14 Feb 2018 • Jaime F. Fisac, Chang Liu, Jessica B. Hamrick, S. Shankar Sastry, J. Karl Hedrick, Thomas L. Griffiths, Anca D. Dragan
We introduce $t$-\ACty{}: a measure that quantifies the accuracy and confidence with which human observers can predict the remaining robot plan from the overall task goal and the observed initial $t$ actions in the plan.
no code implementations • 12 Feb 2018 • Chang Liu, Xiangrui Zeng, Ruogu Lin, Xiaodan Liang, Zachary Freyberg, Eric Xing, Min Xu
Cellular Electron Cryo-Tomography (CECT) is a powerful imaging technique for the 3D visualization of cellular structure and organization at submolecular resolution.
no code implementations • ICLR 2018 • Xinyun Chen, Chang Liu, Dawn Song
We observe that program translation is a modular procedure, in which a sub-tree of the source tree is translated into the corresponding target sub-tree at each step.
no code implementations • 6 Feb 2018 • Chang Liu, Jessica B. Hamrick, Jaime F. Fisac, Anca D. Dragan, J. Karl Hedrick, S. Shankar Sastry, Thomas L. Griffiths
The study of human-robot interaction is fundamental to the design and use of robotics in real-world applications.
no code implementations • 20 Jul 2017 • Jaime F. Fisac, Monica A. Gates, Jessica B. Hamrick, Chang Liu, Dylan Hadfield-Menell, Malayandi Palaniappan, Dhruv Malik, S. Shankar Sastry, Thomas L. Griffiths, Anca D. Dragan
In robotics, value alignment is key to the design of collaborative robots that can integrate into human workflows, successfully inferring and adapting to their users' objectives as they go.
no code implementations • 19 Jan 2018 • He-Liang Huang, Xi-Lin Wang, Peter P. Rohde, Yi-Han Luo, You-Wei Zhao, Chang Liu, Li Li, Nai-Le Liu, Chao-Yang Lu, Jian-Wei Pan
Topological data analysis offers a robust way to extract useful information from noisy, unstructured data by identifying its underlying structure.
no code implementations • 11 Sep 2017 • Sattar Vakili, Qing Zhao, Chang Liu, Chen-Nee Chuah
We consider the problem of detecting a few targets among a large number of hierarchical data streams.
no code implementations • 18 Feb 2017 • Chang Liu, Fuchun Sun, Changhu Wang, Feng Wang, Alan Yuille
In this way, the sequential representation of an image can be naturally translated to a sequence of words, as the target sequence of the RNN model.
no code implementations • 23 Jun 2017 • Wenshuo Wang, Chang Liu, Ding Zhao
For projects that cost millions of dollars, it is critical to determine the right amount of data needed.
no code implementations • 28 Mar 2017 • Shun Yang, Wenshuo Wang, Chang Liu, Kevin Deng, J. Karl Hedrick
We collect a large set of data using The Open Racing Car Simulator (TORCS) and classify the image features into three categories (sky-related, roadside-related, and road-related features). We then design two experimental frameworks to investigate the importance of each single feature for training a CNN controller. The first framework uses the training data with all three features included to train a controller, which is then tested with data that has one feature removed to evaluate the feature's effects.
no code implementations • 11 Dec 2016 • Xiaolin Wu, Xi Zhang, Chang Liu
This article is a sequel to our earlier work [25].
no code implementations • NeurIPS 2016 • Xinyun Chen, Chang Liu, Richard Shin, Dawn Song, Mingcheng Chen
Automatic translation from natural language descriptions into programs is a longstanding challenging problem.
no code implementations • 7 Aug 2016 • Chang Liu, Bo Li, Yevgeniy Vorobeychik, Alina Oprea
The effectiveness of supervised learning techniques has made them ubiquitous in research and practice.
no code implementations • 20 Oct 2014 • Chang Liu, Yi Xu
We propose a filter method for unsupervised feature selection which is based on the Confidence Machine.
no code implementations • 29 Oct 2014 • Xudong Liu, Bin Zhang, Ting Zhang, Chang Liu
Rating Prediction is a basic problem in Recommender System, and one of the most widely used method is Factorization Machines(FM).
no code implementations • ECCV 2018 • Chang Liu, Wei Ke, Fei Qin, Qixiang Ye
Hinted by this, we formalize a Linear Span framework, and propose Linear Span Network (LSN) modified by Linear Span Units (LSUs), which minimize the reconstruction error of convolutional network.
no code implementations • 12 Nov 2018 • Sophia Collet, Robert Dadashi, Zahi N. Karam, Chang Liu, Parinaz Sobhani, Yevgeniy Vahlis, Ji Chao Zhang
In this work, two approaches for private model aggregation are proposed that enable the transfer of knowledge from existing models trained on other companies' datasets to a new company with limited labeled data while protecting each client company's underlying individual sensitive information.
no code implementations • 21 Dec 2018 • Chang Liu, Zhaowei Shang, Anyong Qin
To address this issue, here we propose a novel deep residual learning model that combines the dilated residual convolution and multi-scale convolution groups.
no code implementations • NeurIPS 2016 • Chang Liu, Jun Zhu, Yang song
We propose two stochastic gradient MCMC methods for sampling from Bayesian posterior distributions defined on Riemann manifolds with a known geodesic flow, e. g. hyperspheres.
no code implementations • ICLR 2019 • Xinyun Chen, Chang Liu, Dawn Song
Most existing neural program synthesis approaches employ an encoder-decoder architecture, which uses an encoder to compute the embedding of the given input-output examples, as well as a decoder to generate the program from the embedding following a given syntax.
no code implementations • 29 Mar 2019 • Weiwei Zong, Joon Lee, Chang Liu, Eric Carver, Aharon Feldman, Branislava Janic, Mohamed Elshaikh, Milan Pantelic, David Hearshen, Indrin Chetty, Benjamin Movsas, Ning Wen
Deep learning models have had a great success in disease classifications using large data pools of skin cancer images or lung X-rays.
no code implementations • 4 Apr 2019 • Zhenzhen Dai, Eric Carver, Chang Liu, Joon Lee, Aharon Feldman, Weiwei Zong, Milan Pantelic, Mohamed Elshaikh, Ning Wen
Prostate cancer (PCa) is the most common cancer in men in the United States.
no code implementations • CVPR 2019 • Chang Liu, Fang Wan, Wei Ke, Zhuowei Xiao, Yuan Yao, Xiaosong Zhang, Qixiang Ye
The weight sharing scheme and spatial pooling operations in Convolutional Neural Networks (CNNs) introduce semantic correlation to neighboring pixels on feature maps and therefore deteriorate their pixel-wise classification performance.
no code implementations • WS 2019 • Xinze Guo, Chang Liu, Xiaolong Li, Yiran Wang, Guoliang Li, Feng Wang, Zhitao Xu, Liuyi Yang, Li Ma, Changliang Li
This paper describes the Kingsoft AI Lab{'}s submission to the WMT2019 news translation shared task.
no code implementations • 27 Aug 2019 • Eddie S. J. Du, Chang Liu, David H. Wayne
The ability to accurately predict the fit of fashion items and recommend the correct size is key to reducing merchandise returns in e-commerce.
no code implementations • 30 Aug 2019 • Chang Liu, Yi Dong, Han Yu, Zhiqi Shen, Zhanning Gao, Pan Wang, Changgong Zhang, Peiran Ren, Xuansong Xie, Lizhen Cui, Chunyan Miao
Video contents have become a critical tool for promoting products in E-commerce.
no code implementations • 13 Sep 2019 • Xianlong Zeng, Soheil Moosavinasab, En-Ju D Lin, Simon Lin, Razvan Bunescu, Chang Liu
Efficient representation of patients is very important in the healthcare domain and can help with many tasks such as medical risk prediction.
no code implementations • 25 Sep 2019 • Chang Liu, Yanan Xu, Yanmin Zhu
In this paper, we study the problem of inferring fine-grained bike demands anywhere in a new city before the deployment of bikes.
no code implementations • 5 Oct 2019 • Pengyu Cheng, Chang Liu, Chunyuan Li, Dinghan Shen, Ricardo Henao, Lawrence Carin
The Straight-Through (ST) estimator is a widely used technique for back-propagating gradients through discrete random variables.
no code implementations • 16 Oct 2019 • Wenqiang Xu, Yanjun Fu, Yuchen Luo, Chang Liu, Cewu Lu
Fine-grained recognition task deals with sub-category classification problem, which is important for real-world applications.
no code implementations • 1 Nov 2019 • Xishan Zhang, Shaoli Liu, Rui Zhang, Chang Liu, Di Huang, Shiyi Zhou, Jiaming Guo, Yu Kang, Qi Guo, Zidong Du, Yunji Chen
Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers.
no code implementations • 10 Mar 2019 • Xing Hu, Ling Liang, Lei Deng, Shuangchen Li, Xinfeng Xie, Yu Ji, Yufei Ding, Chang Liu, Timothy Sherwood, Yuan Xie
As neural networks continue their reach into nearly every aspect of software operations, the details of those networks become an increasingly sensitive subject.
Cryptography and Security Hardware Architecture
no code implementations • 23 Dec 2019 • Minoru Kusaba, Chang Liu, Yukinori Koyama, Kiyoyuki Terakura, Ryo Yoshida
In 1869, the first draft of the periodic table was published by Russian chemist Dmitri Mendeleev.
no code implementations • 3 Feb 2020 • Di Huang, Xishan Zhang, Rui Zhang, Tian Zhi, Deyuan He, Jiaming Guo, Chang Liu, Qi Guo, Zidong Du, Shaoli Liu, Tianshi Chen, Yunji Chen
In this paper, we propose a novel Decomposable Winograd Method (DWM), which breaks through the limitation of original Winograd's minimal filtering algorithm to a wide and general convolutions.
no code implementations • 28 Feb 2020 • Yao Mu, Shengbo Eben Li, Chang Liu, Qi Sun, Bingbing Nie, Bo Cheng, Baiyu Peng
This paper presents a mixed reinforcement learning (mixed RL) algorithm by simultaneously using dual representations of environmental dynamics to search the optimal policy with the purpose of improving both learning accuracy and training speed.
no code implementations • 19 Apr 2020 • Chao Qu, Hui Li, Chang Liu, Junwu Xiong, James Zhang, Wei Chu, Weiqiang Wang, Yuan Qi, Le Song
We propose a \emph{collaborative} multi-agent reinforcement learning algorithm named variational policy propagation (VPP) to learn a \emph{joint} policy through the interactions over agents.
Multi-agent Reinforcement Learning reinforcement-learning +2
no code implementations • 12 Dec 2018 • Zhe Li, Caiwen Ding, Siyue Wang, Wujie Wen, Youwei Zhuo, Chang Liu, Qinru Qiu, Wenyao Xu, Xue Lin, Xuehai Qian, Yanzhi Wang
It is a challenging task to have real-time, efficient, and accurate hardware RNN implementations because of the high sensitivity to imprecision accumulation and the requirement of special activation function implementations.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
no code implementations • CVPR 2020 • Gaurav Mittal, Chang Liu, Nikolaos Karianakis, Victor Fragoso, Mei Chen, Yun Fu
To reduce HPO time, we present HyperSTAR (System for Task Aware Hyperparameter Recommendation), a task-aware method to warm-start HPO for deep neural networks.
no code implementations • 26 May 2020 • Lingbo Yang, Pan Wang, Chang Liu, Zhanning Gao, Peiran Ren, Xinfeng Zhang, Shanshe Wang, Siwei Ma, Xian-Sheng Hua, Wen Gao
Human pose transfer (HPT) is an emerging research topic with huge potential in fashion design, media production, online advertising and virtual reality.
no code implementations • 8 May 2020 • Hao Liu, Yan Xu, Jiasheng Wang, Ju Jing, Chang Liu, Jason T. L. Wang, Haimin Wang
By learning the latent patterns in the training data prepared by the physics-based ME tool, the proposed CNN method is able to infer vector magnetic fields from the Stokes profiles of GST/NIRIS.
Solar and Stellar Astrophysics
no code implementations • 4 Jun 2020 • Weihao Jiang, Zhaozhi Xie, Yaoyi Li, Chang Liu, Hongtao Lu
Many of these applications need to perform a real-time and efficient prediction for semantic segmentation with a light-weighted network.
no code implementations • 22 Jun 2020 • Gelu Nita, Manolis Georgoulis, Irina Kitiashvili, Viacheslav Sadykov, Enrico Camporeale, Alexander Kosovichev, Haimin Wang, Vincent Oria, Jason Wang, Rafal Angryk, Berkay Aydin, Azim Ahmadzadeh, Xiaoli Bai, Timothy Bastian, Soukaina Filali Boubrahimi, Bin Chen, Alisdair Davey, Sheldon Fereira, Gregory Fleishman, Dale Gary, Andrew Gerrard, Gregory Hellbourg, Katherine Herbert, Jack Ireland, Egor Illarionov, Natsuha Kuroda, Qin Li, Chang Liu, Yuexin Liu, Hyomin Kim, Dustin Kempton, Ruizhe Ma, Petrus Martens, Ryan McGranaghan, Edward Semones, John Stefan, Andrey Stejko, Yaireska Collado-Vega, Meiqi Wang, Yan Xu, Sijie Yu
The authors of this white paper met on 16-17 January 2020 at the New Jersey Institute of Technology, Newark, NJ, for a 2-day workshop that brought together a group of heliophysicists, data providers, expert modelers, and computer/data scientists.
no code implementations • 22 Jun 2020 • Yaolong Wang, Mingqing Xiao, Chang Liu, Shuxin Zheng, Tie-Yan Liu
Specifically, ILC introduces an invertible encoding module to replace the encoder-decoder structure to produce the low dimensional informative latent representation, meanwhile, transform the lost information into an auxiliary latent variable that won't be further coded or stored.
no code implementations • 20 Jun 2020 • Lixin Fan, Kam Woh Ng, Ce Ju, Tianyu Zhang, Chang Liu, Chee Seng Chan, Qiang Yang
This paper investigates capabilities of Privacy-Preserving Deep Learning (PPDL) mechanisms against various forms of privacy attacks.
no code implementations • ICML 2020 • Michael Zhu, Chang Liu, Jun Zhu
Particle-based Variational Inference methods (ParVIs), like Stein Variational Gradient Descent, are nonparametric variational inference methods that optimize a set of particles to best approximate a target distribution.
no code implementations • 11 Sep 2020 • Jianan Li, Jimei Yang, Jianming Zhang, Chang Liu, Christina Wang, Tingfa Xu
In this paper, we introduce Attribute-conditioned Layout GAN to incorporate the attributes of design elements for graphic layout generation by forcing both the generator and the discriminator to meet attribute conditions.
no code implementations • 17 Sep 2020 • Chang Liu, Huichu Zhang, Wei-Nan Zhang, Guanjie Zheng, Yong Yu
The heavy traffic congestion problem has always been a concern for modern cities.
no code implementations • 1 Jan 2021 • Chang Liu, Kai Li, Yun Fu
Unsupervised domain adaptation (UDA) is to make predictions for unlabeled data in a target domain with labeled data from source domain available.
no code implementations • 4 Nov 2020 • Xinwei Sun, Botong Wu, Xiangyu Zheng, Chang Liu, Wei Chen, Tao Qin, Tie-Yan Liu
To avoid spurious correlation, we propose a Latent Causal Invariance Model (LaCIM) which pursues causal prediction.
no code implementations • 10 Nov 2020 • Chang Liu, Wenzhong Yan, Ankur Mehta
Based on an equivalent plate model, we develop and validate analytical formulas for the behavioral specifications of OADLC mechanisms; the analytical formulas can be described as expressions of design parameters.
Robotics
no code implementations • 19 Nov 2020 • Yuanqiang Cai, Chang Liu, Weiqiang Wang, Qixiang Ye
With only bounding-box annotations in the spatial domain, existing video scene text detection (VSTD) benchmarks lack temporal relation of text instances among video frames, which hinders the development of video text-related applications.
no code implementations • 2 Dec 2020 • Tangqing Cao, Wenqi Guo, Wang Lu, Yunfei Xue, Wenjun Lu, Jing Su, Christian H. Liebscher, Chang Liu, Gerhard Dehm
Such a softening behavior can be related to the interaction of dislocations with short-range clustering.
Materials Science
no code implementations • 11 Sep 2020 • Chang Liu, Zhiqiang Wei, Derrick Wing Kwan Ng, Jinhong Yuan, Ying-Chang Liang
To eliminate the requirement of channel estimation and to improve the system performance, in this paper, we adopt a deep transfer learning (DTL) approach to implicitly extract the features of channel and directly recover tag symbols.
no code implementations • 16 Sep 2020 • Chang Liu, Weijie Yuan, Zhiqiang Wei, Xuemeng Liu, Derrick Wing Kwan Ng
Unmanned aerial vehicle (UAV)-assisted communication becomes a promising technique to realize the beyond fifth generation (5G) wireless networks, due to the high mobility and maneuverability of UAVs which can adapt to heterogeneous requirements of different applications.
no code implementations • 10 Nov 2020 • Chang Liu, Xuemeng Liu, Zhiqiang Wei, Derrick Wing Kwan Ng, Jinhong Yuan, Ying-Chang Liang
Existing tag signal detection algorithms inevitably suffer from a high bit error rate (BER) due to the difficulties in estimating the channel state information (CSI).
no code implementations • 7 Dec 2020 • Chang Liu, Yixing Huang, Joscha Maier, Laura Klein, Marc Kachelrieß, Andreas Maier
For organ-specific AEC, a preliminary CT reconstruction is necessary to estimate organ shapes for dose optimization, where only a few projections are allowed for real-time reconstruction.
no code implementations • 4 Jan 2021 • Ya'nan Wang, Zhuqing Jiang, Chang Liu, Kai Li, Aidong Men, Haiying Wang
This paper proposes a neural network for multi-level low-light image enhancement, which is user-friendly to meet various requirements by selecting different images as brightness reference.
no code implementations • 6 Jan 2021 • Bernhard Kliem, Jeongwoo Lee, Rui Liu, Stephen M. White, Chang Liu, Satoshi Masuda
We present evidence that a magnetic flux rope was formed before a coronal mass ejection (CME) and its associated long-duration flare during a pair of preceding confined eruptions and associated impulsive flares in a compound event in NOAA Active Region 12371.
Solar and Stellar Astrophysics
no code implementations • 20 Jan 2021 • Zhuqing Jiang, Chang Liu, Ya'nan Wang, Kai Li, Aidong Men, Haiying Wang, Haiyong Luo
With the goal of tuning up the brightness, low-light image enhancement enjoys numerous applications, such as surveillance, remote sensing and computational photography.
no code implementations • 22 Jan 2021 • Chang Liu, Henghui Ding, Xudong Jiang
In this paper, we argue that recovering these microscopic details relies on low-level but high-definition texture features.
no code implementations • 5 Mar 2021 • Chang Liu, Xiaoguang Li, Guohao Cai, Zhenhua Dong, Hong Zhu, Lifeng Shang
It is still an open question to leverage various types of information under the BERT framework.
no code implementations • 3 Mar 2021 • Jindong Wang, Wenjie Feng, Chang Liu, Chaohui Yu, Mingxuan Du, Renjun Xu, Tao Qin, Tie-Yan Liu
Being expensive and time-consuming to collect massive COVID-19 image samples to train deep classification models, transfer learning is a promising approach by transferring knowledge from the abundant typical pneumonia datasets for COVID-19 image classification.
no code implementations • 6 Mar 2021 • Jeremy Beauchamp, Razvan Bunescu, Cindy Marling, Zhongen Li, Chang Liu
In this work, we invert the "what-if" scenario and introduce a similar architecture based on chaining two LSTMs that can be trained to make either insulin or carbohydrate recommendations aimed at reaching a desired BG level in the future.
no code implementations • 16 Mar 2021 • Chang Liu, Lixin Fan, Kam Woh Ng, Yilun Jin, Ce Ju, Tianyu Zhang, Chee Seng Chan, Qiang Yang
This paper proposes a novel ternary hash encoding for learning to hash methods, which provides a principled more efficient coding scheme with performances better than those of the state-of-the-art binary hashing counterparts.
no code implementations • 17 Mar 2021 • Juntao Li, Chang Liu, Chongyang Tao, Zhangming Chan, Dongyan Zhao, Min Zhang, Rui Yan
To fill the gap between these up-to-date methods and the real-world applications, we incorporate user-specific dialogue history into the response selection and propose a personalized hybrid matching network (PHMN).
no code implementations • 22 Mar 2021 • Hua Wei, Chacha Chen, Chang Liu, Guanjie Zheng, Zhenhui Li
Simulation of the real-world traffic can be used to help validate the transportation policies.
no code implementations • 22 Mar 2021 • Chang Liu, Xiaojuan Qi, Edmund Lam, Ngai Wong
The neuromorphic event cameras, which capture the optical changes of a scene, have drawn increasing attention due to their high speed and low power consumption.
no code implementations • 8 Oct 2020 • Yunfan Jiang, Jingjing Si, Rui Zhang, Godwin Enemali, Bin Zhou, Hugh McCann, Chang Liu
Chemical Species Tomography (CST) has been widely used for in situ imaging of critical parameters, e. g. species concentration and temperature, in reactive flows.
no code implementations • 6 Apr 2021 • Boyu Yang, Mingbao Lin, Binghao Liu, Mengying Fu, Chang Liu, Rongrong Ji, Qixiang Ye
By tentatively expanding network nodes, LEC-Net enlarges the representation capacity of features, alleviating feature drift of old network from the perspective of model regularization.
no code implementations • 29 Nov 2020 • Yan He, Jifang Qiu, Chang Liu, Yue Liu, Jian Wu
The latest theoretical advances in the field of unlimited sampling framework (USF) show the potential to avoid clipping problems of analog-to-digital converters (ADC).
no code implementations • 20 Nov 2020 • Godwin Enemali, Rui Zhang, Hugh McCann, Chang Liu
Although a fully parallel data acquisition (DAQ) and signal processing system can achieve these functionalities with maximised temporal response, it leads to a highly complex, expensive and power-consuming instrumentation system with high potential for inconsistency between the sampled beams due to the electronics alone.
no code implementations • 4 Sep 2020 • Chang Liu, Jiahui Sun, Haiming Jin, Meng Ai, Qun Li, Cheng Zhang, Kehua Sheng, Guobin Wu, XiaoHu Qie, Xinbing Wang
Thus, in this paper, we exploit adaptive dispatching intervals to boost the platform's profit under a guarantee of the maximum passenger waiting time.
no code implementations • 26 Apr 2021 • Jie Chen, Jie Liu, Chang Liu, Jian Zhang, Bing Han
To overcome this issue and to further improve the recognition performance, we adopt a deep learning approach for underwater target recognition and propose a LOFAR spectrum enhancement (LSE)-based underwater target recognition scheme, which consists of preprocessing, offline training, and online testing.
no code implementations • 18 May 2021 • Chang Liu, Guanjie Zheng, Zhenhui Li
Therefore, in this paper, we propose to learn the human routing model, which is one of the most essential part in the traffic simulator.
no code implementations • 10 Feb 2021 • Rui Zhang, Jingjing Si, Godwin Enemali, Yong Bao, Chang Liu
The proposed scheme was both numerically and experimentally validated using a CST sensor with 32 laser beams using a variety of computational tomographic algorithms.
no code implementations • 5 Jun 2021 • Jingjing Si, Guoliang Li, Yinbo Cheng, Rui Zhang, Godwin Enemali, Chang Liu
As an in situ combustion diagnostic tool, Tunable Diode Laser Absorption Spectroscopy (TDLAS) tomography has been widely used for imaging of two-dimensional temperature distributions in reactive flows.
no code implementations • SEMEVAL 2020 • Chang Liu, Dong Yu
We demonstrate the effectiveness of our approaches, which achieves 0. 95 of subtask 1 in F1 while using only a subset of giving training set to fine-tune the BERT model, and our official submission achieves F1 0. 802, which ranks us 16th in the competition.
no code implementations • 18 Jun 2021 • Chang Liu, Xiaolin Wu
Nighttime photographers are often troubled by light pollution of unwanted artificial lights.
no code implementations • 24 Jun 2021 • Xianlong Zeng, Simon Lin, Chang Liu
In addition, our framework showed a great generalizability potential to transfer learned knowledge from one institution to another, paving the way for future healthcare model pre-training across institutions.
no code implementations • 23 Jun 2021 • Xianlong Zeng, Simon Lin, Chang Liu
The claims data, containing medical codes, services information, and incurred expenditure, can be a good resource for estimating an individual's health condition and medical risk level.
no code implementations • 25 Jun 2021 • Tianle Yue, Hang Yang, Zongliang Du, Chang Liu, Khalil I. Elkhodary, Shan Tang, Xu Guo
During offline training, a mapping function is built between high and low resolution representations of a given design domain.
no code implementations • 16 Jul 2021 • Haodi Jiang, Ju Jing, Jiasheng Wang, Chang Liu, Qin Li, Yan Xu, Jason T. L. Wang, Haimin Wang
Our method consists of a data pre-processing component that prepares training data from a threshold-based tool, a deep learning model implemented as a Bayesian convolutional neural network for probabilistic image segmentation with uncertainty quantification to predict fibrils, and a post-processing component containing a fibril-fitting algorithm to determine fibril orientations.
no code implementations • 3 Aug 2021 • Chang Liu, Han Yu, Boyang Li, Zhiqi Shen, Zhanning Gao, Peiran Ren, Xuansong Xie, Lizhen Cui, Chunyan Miao
Noisy labels are commonly found in real-world data, which cause performance degradation of deep neural networks.
no code implementations • ICML Workshop AML 2021 • Wenzhao Xiang, Chang Liu, Shibao Zheng
Traditional adversarial examples are typically generated by adding perturbation noise to the input image within a small matrix norm.
no code implementations • 26 Aug 2021 • Chang Liu, Weijie Yuan, Shuangyang Li, Xuemeng Liu, Husheng Li, Derrick Wing Kwan Ng, Yonghui Li
Specifically, the convolution and LSTM modules are successively adopted in the proposed HCL-Net to exploit the spatial and temporal dependencies of communication channels to further improve the learning performance.
no code implementations • 13 Sep 2021 • Wenzhao Xiang, Hang Su, Chang Liu, Yandong Guo, Shibao Zheng
As designers of artificial intelligence try to outwit hackers, both sides continue to hone in on AI's inherent vulnerabilities.
no code implementations • 30 Sep 2021 • Zijian Zhu, Hang Su, Chang Liu, Wenzhao Xiang, Shibao Zheng
Fortunately, most existing adversarial patches can be outwitted, disabled and rejected by a simple classification network called an adversarial patch detector, which distinguishes adversarial patches from original images.
no code implementations • 1 Oct 2021 • Chongyang Tao, Jiazhan Feng, Chang Liu, Juntao Li, Xiubo Geng, Daxin Jiang
For this task, the adoption of pre-trained language models (such as BERT) has led to remarkable progress in a number of benchmarks.
no code implementations • CVPR 2022 • Chang Liu, Xiang Yu, Yi-Hsuan Tsai, Ramin Moslemi, Masoud Faraki, Manmohan Chandraker, Yun Fu
Convolutional Neural Networks have achieved remarkable success in face recognition, in part due to the abundant availability of data.
no code implementations • 29 Sep 2021 • Qiwei Ye, Yuxuan Song, Chang Liu, Fangyun Wei, Tao Qin, Tie-Yan Liu
Stochastic polic have been widely applied for their good property in exploration and uncertainty quantification.
Ranked #1 on MuJoCo Games on Ant-v3
no code implementations • 11 Oct 2021 • Chang Liu
The principal knows the reward of the task and provides information to the agent over time in order to motivate effort.
no code implementations • 14 Oct 2021 • Chang Liu, Hairong Tang, Luyan Ji, Yongchao Zhao
Based on the mapping results, we analyzed the changes of Miyun Reservoir from 1984 to 2020 and the driving factors of them.
no code implementations • 1 Nov 2021 • Chang Liu, Chen Gao, Depeng Jin, Yong Li
We first conduct information propagation on two sub-graphs to learn the representations of POIs and users.
no code implementations • 1 Nov 2021 • Haoji Liu, Weichao Zhuang, Guodong Yin, Rongcan Li, Chang Liu, Shanxing Zhou
We first formulate the optimal merging control problem, which includes the constraints of safety and vehicle dynamics, with the objectives of minimizing travel time and energy consumption.
no code implementations • CCL 2020 • Chang Liu, Shengxiang Gao, Zhengtao Yu, Yuxin Huang, Congcong You
汉越平行句对抽取是缓解汉越平行语料库数据稀缺的重要方法。平行句对抽取可转换为同一语义空间下的句子相似性分类任务, 其核心在于双语语义空间对齐。传统语义空间对齐方法依赖于大规模的双语平行语料, 越南语作为低资源语言获取大规模平行语料相对困难。针对这个问题本文提出一种利用种子词典进行跨语言双语预训练及Bi-LSTM(Bi-directional Long Short-Term Memory)的汉-越平行句对抽取方法。预训练中仅需要大量的汉越单语和一个汉越种子词典, 通过利用汉越种子词典将汉越双语映射到公共语义空间进行词对齐。再利用Bi-LSTM和CNN(Convolutional Neural Networks)分别提取句子的全局特征和局部特征从而最大化表示汉-越句对之间的语义相关性。实验结果表明, 本文模型在F1得分上提升7. 1%, 优于基线模型。
no code implementations • CCL 2020 • Hongrui Wang, Chang Liu, Dong Yu
道德词典资源的建设是人工智能伦理计算的一个研究重点。由于道德行为复杂多样, 现有的英文道德词典分类体系并不完善, 而中文方面目前尚未有相关的词典资源, 理论体系和构建方法仍待探究。针对以上问题, 该文提出了面向人工智能伦理计算的中文道德词典构建任务, 设计了四类标签和四种类型, 得到包含25, 012个词的中文道德词典资源。实验结果表明, 该词典资源不仅能够使机器学会道德知识, 判断词的道德标签和类型, 而且能够为句子级别的道德文本分析提供数据支持。
no code implementations • 20 Nov 2021 • Long Gao, Chang Liu, Dooman Arefan, Ashok Panigrahy, Margarita L. Zuley, Shandong Wu
To address this challenge, we propose a medical-knowledge-guided one-class classification approach that leverages domain-specific knowledge of classification tasks to boost the model's performance.
no code implementations • 20 Nov 2021 • Long Gao, Chang Liu, Dooman Arefan, Ashok Panigrahy, Shandong Wu
These methods mainly focus on capturing either compact or descriptive features, where the information of the samples of a given one class is not sufficiently utilized.
no code implementations • 19 Jan 2022 • Zhongyuan Guo, Hong Zheng, Changhui You, Tianyu Wang, Chang Liu
We first analyze the production principle of anti-counterfeiting QR code, and convert the identification of copy forgery to device category forensics, and then a Dual-Branch Multi-Scale Feature Fusion network is proposed.
no code implementations • 9 Feb 2022 • Jie Chen, Chang Liu, Jiawu Xie, Jie An, Nan Huang
In particular, this method breaks through the limitations of the existing methods, not only achieves good results in multivariate separation, but also effectively separates signals when mixed with 40dB Gaussian noise signals.
no code implementations • 26 Feb 2022 • Vikram Shree, Carlos Diaz-Ruiz, Chang Liu, Bharath Hariharan, Mark Campbell
This paper focuses on the problem of decentralized pedestrian tracking using a sensor network.
no code implementations • 6 Mar 2022 • Jiayi Zhang, Chang Liu, Junchi Yan, Xijun Li, Hui-Ling Zhen, Mingxuan Yuan
This paper surveys the trend of leveraging machine learning to solve mixed integer programming (MIP) problems.
no code implementations • 10 Mar 2022 • Chang Liu, Chun Yang, Hai-Bo Qin, Xiaobin Zhu, Cheng-Lin Liu, Xu-Cheng Yin
Scene text recognition is a popular topic and extensively used in the industry.
no code implementations • 28 Feb 2022 • Yu Shi, Shuxin Zheng, Guolin Ke, Yifei Shen, Jiacheng You, Jiyan He, Shengjie Luo, Chang Liu, Di He, Tie-Yan Liu
This technical note describes the recent updates of Graphormer, including architecture design modifications, and the adaption to 3D molecular dynamics simulation.
no code implementations • 29 Mar 2022 • Chang Liu, Xiaoyan Qian, Xiaojuan Qi, Edmund Y. Lam, Siew-Chong Tan, Ngai Wong
While a few previous studies tried to automatically generate 3D bounding boxes from weak labels such as 2D boxes, the quality is sub-optimal compared to human annotators.
no code implementations • 1 Mar 2022 • Qi Zhang, Chang Liu, Stephen Wu, Ryo Yoshida
The design variables consist of a set of reactants in a reaction network and its network topology.
no code implementations • 6 Apr 2022 • Wenhan Cao, Jingliang Duan, Shengbo Eben Li, Chen Chen, Chang Liu, Yu Wang
Both the primal and dual estimators are learned from data using supervised learning techniques, and the explicit sample size is provided, which enables us to guarantee the quality of each learned estimator in terms of feasibility and optimality.
no code implementations • 26 Apr 2022 • Chang Liu, Xudong Jiang, Henghui Ding
In this work, we propose a novel framework that simultaneously detects the target-of-interest via feature propagation and generates a fine-grained segmentation mask.
no code implementations • 28 Apr 2022 • Sijia Li, Gaopeng Gou, Chang Liu, Chengshang Hou, Zhenzhen Li, Gang Xiong
In this paper, we propose a Temporal Transaction Aggregation Graph Network (TTAGN) to enhance phishing scams detection performance on Ethereum.
no code implementations • ACL 2022 • Chang Liu, Xu Tan, Chongyang Tao, Zhenxin Fu, Dongyan Zhao, Tie-Yan Liu, Rui Yan
To enable the chatbot to foresee the dialogue future, we design a beam-search-like roll-out strategy for dialogue future simulation using a typical dialogue generation model and a dialogue selector.
no code implementations • 25 May 2022 • Eduardo Pérez-Pellitero, Sibi Catley-Chandar, Richard Shaw, Aleš Leonardis, Radu Timofte, Zexin Zhang, Cen Liu, Yunbo Peng, Yue Lin, Gaocheng Yu, Jin Zhang, Zhe Ma, Hongbin Wang, Xiangyu Chen, Xintao Wang, Haiwei Wu, Lin Liu, Chao Dong, Jiantao Zhou, Qingsen Yan, Song Zhang, Weiye Chen, Yuhang Liu, Zhen Zhang, Yanning Zhang, Javen Qinfeng Shi, Dong Gong, Dan Zhu, Mengdi Sun, Guannan Chen, Yang Hu, Haowei Li, Baozhu Zou, Zhen Liu, Wenjie Lin, Ting Jiang, Chengzhi Jiang, Xinpeng Li, Mingyan Han, Haoqiang Fan, Jian Sun, Shuaicheng Liu, Juan Marín-Vega, Michael Sloth, Peter Schneider-Kamp, Richard Röttger, Chunyang Li, Long Bao, Gang He, Ziyao Xu, Li Xu, Gen Zhan, Ming Sun, Xing Wen, Junlin Li, Shuang Feng, Fei Lei, Rui Liu, Junxiang Ruan, Tianhong Dai, Wei Li, Zhan Lu, Hengyan Liu, Peian Huang, Guangyu Ren, Yonglin Luo, Chang Liu, Qiang Tu, Fangya Li, Ruipeng Gang, Chenghua Li, Jinjing Li, Sai Ma, Chenming Liu, Yizhen Cao, Steven Tel, Barthelemy Heyrman, Dominique Ginhac, Chul Lee, Gahyeon Kim, Seonghyun Park, An Gia Vien, Truong Thanh Nhat Mai, Howoon Yoon, Tu Vo, Alexander Holston, Sheir Zaheer, Chan Y. Park
The challenge is composed of two tracks with an emphasis on fidelity and complexity constraints: In Track 1, participants are asked to optimize objective fidelity scores while imposing a low-complexity constraint (i. e. solutions can not exceed a given number of operations).
no code implementations • 2 Jun 2022 • Chang Liu, Zhen-Hua Ling, Ling-Hui Chen
This paper proposes a multilingual speech synthesis method which combines unsupervised phonetic representations (UPR) and supervised phonetic representations (SPR) to avoid reliance on the pronunciation dictionaries of target languages.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 25 Jun 2022 • HongBing Zhang, Xinyi Liu, Chang Liu, HongTao Fan, YaJing Li, Xinyun Zhu
The proposed function is generalized to tensor cases, yielding tensor MLCP and weighted tensor $L\gamma$-norm.
no code implementations • 4 Jul 2022 • Chang Liu, Yugong Luo, Pengfei Li, Chunhui Xing, Weiwei Kong
To deal with this problem, this paper introduces a two-dimensional maneuver management framework with a fault-tolerant mechanism on the basis of the proposed hierarchical architecture for the platoon control system.
no code implementations • 6 Jul 2022 • Zhennan Wang, Kehan Li, Runyi Yu, Yian Zhao, Pengchong Qiao, Chang Liu, Fan Xu, Xiangyang Ji, Guoli Song, Jie Chen
In this paper, we analyze batch normalization from the perspective of discriminability and find the disadvantages ignored by previous studies: the difference in $l_2$ norms of sample features can hinder batch normalization from obtaining more distinguished inter-class features and more compact intra-class features.
no code implementations • 26 Sep 2022 • Chang Liu, Xuemeng Liu, Shuangyang Li, Weijie Yuan, Derrick Wing Kwan Ng
Predictive beamforming design is an essential task in realizing high-mobility integrated sensing and communication (ISAC), which highly depends on the accuracy of the channel prediction (CP), i. e., predicting the angular parameters of users.
no code implementations • 5 Oct 2022 • Wenhan Cao, Chang Liu, Zhiqian Lan, Shengbo Eben Li, Wei Pan, Angelo Alessandri
The accuracy of moving horizon estimation (MHE) suffers significantly in the presence of measurement outliers.
no code implementations • 7 Oct 2022 • Chang Liu, Terence Jie Chua, Jun Zhao
Therefore, we formulate a joint learning and communication optimization problem to minimize total model parameter communication and computation delay, by optimizing local iteration counts and edge iteration counts.
no code implementations • COLING 2022 • Jiazhan Feng, Chongyang Tao, Zhen Li, Chang Liu, Tao Shen, Dongyan Zhao
In this paper, we propose a reciprocal learning approach to jointly optimize a knowledge retriever and a response ranker for knowledge-grounded response retrieval without ground-truth knowledge labels.
no code implementations • CVPR 2023 • Kehan Li, Zhennan Wang, Zesen Cheng, Runyi Yu, Yian Zhao, Guoli Song, Chang Liu, Li Yuan, Jie Chen
Recently, self-supervised large-scale visual pre-training models have shown great promise in representing pixel-level semantic relationships, significantly promoting the development of unsupervised dense prediction tasks, e. g., unsupervised semantic segmentation (USS).
no code implementations • 13 Oct 2022 • Chang Liu, Yuwen Yang, Yue Ding, Hongtao Lu
The normalizing layer has become one of the basic configurations of deep learning models, but it still suffers from computational inefficiency, interpretability difficulties, and low generality.
no code implementations • CVPR 2023 • Pengchong Qiao, Zhidan Wei, Yu Wang, Zhennan Wang, Guoli Song, Fan Xu, Xiangyang Ji, Chang Liu, Jie Chen
Semi-supervised learning (SSL) essentially pursues class boundary exploration with less dependence on human annotations.
no code implementations • 28 Oct 2022 • Ligen Shi, Chang Liu, Di He, Xing Zhao, Jun Qiu
A major challenge for matching-based depth estimation is to prevent mismatches in occlusion and smooth regions.
no code implementations • 28 Oct 2022 • Chang Liu, Yuwen Yang, Xun Cai, Yue Ding, Hongtao Lu
Federated learning (FL) faces three major difficulties: cross-domain, heterogeneous models, and non-i. i. d.
no code implementations • 2 Nov 2022 • Yifei Zhang, Chang Liu, Yu Zhou, Weiping Wang, Qixiang Ye, Xiangyang Ji
In this paper, we present relation-aware contrastive self-supervised learning (ReCo) to integrate instance relations, i. e., global distribution relation and local interpolation relation, into the CSL framework in a plug-and-play fashion.
no code implementations • 5 Nov 2022 • Yu Yang, Wing Yin Cheung, Chang Liu, Xiangyang Ji
Multiview self-supervised representation learning roots in exploring semantic consistency across data of complex intra-class variation.
no code implementations • 15 Nov 2022 • Kun He, Chang Liu, Stephen Lin, John E. Hopcroft
And further combination with our feature augmentation techniques, termed LOMA_IF&FO, can continue to strengthen the model and outperform advanced intensity transformation methods for data augmentation.
no code implementations • 16 Nov 2022 • Xinyu Zhou, Chang Liu, Jun Zhao
The Metaverse has received much attention recently.
no code implementations • 19 Nov 2022 • Chang Liu, Yuwen Yang, Yue Ding, Hongtao Lu
While most existing message-passing graph neural networks (MPNNs) are permutation-invariant in graph-level representation learning and permutation-equivariant in node- and edge-level representation learning, their expressive power is commonly limited by the 1-Weisfeiler-Lehman (1-WL) graph isomorphism test.
no code implementations • CVPR 2023 • Zesen Cheng, Pengchong Qiao, Kehan Li, Siheng Li, Pengxu Wei, Xiangyang Ji, Li Yuan, Chang Liu, Jie Chen
Weakly supervised semantic segmentation is typically inspired by class activation maps, which serve as pseudo masks with class-discriminative regions highlighted.
Optical Character Recognition (OCR) Weakly supervised Semantic Segmentation +1
no code implementations • 23 Nov 2022 • Chang Liu, Xuemeng Liu, Zhiqiang Wei, Derrick Wing Kwan Ng, Robert Schober
With the proposed predictive approach, we can avoid full-scale CSI estimation and facilitate low-dimensional CE for transmit beamforming design such that the signaling overhead is reduced by a scale of $\frac{1}{N}$, where $N$ is the number of IRS elements.
no code implementations • 30 Nov 2022 • Chang Liu
We study a two-period moral hazard problem; there are two agents, with action sets that are unknown to the principal.
no code implementations • 16 Dec 2022 • Wei Sun, Chengao Liu, Linyan Zhang, Yu Li, Pengxu Wei, Chang Liu, Jialing Zou, Jianbin Jiao, Qixiang Ye
Optimizing a convolutional neural network (CNN) for camouflaged object detection (COD) tends to activate local discriminative regions while ignoring complete object extent, causing the partial activation issue which inevitably leads to missing or redundant regions of objects.
no code implementations • 20 Dec 2022 • Chang Liu, Chongyang Tao, Xiubo Geng, Tao Shen, Dongyan Zhao, Can Xu, Binxing Jiao, Daxin Jiang
Different from previous works that only rely on one positive and hard negatives as candidate passages, we create dark examples that all have moderate relevance to the query through mixing-up and masking in discrete space.
no code implementations • 14 Dec 2022 • Kun Tang, Xu Cao, Zhipeng Cao, Tong Zhou, Erlong Li, Ao Liu, Shengtao Zou, Chang Liu, Shuqi Mei, Elena Sizikova, Chao Zheng
THMA has been deployed by the Tencent Map team to provide services to downstream companies and users, serving over 1, 000 labeling workers and producing more than 30, 000 kilometers of HD map data per day at most.
no code implementations • 28 Dec 2022 • Chang Liu, Shuangyang Li, Weijie Yuan, Xuemeng Liu, Derrick Wing Kwan Ng
This paper investigates the orthogonal time frequency space (OTFS) transmission for enabling ultra-reliable low-latency communications (URLLC).
no code implementations • 28 Feb 2023 • Chang Liu, Rui Zhang, Xishan Zhang, Yifan Hao, Zidong Du, Xing Hu, Ling Li, Qi Guo
The energy-efficient works try to decrease the precision of multiplication or replace the multiplication with energy-efficient operations such as addition or bitwise shift, to reduce the energy consumption of FP32 multiplications.
no code implementations • 28 Feb 2023 • Chang Liu, Wenzhao Xiang, Yuan He, Hui Xue, Shibao Zheng, Hang Su
To address this issue, we proposed a novel method of Augmenting data with Adversarial examples via a Wavelet module (AdvWavAug), an on-manifold adversarial data augmentation technique that is simple to implement.
no code implementations • 28 Feb 2023 • Chang Liu, Yinpeng Dong, Wenzhao Xiang, Xiao Yang, Hang Su, Jun Zhu, Yuefeng Chen, Yuan He, Hui Xue, Shibao Zheng
In our benchmark, we evaluate the robustness of 55 typical deep learning models on ImageNet with diverse architectures (e. g., CNNs, Transformers) and learning algorithms (e. g., normal supervised training, pre-training, adversarial training) under numerous adversarial attacks and out-of-distribution (OOD) datasets.
no code implementations • 28 Feb 2023 • Yifan Yang, Chang Liu, Zheng Zhang
Online optimization has gained increasing interest due to its capability of tracking real-world streaming data.
no code implementations • 13 Mar 2023 • Zesen Cheng, Kehan Li, Peng Jin, Xiangyang Ji, Li Yuan, Chang Liu, Jie Chen
An intuitive materialization of our paradigm is Parallel Vertex Diffusion (PVD) to directly set vertex coordinates as the generation target and use a diffusion model to train and infer.
no code implementations • 7 Mar 2023 • Chang Liu, Sandra Paterlini
Stock price prediction is a crucial element in financial trading as it allows traders to make informed decisions about buying, selling, and holding stocks.
no code implementations • ICCV 2023 • Kehan Li, Yian Zhao, Zhennan Wang, Zesen Cheng, Peng Jin, Xiangyang Ji, Li Yuan, Chang Liu, Jie Chen
Interactive segmentation enables users to segment as needed by providing cues of objects, which introduces human-computer interaction for many fields, such as image editing and medical image analysis.
no code implementations • 27 Mar 2023 • Xiaoyan Qian, Chang Liu, Xiaojuan Qi, Siew-Chong Tan, Edmund Lam, Ngai Wong
3D automatic annotation has received increased attention since manually annotating 3D point clouds is laborious.
no code implementations • 12 May 2023 • Chang Liu
This paper studies the equilibrium behavior in contests with stochastic progress.
no code implementations • 17 May 2023 • Hao Li, Peng Jin, Zesen Cheng, Songyang Zhang, Kai Chen, Zhennan Wang, Chang Liu, Jie Chen
Video question answering aims at answering a question about the video content by reasoning the alignment semantics within them.
no code implementations • 22 May 2023 • Jinsong Chen, Chang Liu, Kaiyuan Gao, Gaichao Li, Kun He
Graph Transformers, emerging as a new architecture for graph representation learning, suffer from the quadratic complexity on the number of nodes when handling large graphs.
no code implementations • 24 May 2023 • Chang Liu, Henghui Ding, Yulun Zhang, Xudong Jiang
However, the generic attention mechanism in Transformer only uses the language input for attention weight calculation, which does not explicitly fuse language features in its output.
no code implementations • 28 May 2023 • Quzhe Huang, Yutong Hu, Shengqi Zhu, Yansong Feng, Chang Liu, Dongyan Zhao
After examining the relation definitions in various ETRE tasks, we observe that all relations can be interpreted using the start and end time points of events.
no code implementations • 8 Jun 2023 • Shuxin Zheng, Jiyan He, Chang Liu, Yu Shi, Ziheng Lu, Weitao Feng, Fusong Ju, Jiaxi Wang, Jianwei Zhu, Yaosen Min, He Zhang, Shidi Tang, Hongxia Hao, Peiran Jin, Chi Chen, Frank Noé, Haiguang Liu, Tie-Yan Liu
In this paper, we introduce a novel deep learning framework, called Distributional Graphormer (DiG), in an attempt to predict the equilibrium distribution of molecular systems.
no code implementations • 19 Jun 2023 • Zesen Cheng, Peng Jin, Hao Li, Kehan Li, Siheng Li, Xiangyang Ji, Chang Liu, Jie Chen
Bottom-up methods are mainly perturbed by Inferior Positive (IP) errors due to the lack of prior object information.
no code implementations • 11 Jul 2023 • Dongbo Wang, Chang Liu, Zhixiao Zhao, Si Shen, Liu Liu, Bin Li, Haotian Hu, Mengcheng Wu, Litao Lin, Xue Zhao, Xiyu Wang
In the context of the rapid development of large language models, we have meticulously trained and introduced the GujiBERT and GujiGPT language models, which are foundational models specifically designed for intelligent information processing of ancient texts.
no code implementations • 16 Sep 2023 • Yuwen Yang, Chang Liu, Xun Cai, Suizhi Huang, Hongtao Lu, Yue Ding
Federated Learning (FL) has emerged as a promising approach to enable collaborative learning among multiple clients while preserving data privacy.
no code implementations • 19 Sep 2023 • Chang Liu, Yi Niu, Mingming Ma, Fu Li, Guangming Shi
The basic principle of the patch-matching based style transfer is to substitute the patches of the content image feature maps by the closest patches from the style image feature maps.
no code implementations • 19 Sep 2023 • Chang Liu, Giulia Rizzoli, Francesco Barbato, Andrea Maracani, Marco Toldo, Umberto Michieli, Yi Niu, Pietro Zanuttigh
Catastrophic forgetting of previous knowledge is a critical issue in continual learning typically handled through various regularization strategies.
no code implementations • ICCV 2023 • Hongliang He, Jun Wang, Pengxu Wei, Fan Xu, Xiangyang Ji, Chang Liu, Jie Chen
Experiments on three nuclear instance segmentation datasets justify the superiority of TopoSeg, which achieves state-of-the-art performance.
no code implementations • 28 Sep 2023 • He Zhang, Siyuan Liu, Jiacheng You, Chang Liu, Shuxin Zheng, Ziheng Lu, Tong Wang, Nanning Zheng, Bin Shao
Orbital-free density functional theory (OFDFT) is a quantum chemistry formulation that has a lower cost scaling than the prevailing Kohn-Sham DFT, which is increasingly desired for contemporary molecular research.
no code implementations • 2 Oct 2023 • Shenzhi Wang, Chang Liu, Zilong Zheng, Siyuan Qi, Shuo Chen, Qisen Yang, Andrew Zhao, Chaofei Wang, Shiji Song, Gao Huang
This study utilizes the intricate Avalon game as a testbed to explore LLMs' potential in deceptive environments.
no code implementations • 26 Oct 2023 • Chang Liu, Liguo Zhou, Yanliang Huang, Alois Knoll
Vehicle perception systems strive to achieve comprehensive and rapid visual interpretation of their surroundings for improved safety and navigation.
no code implementations • 29 Nov 2023 • Chang Liu, Peng Hou, AnXiang Zeng, Han Yu
Since its deployment in March 2023, it has achieved a remarkable 9. 90% increase in terms of clicks per user and a 4. 23% boost in terms of orders per user for the image search feature on the Shopee e-commerce platform.
no code implementations • 4 Dec 2023 • Akshay Goel, Almog Gueta, Omry Gilon, Chang Liu, Sofia Erell, Lan Huong Nguyen, Xiaohong Hao, Bolous Jaber, Shashir Reddy, Rupesh Kartha, Jean Steiner, Itay Laish, Amir Feder
The results highlight the potential of using LLMs to improve the utilization of unstructured clinical data, allowing for the swift deployment of tailored NLP solutions in healthcare.
no code implementations • 6 Dec 2023 • Chang Liu, Tamas Sziranyi
This work focuses on the development of an optimal wildfire escape route planning system specifically designed for drones, considering dynamic fire and smoke models.
no code implementations • 6 Dec 2023 • Chang Liu, Tamas Sziranyi
Taking the Chongqing wildfire on August 24, 2022, as a case study, the results demonstrate that the dynamic escape route planning algorithm can provide an optimal real-time navigation path for humans in the presence of fire through the information fusion of UAVs and satellites.
no code implementations • 7 Dec 2023 • Ruyi Gan, XiaoJun Wu, Junyu Lu, Yuanhe Tian, Dixiang Zhang, Ziwei Wu, Renliang Sun, Chang Liu, Jiaxing Zhang, Pingjian Zhang, Yan Song
However, there are few specialized models in certain domains, such as interior design, which is attributed to the complex textual descriptions and detailed visual elements inherent in design, alongside the necessity for adaptable resolution.
no code implementations • 11 Dec 2023 • Yitong Wang, Chang Liu, Jun Zhao
In pursuit of enhancing the accessibility of AIGC services, the deployment of AIGC models (e. g., diffusion models) to edge servers and local devices has become a prevailing trend.
no code implementations • 14 Dec 2023 • Stephen Wu, Yu Otake, Daijiro Mizutani, Chang Liu, Kotaro Asano, Nana Sato, Hidetoshi Baba, Yusuke Fukunaga, Yosuke Higo, Akiyoshi Kamura, Shinnosuke Kodama, Masataka Metoki, Tomoka Nakamura, Yuto Nakazato, Taiga Saito, Akihiro Shioi, Masahiro Takenobu, Keigo Tsukioka, Ryo Yoshikawa
The integration of Large Language Models (LLMs) like ChatGPT into the workflows of geotechnical engineering has a high potential to transform how the discipline approaches problem-solving and decision-making.
no code implementations • 28 Dec 2023 • Jason Chun Lok Li, Chang Liu, Binxiao Huang, Ngai Wong
Existing approaches to Implicit Neural Representation (INR) can be interpreted as a global scene representation via a linear combination of Fourier bases of different frequencies.
no code implementations • 5 Jan 2024 • Pengying Wu, Yao Mu, Bingxian Wu, Yi Hou, Ji Ma, Shanghang Zhang, Chang Liu
In the realm of household robotics, the Zero-Shot Object Navigation (ZSON) task empowers agents to adeptly traverse unfamiliar environments and locate objects from novel categories without prior explicit training.
no code implementations • 21 Jan 2024 • Ruochi Zhang, Haoran Wu, Chang Liu, Huaping Li, Yuqian Wu, Kewei Li, Yifan Wang, Yifan Deng, Jiahui Chen, Fengfeng Zhou, Xin Gao
This study introduces a novel multi-view contrastive learning framework PepHarmony for the sequence-based peptide encoding task.
no code implementations • 20 Jan 2024 • Chang Liu, Antwan D. Clark
Constant gain least-mean-squares (LMS) algorithms have a wide range of applications in trajectory tracking problems, but the formal convergence of LMS in mean square is not yet fully established.
no code implementations • 23 Jan 2024 • Chang Liu, Laura Klein, Yixing Huang, Edith Baader, Michael Lell, Marc Kachelrieß, Andreas Maier
The average organ dice of the proposed method is 0. 71 compared with 0. 63 in baseline model, indicating the enhancement of anatomical structures.
no code implementations • 22 Feb 2024 • Yanliang Huang, Liguo Zhou, Chang Liu, Alois Knoll
The implementation of Autonomous Driving (AD) technologies within urban environments presents significant challenges.
no code implementations • 23 Feb 2024 • Jingtao Ding, Chang Liu, Yu Zheng, Yunke Zhang, Zihan Yu, Ruikun Li, Hongyi Chen, Jinghua Piao, Huandong Wang, Jiazhen Liu, Yong Li
Complex networks pervade various real-world systems, from the natural environment to human societies.
no code implementations • 29 Feb 2024 • Ji Ma, Hongming Dai, Yao Mu, Pengying Wu, Hao Wang, Xiaowei Chi, Yang Fei, Shanghang Zhang, Chang Liu
Zero-Shot Object Navigation (ZSON) requires agents to autonomously locate and approach unseen objects in unfamiliar environments and has emerged as a particularly challenging task within the domain of Embodied AI.
no code implementations • 5 Mar 2024 • Chang Liu, Fuxin Fan, Annette Schwarz, Andreas Maier
Multi-organ segmentation in medical images is a widely researched task and can save much manual efforts of clinicians in daily routines.
no code implementations • 14 Mar 2024 • He Zhang, Chang Liu, Zun Wang, Xinran Wei, Siyuan Liu, Nanning Zheng, Bin Shao, Tie-Yan Liu
This merit addresses the data scarcity difficulty, and distinguishes the task from other property prediction formulations with unique benefits: (1) self-consistency training enables the model to be trained on a large amount of unlabeled data, hence substantially enhances generalization; (2) self-consistency training is more efficient than labeling data with DFT for supervised training, since it is an amortization of DFT calculation over a set of molecular structures.
no code implementations • 26 Mar 2024 • Chang Liu, Wei Han, Guangyu Yan, Bowang Zhang, Chunlin Li
The swift response of SCC and two-step perturb-and-observe algorithm mitigate output disturbances, thereby expediting the frequency tuning process.
no code implementations • 30 Mar 2024 • Wenhan Cao, Shiqi Liu, Chang Liu, Zeyu He, Stephen S. -T. Yau, Shengbo Eben Li
In this paper, we find that by adding an additional event that stipulates an inequality condition, we can transform the conditional probability into a special integration that is analogous to convolution.
no code implementations • 10 Apr 2024 • Ligen Shi, Chang Liu, Ping Yang, Jun Qiu, Xing Zhao
In spectral CT reconstruction, the basis materials decomposition involves solving a large-scale nonlinear system of integral equations, which is highly ill-posed mathematically.
no code implementations • 19 Apr 2024 • Chang Liu, Xiaohui Xie, Xinggong Zhang, Yong Cui
The networking field is characterized by its high complexity and rapid iteration, requiring extensive expertise to accomplish network tasks, ranging from network design, configuration, diagnosis and security.
no code implementations • 1 May 2024 • Yian Zhao, Kehan Li, Zesen Cheng, Pengchong Qiao, Xiawu Zheng, Rongrong Ji, Chang Liu, Li Yuan, Jie Chen
In this work, we introduce Granularity-Controllable Interactive Segmentation (GraCo), a novel approach that allows precise control of prediction granularity by introducing additional parameters to input.
1 code implementation • 17 May 2020 • Juntao Li, Chang Liu, Jian Wang, Lidong Bing, Hongsong Li, Xiaozhong Liu, Dongyan Zhao, Rui Yan
We manually collect a new and high-quality paired dataset, where each pair contains an unordered product attribute set in the source language and an informative product description in the target language.
1 code implementation • 6 Jul 2020 • Yifei Zhang, Chang Liu, Yu Zhou, Wei Wang, Weiping Wang, Qixiang Ye
In this work, we propose a novel clustering based method, which, by iteratively excluding class inconsistent samples during progressive cluster formation, alleviates the impact of noise samples in a simple-yet-effective manner.
1 code implementation • 4 Sep 2020 • Yasser Abduallah, Jason T. L. Wang, Yang Nie, Chang Liu, Haimin Wang
Solar flare prediction plays an important role in understanding and forecasting space weather.
1 code implementation • 24 Jun 2021 • Xianlong Zeng, Fanghao Song, Zhongen Li, Krerkkiat Chusap, Chang Liu
Our method can be divided into three stages: 1) a neighborhood generation stage, which generates instances based on the given sample; 2) a classification stage, which yields classifications on the generated instances to carve out the local decision boundary and delineate the model behavior; and 3) a human-in-the-loop stage, which involves human to refine and explore the neighborhood of interest.
BIG-bench Machine Learning Explainable artificial intelligence +1
1 code implementation • 18 May 2023 • Yitong Li, Chang Liu, Jie Ma
Weakly supervised learning based on scribble annotations in target extraction of remote sensing images has drawn much interest due to scribbles' flexibility in denoting winding objects and low cost of manually labeling.
1 code implementation • 17 Apr 2024 • Ziyu Zhou, Wenyuan Shen, Chang Liu
Colorectal cancer (CRC), which frequently originates from initially benign polyps, remains a significant contributor to global cancer-related mortality.
3 code implementations • 22 Feb 2020 • Hao Liu, Chang Liu, Jason T. L. Wang, Haimin Wang
We present two recurrent neural networks (RNNs), one based on gated recurrent units and the other based on long short-term memory, for predicting whether an active region (AR) that produces an M- or X-class flare will also produce a coronal mass ejection (CME).
1 code implementation • 29 Nov 2021 • Mengnan Shi, Chang Liu, Qixiang Ye, Jianbin Jiao
Gating modules have been widely explored in dynamic network pruning to reduce the run-time computational cost of deep neural networks while preserving the representation of features.