Search Results for author: Zheng Xu

Found 53 papers, 17 papers with code

Exploiting Lists of Names for Named Entity Identification of Financial Institutions from Unstructured Documents

no code implementations14 Feb 2016 Zheng Xu, Douglas Burdick, Louiqa Raschid

To our knowledge, our proposed solutions, Dict-based NER and Rank-based ER, and the root and suffix dictionaries, are the first attempt to exploit specialized knowledge, i. e., lists of FI names, for rule-based NER and

Entity Resolution named-entity-recognition +2

Training Neural Networks Without Gradients: A Scalable ADMM Approach

2 code implementations6 May 2016 Gavin Taylor, Ryan Burmeister, Zheng Xu, Bharat Singh, Ankit Patel, Tom Goldstein

With the growing importance of large network models and enormous training datasets, GPUs have become increasingly necessary to train neural networks.

Adaptive ADMM with Spectral Penalty Parameter Selection

no code implementations24 May 2016 Zheng Xu, Mario A. T. Figueiredo, Tom Goldstein

The alternating direction method of multipliers (ADMM) is a versatile tool for solving a wide range of constrained optimization problems, with differentiable or non-differentiable objective functions.

An Empirical Study of ADMM for Nonconvex Problems

no code implementations10 Dec 2016 Zheng Xu, Soham De, Mario Figueiredo, Christoph Studer, Tom Goldstein

The alternating direction method of multipliers (ADMM) is a common optimization tool for solving constrained and non-differentiable problems.

Image Denoising regression +1

Non-negative Factorization of the Occurrence Tensor from Financial Contracts

1 code implementation10 Dec 2016 Zheng Xu, Furong Huang, Louiqa Raschid, Tom Goldstein

We propose an algorithm for the non-negative factorization of an occurrence tensor built from heterogeneous networks.

Adaptive Relaxed ADMM: Convergence Theory and Practical Implementation

no code implementations CVPR 2017 Zheng Xu, Mario A. T. Figueiredo, Xiaoming Yuan, Christoph Studer, Tom Goldstein

Relaxed ADMM is a generalization of ADMM that often achieves better performance, but its efficiency depends strongly on algorithm parameters that must be chosen by an expert user.

Stabilizing Adversarial Nets With Prediction Methods

1 code implementation ICLR 2018 Abhay Yadav, Sohil Shah, Zheng Xu, David Jacobs, Tom Goldstein

Adversarial neural networks solve many important problems in data science, but are notoriously difficult to train.

Training Quantized Nets: A Deeper Understanding

no code implementations NeurIPS 2017 Hao Li, Soham De, Zheng Xu, Christoph Studer, Hanan Samet, Tom Goldstein

Currently, deep neural networks are deployed on low-power portable devices by first training a full-precision model using powerful hardware, and then deriving a corresponding low-precision model for efficient inference on such systems.

Adaptive Consensus ADMM for Distributed Optimization

no code implementations ICML 2017 Zheng Xu, Gavin Taylor, Hao Li, Mario Figueiredo, Xiaoming Yuan, Tom Goldstein

The alternating direction method of multipliers (ADMM) is commonly used for distributed model fitting problems, but its performance and reliability depend strongly on user-defined penalty parameters.

Distributed Optimization

Group-driven Reinforcement Learning for Personalized mHealth Intervention

1 code implementation14 Aug 2017 Feiyun Zhu, Jun Guo, Zheng Xu, Peng Liao, Junzhou Huang

Due to the popularity of smartphones and wearable devices nowadays, mobile health (mHealth) technologies are promising to bring positive and wide impacts on people's health.

Clustering Decision Making +2

Visualizing the Loss Landscape of Neural Nets

11 code implementations ICLR 2018 Hao Li, Zheng Xu, Gavin Taylor, Christoph Studer, Tom Goldstein

Neural network training relies on our ability to find "good" minimizers of highly non-convex loss functions.

Learning to Cluster for Proposal-Free Instance Segmentation

1 code implementation17 Mar 2018 Yen-Chang Hsu, Zheng Xu, Zsolt Kira, Jiawei Huang

We utilize the most fundamental property of instance labeling -- the pairwise relationship between pixels -- as the supervision to formulate the learning objective, then apply it to train a fully convolutional network (FCN) for learning to perform pixel-wise clustering.

Autonomous Driving Clustering +6

The Effectiveness of Instance Normalization: a Strong Baseline for Single Image Dehazing

no code implementations8 May 2018 Zheng Xu, Xitong Yang, Xue Li, Xiaoshuai Sun

We propose a novel deep neural network architecture for the challenging problem of single image dehazing, which aims to recover the clear image from a degraded hazy image.

Image Dehazing Single Image Dehazing

Learning from Multi-domain Artistic Images for Arbitrary Style Transfer

1 code implementation25 May 2018 Zheng Xu, Michael Wilber, Chen Fang, Aaron Hertzmann, Hailin Jin

We propose a fast feed-forward network for arbitrary style transfer, which can generate stylized image for previously unseen content and style image pairs.

Style Transfer

Accurate and efficient video de-fencing using convolutional neural networks and temporal information

1 code implementation28 Jun 2018 Chen Du, Byeongkeun Kang, Zheng Xu, Ji Dai, Truong Nguyen

To overcome these problems, we propose a novel method consisting of segmentation using convolutional neural networks and a fast/robust recovery algorithm.

object-detection Object Detection +2

Universal Adversarial Training

no code implementations27 Nov 2018 Ali Shafahi, Mahyar Najibi, Zheng Xu, John Dickerson, Larry S. Davis, Tom Goldstein

Standard adversarial attacks change the predicted class label of a selected image by adding specially tailored small perturbations to its pixels.

The Impact of Neural Network Overparameterization on Gradient Confusion and Stochastic Gradient Descent

no code implementations15 Apr 2019 Karthik A. Sankararaman, Soham De, Zheng Xu, W. Ronny Huang, Tom Goldstein

Our results show that, for popular initialization techniques, increasing the width of neural networks leads to lower gradient confusion, and thus faster model training.

The Effect of Neural Net Architecture on Gradient Confusion & Training Performance

no code implementations25 Sep 2019 Karthik A. Sankararaman, Soham De, Zheng Xu, W. Ronny Huang, Tom Goldstein

Through novel theoretical and experimental results, we show how the neural net architecture affects gradient confusion, and thus the efficiency of training.

Exploring Model Robustness with Adaptive Networks and Improved Adversarial Training

no code implementations30 May 2020 Zheng Xu, Ali Shafahi, Tom Goldstein

Our adaptive networks also outperform larger widened non-adaptive architectures that have 1. 5 times more parameters.

Towards Accurate Quantization and Pruning via Data-free Knowledge Transfer

no code implementations14 Oct 2020 Chen Zhu, Zheng Xu, Ali Shafahi, Manli Shu, Amin Ghiasi, Tom Goldstein

Further, we demonstrate that the compact structure and corresponding initialization from the Lottery Ticket Hypothesis can also help in data-free training.

Data Free Quantization Transfer Learning

Local Adaptivity in Federated Learning: Convergence and Consistency

no code implementations4 Jun 2021 Jianyu Wang, Zheng Xu, Zachary Garrett, Zachary Charles, Luyang Liu, Gauri Joshi

Popular optimization algorithms of FL use vanilla (stochastic) gradient descent for both local updates at clients and global updates at the aggregating server.

Federated Learning

Local-Global Knowledge Distillation in Heterogeneous Federated Learning with Non-IID Data

no code implementations30 Jun 2021 Dezhong Yao, Wanning Pan, Yutong Dai, Yao Wan, Xiaofeng Ding, Hai Jin, Zheng Xu, Lichao Sun

Federated learning enables multiple clients to collaboratively learn a global model by periodically aggregating the clients' models without transferring the local data.

Federated Learning Knowledge Distillation

Efficient and Private Federated Learning with Partially Trainable Networks

no code implementations6 Oct 2021 Hakim Sidahmed, Zheng Xu, Ankush Garg, Yuan Cao, Mingqing Chen

Through extensive experiments, we empirically show that Federated learning of Partially Trainable neural networks (FedPT) can result in superior communication-accuracy trade-offs, with up to $46\times$ reduction in communication cost, at a small accuracy cost.

Federated Learning

Digging into Primary Financial Market: Challenges and Opportunities of Adopting Blockchain

no code implementations20 Apr 2022 Ji Liu, Zheng Xu, Yanmei Zhang, Wei Dai, Hao Wu, Shiping Chen

Since the emergence of blockchain technology, its application in the financial market has always been an area of focus and exploration by all parties.

On the Unreasonable Effectiveness of Federated Averaging with Heterogeneous Data

no code implementations9 Jun 2022 Jianyu Wang, Rudrajit Das, Gauri Joshi, Satyen Kale, Zheng Xu, Tong Zhang

Motivated by this observation, we propose a new quantity, average drift at optimum, to measure the effects of data heterogeneity, and explicitly use it to present a new theoretical analysis of FedAvg.

Federated Learning

Beyond Uniform Lipschitz Condition in Differentially Private Optimization

no code implementations21 Jun 2022 Rudrajit Das, Satyen Kale, Zheng Xu, Tong Zhang, Sujay Sanghavi

Most prior results on differentially private stochastic gradient descent (DP-SGD) are derived under the simplistic assumption of uniform Lipschitzness, i. e., the per-sample gradients are uniformly bounded.

Benchmarking regression

Multimodal Open-Vocabulary Video Classification via Pre-Trained Vision and Language Models

no code implementations15 Jul 2022 Rui Qian, Yeqing Li, Zheng Xu, Ming-Hsuan Yang, Serge Belongie, Yin Cui

Utilizing vision and language models (VLMs) pre-trained on large-scale image-text pairs is becoming a promising paradigm for open-vocabulary visual recognition.

Optical Flow Estimation Video Classification +1

Learning to Generate Image Embeddings with User-level Differential Privacy

1 code implementation CVPR 2023 Zheng Xu, Maxwell Collins, Yuxiao Wang, Liviu Panait, Sewoong Oh, Sean Augenstein, Ting Liu, Florian Schroff, H. Brendan McMahan

Small on-device models have been successfully trained with user-level differential privacy (DP) for next word prediction and image classification tasks in the past.

Federated Learning Image Classification

Decentralized Energy Market Integrating Carbon Allowance Trade and Uncertainty Balance in Energy Communities

no code implementations28 Jan 2023 Yuanxi Wu, Zhi Wu, Wei Gu, Zheng Xu, Shu Zheng, Qirun Sun

With the sustained attention on carbon neutrality, the personal carbon trading (PCT) scheme has been embraced as an auspicious paradigm for scaling down carbon emissions.

energy trading

On the Convergence of Federated Averaging with Cyclic Client Participation

no code implementations6 Feb 2023 Yae Jee Cho, Pranay Sharma, Gauri Joshi, Zheng Xu, Satyen Kale, Tong Zhang

Federated Averaging (FedAvg) and its variants are the most popular optimization algorithms in federated learning (FL).

Federated Learning

How to DP-fy ML: A Practical Guide to Machine Learning with Differential Privacy

1 code implementation1 Mar 2023 Natalia Ponomareva, Hussein Hazimeh, Alex Kurakin, Zheng Xu, Carson Denison, H. Brendan McMahan, Sergei Vassilvitskii, Steve Chien, Abhradeep Thakurta

However, while some adoption of DP has happened in industry, attempts to apply DP to real world complex ML models are still few and far between.

An Empirical Evaluation of Federated Contextual Bandit Algorithms

1 code implementation17 Mar 2023 Alekh Agarwal, H. Brendan McMahan, Zheng Xu

As the adoption of federated learning increases for learning from sensitive data local to user devices, it is natural to ask if the learning can be done using implicit signals generated as users interact with the applications of interest, rather than requiring access to explicit labels which can be difficult to acquire in many tasks.

Federated Learning Multi-Armed Bandits

User Inference Attacks on Large Language Models

no code implementations13 Oct 2023 Nikhil Kandpal, Krishna Pillutla, Alina Oprea, Peter Kairouz, Christopher A. Choquette-Choo, Zheng Xu

Fine-tuning is a common and effective method for tailoring large language models (LLMs) to specialized tasks and applications.

Analyze Drivers' Intervention Behavior During Autonomous Driving -- A VR-incorporated Approach

no code implementations4 Dec 2023 Zheng Xu

Given the rapid advance in ITS technologies, future mobility is pointing to vehicular autonomy.

Autonomous Driving

InstructPipe: Building Visual Programming Pipelines with Human Instructions

no code implementations15 Dec 2023 Zhongyi Zhou, Jing Jin, Vrushank Phadnis, Xiuxiu Yuan, Jun Jiang, Xun Qian, Jingtao Zhou, Yiyi Huang, Zheng Xu, yinda zhang, Kristen Wright, Jason Mayes, Mark Sherwood, Johnny Lee, Alex Olwal, David Kim, Ram Iyengar, Na Li, Ruofei Du

Our user study (N=16) showed that InstructPipe empowers novice users to streamline their workflow in creating desired ML pipelines, reduce their learning curve, and spark innovative ideas with open-ended commands.

Heterogeneous LoRA for Federated Fine-tuning of On-Device Foundation Models

no code implementations12 Jan 2024 Yae Jee Cho, Luyang Liu, Zheng Xu, Aldi Fahrezi, Gauri Joshi

Foundation models (FMs) adapt well to specific domains or tasks with fine-tuning, and federated learning (FL) enables the potential for privacy-preserving fine-tuning of the FMs with on-device local data.

Federated Learning Privacy Preserving

PaLM2-VAdapter: Progressively Aligned Language Model Makes a Strong Vision-language Adapter

no code implementations16 Feb 2024 Junfei Xiao, Zheng Xu, Alan Yuille, Shen Yan, Boyu Wang

Our research undertakes a thorough exploration of the state-of-the-art perceiver resampler architecture and builds a strong baseline.

Language Modelling Question Answering +1

Privacy-Preserving Instructions for Aligning Large Language Models

no code implementations21 Feb 2024 Da Yu, Peter Kairouz, Sewoong Oh, Zheng Xu

Service providers of large language model (LLM) applications collect user instructions in the wild and use them in further aligning LLMs with users' intentions.

Language Modelling Large Language Model +1

Enhancing Kubernetes Automated Scheduling with Deep Learning and Reinforcement Techniques for Large-Scale Cloud Computing Optimization

no code implementations26 Feb 2024 Zheng Xu, Yulu Gong, Yanlin Zhou, Qiaozhi Bao, Wenpin Qian

With the continuous expansion of the scale of cloud computing applications, artificial intelligence technologies such as Deep Learning and Reinforcement Learning have gradually become the key tools to solve the automated task scheduling of large-scale cloud computing systems.

Cloud Computing reinforcement-learning +1

Efficient Language Model Architectures for Differentially Private Federated Learning

no code implementations12 Mar 2024 Jae Hun Ro, Srinadh Bhojanapalli, Zheng Xu, Yanxiang Zhang, Ananda Theertha Suresh

Cross-device federated learning (FL) is a technique that trains a model on data distributed across typically millions of edge devices without data leaving the devices.

Computational Efficiency Federated Learning +1

Prompt Public Large Language Models to Synthesize Data for Private On-device Applications

no code implementations5 Apr 2024 Shanshan Wu, Zheng Xu, Yanxiang Zhang, Yuanbo Zhang, Daniel Ramage

Pre-training on public data is an effective method to improve the performance for federated learning (FL) with differential privacy (DP).

Federated Learning

Analyzing the effect of neural network architecture on training performance

no code implementations ICML 2020 Karthik Abinav Sankararaman, Soham De, Zheng Xu, W. Ronny Huang, Tom Goldstein

Through novel theoretical and experimental results, we show how the neural net architecture affects gradient confusion, and thus the efficiency of training.

Cannot find the paper you are looking for? You can Submit a new open access paper.