Search Results for author: Qirong Ho

Found 30 papers, 3 papers with code

Learning in Chaos: Efficient Autoscaling and Self-healing for Distributed Training at the Edge

no code implementations19 May 2025 Wenjiao Feng, Rongxing Xiao, Zonghang Li, Hongfang Yu, Gang Sun, Long Luo, Mohsen Guizani, Qirong Ho

Frequent node and link changes in edge AI clusters disrupt distributed training, while traditional checkpoint-based recovery and cloud-centric autoscaling are too slow for scale-out and ill-suited to chaotic and self-governed edge.

Scheduling

Token Level Routing Inference System for Edge Devices

no code implementations10 Apr 2025 Jianshu She, Wenhao Zheng, Zhengzhong Liu, Hongyi Wang, Eric Xing, Huaxiu Yao, Qirong Ho

This paradigm leverages the strengths of both model types by enabling high-quality inference through selective intervention of the large model, while maintaining the speed and efficiency of the smaller model.

Language Modeling Language Modelling +1

Hawkeye:Efficient Reasoning with Model Collaboration

no code implementations1 Apr 2025 Jianshu She, Zhuohao Li, Zhemin Huang, Qi Li, Peiran Xu, Haonan Li, Qirong Ho

Chain-of-Thought (CoT) reasoning has demonstrated remarkable effectiveness in enhancing the reasoning abilities of large language models (LLMs).

Math model +1

J-Invariant Volume Shuffle for Self-Supervised Cryo-Electron Tomogram Denoising on Single Noisy Volume

no code implementations22 Nov 2024 Xiwei Liu, Mohamad Kassab, Min Xu, Qirong Ho

Cryo-Electron Tomography (Cryo-ET) enables detailed 3D visualization of cellular structures in near-native states but suffers from low signal-to-noise ratio due to imaging constraints.

Denoising Electron Tomography +1

FIAS: Feature Imbalance-Aware Medical Image Segmentation with Dynamic Fusion and Mixing Attention

no code implementations16 Nov 2024 Xiwei Liu, Min Xu, Qirong Ho

To address these issues, we propose a Feaure Imbalance-Aware Segmentation (FIAS) network, which incorporates a dual-path encoder and a novel Mixing Attention (MixAtt) decoder.

Decoder Image Segmentation +3

Reducing Hyperparameter Tuning Costs in ML, Vision and Language Model Training Pipelines via Memoization-Awareness

1 code implementation6 Nov 2024 Abdelmajid Essofi, Ridwan Salahuddeen, Munachiso Nwadike, Elnura Zhalieva, Kun Zhang, Eric Xing, Willie Neiswanger, Qirong Ho

The training or fine-tuning of machine learning, vision, and language models is often implemented as a pipeline: a sequence of stages encompassing data preparation, model training and evaluation.

Bayesian Optimization Language Modeling +1

Continual Learning of Nonlinear Independent Representations

no code implementations11 Aug 2024 Boyang Sun, Ignavier Ng, Guangyi Chen, Yifan Shen, Qirong Ho, Kun Zhang

Identifying the causal relations between interested variables plays a pivotal role in representation learning as it provides deep insights into the dataset.

Continual Learning Lifelong learning +1

Multi-level Adaptive Contrastive Learning for Knowledge Internalization in Dialogue Generation

no code implementations13 Oct 2023 Chenxu Yang, Zheng Lin, Lanrui Wang, Chong Tian, Liang Pang, Jiangnan Li, Qirong Ho, Yanan Cao, Weiping Wang

Knowledge-grounded dialogue generation aims to mitigate the issue of text degeneration by incorporating external knowledge to supplement the context.

Contrastive Learning Dialogue Generation

On Optimizing the Communication of Model Parallelism

no code implementations10 Nov 2022 Yonghao Zhuang, Hexu Zhao, Lianmin Zheng, Zhuohan Li, Eric P. Xing, Qirong Ho, Joseph E. Gonzalez, Ion Stoica, Hao Zhang

This pattern emerges when the two paradigms of model parallelism - intra-operator and inter-operator parallelism - are combined to support large models on large clusters.

model

Pollux: Co-adaptive Cluster Scheduling for Goodput-Optimized Deep Learning

2 code implementations27 Aug 2020 Aurick Qiao, Sang Keun Choe, Suhas Jayaram Subramanya, Willie Neiswanger, Qirong Ho, Hao Zhang, Gregory R. Ganger, Eric P. Xing

Some recent schedulers choose job resources for users, but do so without awareness of how DL training can be re-optimized to better utilize the provided resources.

Deep Learning Fairness +1

Cavs: A Vertex-centric Programming Interface for Dynamic Neural Networks

no code implementations11 Dec 2017 Hao Zhang, Shizhen Xu, Graham Neubig, Wei Dai, Qirong Ho, Guangwen Yang, Eric P. Xing

Recent deep learning (DL) models have moved beyond static network architectures to dynamic ones, handling data where the network structure changes every example, such as sequences of variable lengths, trees, and graphs.

Dynamic neural networks graph construction +2

Poseidon: An Efficient Communication Architecture for Distributed Deep Learning on GPU Clusters

no code implementations11 Jun 2017 Hao Zhang, Zeyu Zheng, Shizhen Xu, Wei Dai, Qirong Ho, Xiaodan Liang, Zhiting Hu, Jinliang Wei, Pengtao Xie, Eric P. Xing

We show that Poseidon enables Caffe and TensorFlow to achieve 15. 5x speed-up on 16 single-GPU machines, even with limited bandwidth (10GbE) and the challenging VGG19-22K network for image classification.

image-classification Image Classification

Distributed Multi-Task Relationship Learning

no code implementations13 Dec 2016 Sulin Liu, Sinno Jialin Pan, Qirong Ho

Due to heavy communication caused by transmitting the data and the issue of data privacy and security, it is impossible to send data of different task to a master machine to perform multi-task learning.

Distributed Optimization Multi-Task Learning

Strategies and Principles of Distributed Machine Learning on Big Data

no code implementations31 Dec 2015 Eric P. Xing, Qirong Ho, Pengtao Xie, Wei Dai

Taking the view that Big ML systems can benefit greatly from ML-rooted statistical and algorithmic insights --- and that ML researchers should therefore not shy away from such systems design --- we discuss a series of principles and strategies distilled from our recent efforts on industrial-scale ML solutions.

BIG-bench Machine Learning

Poseidon: A System Architecture for Efficient GPU-based Deep Learning on Multiple Machines

no code implementations19 Dec 2015 Hao Zhang, Zhiting Hu, Jinliang Wei, Pengtao Xie, Gunhee Kim, Qirong Ho, Eric Xing

To investigate how to adapt existing frameworks to efficiently support distributed GPUs, we propose Poseidon, a scalable system architecture for distributed inter-machine communication in existing DL frameworks.

Object Recognition

Distributed Machine Learning via Sufficient Factor Broadcasting

no code implementations26 Nov 2015 Pengtao Xie, Jin Kyu Kim, Yi Zhou, Qirong Ho, Abhimanu Kumar, Yao-Liang Yu, Eric Xing

Matrix-parametrized models, including multiclass logistic regression and sparse coding, are used in machine learning (ML) applications ranging from computer vision to computational biology.

BIG-bench Machine Learning

LightLDA: Big Topic Models on Modest Compute Clusters

1 code implementation4 Dec 2014 Jinhui Yuan, Fei Gao, Qirong Ho, Wei Dai, Jinliang Wei, Xun Zheng, Eric P. Xing, Tie-Yan Liu, Wei-Ying Ma

When building large-scale machine learning (ML) programs, such as big topic models or deep neural nets, one usually assumes such tasks can only be attempted with industrial-sized clusters with thousands of nodes, which are out of reach for most practitioners or academic researchers.

Topic Models

On Model Parallelization and Scheduling Strategies for Distributed Machine Learning

no code implementations NeurIPS 2014 Seunghak Lee, Jin Kyu Kim, Xun Zheng, Qirong Ho, Garth A. Gibson, Eric P. Xing

Distributed machine learning has typically been approached from a data parallel perspective, where big data are partitioned to multiple workers and an algorithm is executed concurrently over different data subsets under various synchronization schemes to ensure speed-up and/or correctness.

BIG-bench Machine Learning Scheduling

Model-Parallel Inference for Big Topic Models

no code implementations10 Nov 2014 Xun Zheng, Jin Kyu Kim, Qirong Ho, Eric P. Xing

In real world industrial applications of topic modeling, the ability to capture gigantic conceptual space by learning an ultra-high dimensional topical representation, i. e., the so-called "big model", is becoming the next desideratum after enthusiasms on "big data", especially for fine-grained downstream tasks such as online advertising, where good performances are usually achieved by regression-based predictors built on millions if not billions of input features.

model Topic Models

High-Performance Distributed ML at Scale through Parameter Server Consistency Models

no code implementations29 Oct 2014 Wei Dai, Abhimanu Kumar, Jinliang Wei, Qirong Ho, Garth Gibson, Eric P. Xing

As Machine Learning (ML) applications increase in data size and model complexity, practitioners turn to distributed clusters to satisfy the increased computational and memory demands.

Vocal Bursts Intensity Prediction

Distributed Machine Learning via Sufficient Factor Broadcasting

no code implementations19 Sep 2014 Pengtao Xie, Jin Kyu Kim, Yi Zhou, Qirong Ho, Abhimanu Kumar, Yao-Liang Yu, Eric Xing

Matrix-parametrized models, including multiclass logistic regression and sparse coding, are used in machine learning (ML) applications ranging from computer vision to computational biology.

BIG-bench Machine Learning

Primitives for Dynamic Big Model Parallelism

no code implementations18 Jun 2014 Seunghak Lee, Jin Kyu Kim, Xun Zheng, Qirong Ho, Garth A. Gibson, Eric P. Xing

When training large machine learning models with many variables or parameters, a single machine is often inadequate since the model may be too large to fit in memory, while training can take a long time even with stochastic updates.

model Scheduling

Consistent Bounded-Asynchronous Parameter Servers for Distributed ML

no code implementations30 Dec 2013 Jinliang Wei, Wei Dai, Abhimanu Kumar, Xun Zheng, Qirong Ho, Eric P. Xing

Many ML algorithms fall into the category of \emph{iterative convergent algorithms} which start from a randomly chosen initial point and converge to optima by repeating iteratively a set of procedures.

Petuum: A New Platform for Distributed Machine Learning on Big Data

no code implementations30 Dec 2013 Eric P. Xing, Qirong Ho, Wei Dai, Jin Kyu Kim, Jinliang Wei, Seunghak Lee, Xun Zheng, Pengtao Xie, Abhimanu Kumar, Yao-Liang Yu

What is a systematic way to efficiently apply a wide spectrum of advanced ML programs to industrial scale problems, using Big Models (up to 100s of billions of parameters) on Big Data (up to terabytes or petabytes)?

BIG-bench Machine Learning Scheduling

Structure-Aware Dynamic Scheduler for Parallel Machine Learning

no code implementations19 Dec 2013 Seunghak Lee, Jin Kyu Kim, Qirong Ho, Garth A. Gibson, Eric P. Xing

Training large machine learning (ML) models with many variables or parameters can take a long time if one employs sequential procedures even with stochastic updates.

BIG-bench Machine Learning Distributed Computing

More Effective Distributed ML via a Stale Synchronous Parallel Parameter Server

no code implementations NeurIPS 2013 Qirong Ho, James Cipar, Henggang Cui, Seunghak Lee, Jin Kyu Kim, Phillip B. Gibbons, Garth A. Gibson, Greg Ganger, Eric P. Xing

We propose a parameter server system for distributed ML, which follows a Stale Synchronous Parallel (SSP) model of computation that maximizes the time computational workers spend doing useful work on ML algorithms, while still providing correctness guarantees.

On Triangular versus Edge Representations --- Towards Scalable Modeling of Networks

no code implementations NeurIPS 2012 Qirong Ho, Junming Yin, Eric P. Xing

A triangular motif is a vertex triple containing 2 or 3 edges, and the number of such motifs is $\Theta(\sum_{i}D_{i}^{2})$ (where $D_i$ is the degree of vertex $i$), which is much smaller than $N^2$ for low-maximum-degree networks.

Community Detection

Cannot find the paper you are looking for? You can Submit a new open access paper.