Search Results for author: Randal Burns

Found 19 papers, 7 papers with code

FlashR: R-Programmed Parallel and Scalable Machine Learning using SSDs

2 code implementations21 Apr 2016 Da Zheng, Disa Mhembere, Joshua T. Vogelstein, Carey E. Priebe, Randal Burns

R is one of the most popular programming languages for statistics and machine learning, but the R framework is relatively slow and unable to scale to large datasets.

Distributed, Parallel, and Cluster Computing

Semi-External Memory Sparse Matrix Multiplication for Billion-Node Graphs

2 code implementations9 Feb 2016 Da Zheng, Disa Mhembere, Vince Lyzinski, Joshua Vogelstein, Carey E. Priebe, Randal Burns

In contrast, we scale sparse matrix multiplication beyond memory capacity by implementing sparse matrix dense matrix multiplication (SpMM) in a semi-external memory (SEM) fashion; i. e., we keep the sparse matrix on commodity SSDs and dense matrices in memory.

Distributed, Parallel, and Cluster Computing

Active Community Detection in Massive Graphs

2 code implementations30 Dec 2014 Heng Wang, Da Zheng, Randal Burns, Carey Priebe

A canonical problem in graph mining is the detection of dense communities.

Social and Information Networks Physics and Society

Sparse Projection Oblique Randomer Forests

2 code implementations10 Jun 2015 Tyler M. Tomita, James Browne, Cencheng Shen, Jaewon Chung, Jesse L. Patsolic, Benjamin Falk, Jason Yim, Carey E. Priebe, Randal Burns, Mauro Maggioni, Joshua T. Vogelstein

Unfortunately, these extensions forfeit one or more of the favorable properties of decision forests based on axis-aligned splits, such as robustness to many noise dimensions, interpretability, or computational efficiency.

Computational Efficiency

Supervised Dimensionality Reduction for Big Data

1 code implementation5 Sep 2017 Joshua T. Vogelstein, Eric Bridgeford, Minh Tang, Da Zheng, Christopher Douville, Randal Burns, Mauro Maggioni

To solve key biomedical problems, experimentalists now routinely measure millions or billions of features (dimensions) per sample, with the hope that data science techniques will be able to build accurate data-driven inferences.

Computational Efficiency General Classification +2

knor: A NUMA-Optimized In-Memory, Distributed and Semi-External-Memory k-means Library

1 code implementation28 Jun 2016 Disa Mhembere, Da Zheng, Carey E. Priebe, Joshua T. Vogelstein, Randal Burns

The \textit{k-means NUMA Optimized Routine} (\textsf{knor}) library has (i) in-memory (\textsf{knori}), (ii) distributed memory (\textsf{knord}), and (iii) semi-external memory (\textsf{knors}) modules that radically improve the performance of k-means for varying memory and hardware budgets.

Distributed, Parallel, and Cluster Computing

Edge-Parallel Graph Encoder Embedding

1 code implementation6 Feb 2024 Ariel Lubonja, Cencheng Shen, Carey Priebe, Randal Burns

New algorithms for embedding graphs have reduced the asymptotic complexity of finding low-dimensional representations.

VESICLE: Volumetric Evaluation of Synaptic Interfaces using Computer vision at Large Scale

no code implementations14 Mar 2014 William Gray Roncal, Michael Pekala, Verena Kaynig-Fittkau, Dean M. Kleissas, Joshua T. Vogelstein, Hanspeter Pfister, Randal Burns, R. Jacob Vogelstein, Mark A. Chevillet, Gregory D. Hager

An open challenge problem at the forefront of modern neuroscience is to obtain a comprehensive mapping of the neural pathways that underlie human brain function; an enhanced understanding of the wiring diagram of the brain promises to lead to new breakthroughs in diagnosing and treating neurological disorders.

object-detection Object Detection

Automatic Annotation of Axoplasmic Reticula in Pursuit of Connectomes

no code implementations16 Apr 2014 Ayushi Sinha, William Gray Roncal, Narayanan Kasthuri, Ming Chuang, Priya Manavalan, Dean M. Kleissas, Joshua T. Vogelstein, R. Jacob Vogelstein, Randal Burns, Jeff W. Lichtman, Michael Kazhdan

The contribution of this work is the introduction of a straightforward and robust pipeline which annotates axoplasmic reticula with high precision, contributing towards advancements in automatic feature annotations in neural EM data.

Geodesic Learning via Unsupervised Decision Forests

no code implementations5 Jul 2019 Meghana Madhyastha, Percy Li, James Browne, Veronika Strnadova-Neeley, Carey E. Priebe, Randal Burns, Joshua T. Vogelstein

Empirical results on simulated and real data demonstrate that URerF is robust to high-dimensional noise, where as other methods, such as Isomap, UMAP, and FLANN, quickly deteriorate in such settings.

Graphyti: A Semi-External Memory Graph Library for FlashGraph

no code implementations7 Jul 2019 Disa Mhembere, Da Zheng, Carey E. Priebe, Joshua T. Vogelstein, Randal Burns

Emerging frameworks avoid the network bottleneck of distributed data with Semi-External Memory (SEM) that uses a single multicore node and operates on graphs larger than memory.

Distributed, Parallel, and Cluster Computing Databases

PACSET (Packed Serialized Trees): Reducing Inference Latency for Tree Ensemble Deployment

no code implementations10 Nov 2020 Meghana Madhyastha, Kunal Lillaney, James Browne, Joshua Vogelstein, Randal Burns

We present methods to serialize and deserialize tree ensembles that optimize inference latency when models are not already loaded into memory.

Understanding Patterns of Deep Learning ModelEvolution in Network Architecture Search

no code implementations22 Sep 2023 Robert Underwood, Meghana Madhastha, Randal Burns, Bogdan Nicolae

We describe how evolutionary patterns appear in distributed settings and opportunities for caching and improved scheduling.

Scheduling

Masked Matrix Multiplication for Emergent Sparsity

no code implementations21 Feb 2024 Brian Wheatman, Meghana Madhyastha, Randal Burns

Artificial intelligence workloads, especially transformer models, exhibit emergent sparsity in which computations perform selective sparse access to dense data.

Cannot find the paper you are looking for? You can Submit a new open access paper.