Search Results for author: Yiren Zhao

Found 52 papers, 19 papers with code

Hardware and Software Platform Inference

no code implementations7 Nov 2024 Cheng Zhang, Hanna Foerster, Robert D. Mullins, Yiren Zhao, Ilia Shumailov

We evaluate HSPI against models served on different real hardware and find that in a white-box setting we can distinguish between different \GPU{}s with between $83. 9\%$ and $100\%$ accuracy.

Large Language Model

Absorb & Escape: Overcoming Single Model Limitations in Generating Genomic Sequences

2 code implementations28 Oct 2024 Zehui Li, Yuhao Ni, Guoxuan Xia, William Beardall, Akashaditya Das, Guy-Bart Stan, Yiren Zhao

To overcome the limitations of both approaches, we propose a post-training sampling method, termed Absorb & Escape (A&E) to perform compositional generation from AR models and DMs.

Scaling Laws for Mixed quantization in Large Language Models

no code implementations9 Oct 2024 Zeyu Cao, Cheng Zhang, Pedro Gimenes, Jianqiao Lu, Jianyi Cheng, Yiren Zhao

Post-training quantization of Large Language Models (LLMs) has proven effective in reducing the computational requirements for running inference on these models.

Quantization

QERA: an Analytical Framework for Quantization Error Reconstruction

no code implementations8 Oct 2024 Cheng Zhang, Jeffrey T. H. Wong, Can Xiao, George A. Constantinides, Yiren Zhao

However, these heuristic methods lack an analytical solution to guide the design of quantization error reconstruction terms.

parameter-efficient fine-tuning Quantization

GV-Rep: A Large-Scale Dataset for Genetic Variant Representation Learning

1 code implementation24 Jul 2024 Zehui Li, Vallijah Subasri, Guy-Bart Stan, Yiren Zhao, Bo wang

Genetic variants (GVs) are defined as differences in the DNA sequences among individuals and play a crucial role in diagnosing and treating genetic diseases.

Representation Learning

Unlocking the Global Synergies in Low-Rank Adapters

no code implementations21 Jun 2024 Zixi Zhang, Cheng Zhang, Xitong Gao, Robert D. Mullins, George A. Constantinides, Yiren Zhao

We present HeteroLoRA, a light-weight search algorithm that leverages zero-cost proxies to allocate the limited LoRA trainable parameters across the model for better fine-tuned performance.

MRPC parameter-efficient fine-tuning

Optimised Grouped-Query Attention Mechanism for Transformers

no code implementations21 Jun 2024 Yuang Chen, Cheng Zhang, Xitong Gao, Robert D. Mullins, George A. Constantinides, Yiren Zhao

In this work, we propose AsymGQA, an activation-informed approach to asymmetrically grouping an MHA to a GQA for better model performance.

MMLU

HASS: Hardware-Aware Sparsity Search for Dataflow DNN Accelerator

1 code implementation5 Jun 2024 Zhewen Yu, Sudarshan Sreeram, Krish Agrawal, Junyi Wu, Alexander Montgomerie-Corcoran, Cheng Zhang, Jianyi Cheng, Christos-Savvas Bouganis, Yiren Zhao

We propose a Hardware-Aware Sparsity Search (HASS) to systematically determine an efficient sparsity solution for dataflow accelerators.

$Δ$-DiT: A Training-Free Acceleration Method Tailored for Diffusion Transformers

no code implementations3 Jun 2024 Pengtao Chen, Mingzhu Shen, Peng Ye, JianJian Cao, Chongjun Tu, Christos-Savvas Bouganis, Yiren Zhao, Tao Chen

Based on this insight, we propose an overall training-free inference acceleration framework $\Delta$-DiT: using a designed cache mechanism to accelerate the rear DiT blocks in the early sampling stages and the front DiT blocks in the later stages.

Image Generation

Locking Machine Learning Models into Hardware

no code implementations31 May 2024 Eleanor Clifford, Adhithya Saravanan, Harry Langford, Cheng Zhang, Yiren Zhao, Robert Mullins, Ilia Shumailov, Jamie Hayes

We demonstrate that locking mechanisms are feasible by either targeting efficiency of model representations, such making models incompatible with quantisation, or tie the model's operation on specific characteristics of hardware, such as number of cycles for arithmetic operations.

Enhancing Node Representations for Real-World Complex Networks with Topological Augmentation

1 code implementation20 Feb 2024 Xiangyu Zhao, Zehui Li, Mingzhu Shen, Guy-Bart Stan, Pietro Liò, Yiren Zhao

In this paper, we present Topological Augmentation (TopoAug), a novel graph augmentation method that builds a combinatorial complex from the original graph by constructing virtual hyperedges directly from the raw data.

Architectural Neural Backdoors from First Principles

no code implementations10 Feb 2024 Harry Langford, Ilia Shumailov, Yiren Zhao, Robert Mullins, Nicolas Papernot

In this work we construct an arbitrary trigger detector which can be used to backdoor an architecture with no human supervision.

DiscDiff: Latent Diffusion Model for DNA Sequence Generation

no code implementations8 Feb 2024 Zehui Li, Yuhao Ni, William A V Beardall, Guoxuan Xia, Akashaditya Das, Guy-Bart Stan, Yiren Zhao

This paper introduces a novel framework for DNA sequence generation, comprising two key components: DiscDiff, a Latent Diffusion Model (LDM) tailored for generating discrete DNA sequences, and Absorb-Escape, a post-training algorithm designed to refine these sequences.

Latent Diffusion Model for DNA Sequence Generation

1 code implementation9 Oct 2023 Zehui Li, Yuhao Ni, Tim August B. Huygelen, Akashaditya Das, Guoxuan Xia, Guy-Bart Stan, Yiren Zhao

On the other hand, Diffusion Models are a promising new class of generative models that are not burdened with these problems, enabling them to reach the state-of-the-art in domains such as image generation.

Text Generation

Revisiting Block-based Quantisation: What is Important for Sub-8-bit LLM Inference?

1 code implementation8 Oct 2023 Cheng Zhang, Jianyi Cheng, Ilia Shumailov, George A. Constantinides, Yiren Zhao

In this work, we explore the statistical and learning properties of the LLM layer and attribute the bottleneck of LLM quantisation to numerical scaling offsets.

Attribute

Will More Expressive Graph Neural Networks do Better on Generative Tasks?

no code implementations23 Aug 2023 Xiandong Zou, Xiangyu Zhao, Pietro Liò, Yiren Zhao

Moreover, we show that GCPN and GraphAF with advanced GNNs can achieve state-of-the-art results across 17 other non-GNN-based graph generative approaches, such as variational autoencoders and Bayesian optimisation models, on the proposed molecular generative objectives (DRD2, Median1, Median2), which are impor- tant metrics for de-novo molecular design.

Bayesian Optimisation Graph Generation +2

Hybrid Graph: A Unified Graph Representation with Datasets and Benchmarks for Complex Graphs

no code implementations8 Jun 2023 Zehui Li, Xiangyu Zhao, Mingzhu Shen, Guy-Bart Stan, Pietro Liò, Yiren Zhao

Additionally, though many Graph Neural Networks (GNNs) have been proposed for representation learning on higher-order graphs, they are usually only evaluated on simple graph datasets.

Graph Learning Representation Learning

The Curse of Recursion: Training on Generated Data Makes Models Forget

1 code implementation27 May 2023 Ilia Shumailov, Zakhar Shumaylov, Yiren Zhao, Yarin Gal, Nicolas Papernot, Ross Anderson

It is now clear that large language models (LLMs) are here to stay, and will bring about drastic change in the whole ecosystem of online text and images.

Descriptive

Revisiting Automated Prompting: Are We Actually Doing Better?

1 code implementation7 Apr 2023 Yulin Zhou, Yiren Zhao, Ilia Shumailov, Robert Mullins, Yarin Gal

Current literature demonstrates that Large Language Models (LLMs) are great few-shot learners, and prompting significantly increases their performance on a range of downstream tasks in a few-shot learning setting.

Few-Shot Learning

Dynamic Stashing Quantization for Efficient Transformer Training

no code implementations9 Mar 2023 Guo Yang, Daniel Lo, Robert Mullins, Yiren Zhao

Large Language Models (LLMs) have demonstrated impressive performance on a range of Natural Language Processing (NLP) tasks.

Quantization

Task-Agnostic Graph Neural Network Evaluation via Adversarial Collaboration

1 code implementation27 Jan 2023 Xiangyu Zhao, Hannes Stärk, Dominique Beaini, Yiren Zhao, Pietro Liò

Existing GNN benchmarking methods for molecular representation learning focus on comparing the GNNs' performances on some node/graph classification/regression tasks on certain datasets.

Benchmarking Graph Classification +4

Flareon: Stealthy any2any Backdoor Injection via Poisoned Augmentation

1 code implementation20 Dec 2022 Tianrui Qin, Xianghuan He, Xitong Gao, Yiren Zhao, Kejiang Ye, Cheng-Zhong Xu

Open software supply chain attacks, once successful, can exact heavy costs in mission-critical applications.

Data Augmentation

Revisiting Structured Dropout

no code implementations5 Oct 2022 Yiren Zhao, Oluwatomisin Dada, Xitong Gao, Robert D Mullins

Large neural networks are often overparameterised and prone to overfitting, Dropout is a widely used regularization technique to combat overfitting and improve model generalization.

Scheduling

DARTFormer: Finding The Best Type Of Attention

no code implementations2 Oct 2022 Jason Ross Brown, Yiren Zhao, Ilia Shumailov, Robert D Mullins

Given the wide and ever growing range of different efficient Transformer attention mechanisms, it is important to identify which attention is most effective when given a task.

ListOps Neural Architecture Search +3

Wide Attention Is The Way Forward For Transformers?

no code implementations2 Oct 2022 Jason Ross Brown, Yiren Zhao, Ilia Shumailov, Robert D Mullins

We demonstrate that wide single layer Transformer models can compete with or outperform deeper ones in a variety of Natural Language Processing (NLP) tasks when both are trained from scratch.

text-classification Text Classification

ImpNet: Imperceptible and blackbox-undetectable backdoors in compiled neural networks

no code implementations30 Sep 2022 Tim Clifford, Ilia Shumailov, Yiren Zhao, Ross Anderson, Robert Mullins

These backdoors are impossible to detect during the training or data preparation processes, because they are not yet present.

Augmentation Backdoors

1 code implementation29 Sep 2022 Joseph Rance, Yiren Zhao, Ilia Shumailov, Robert Mullins

It is well known that backdoors can be inserted into machine learning models through serving a modified dataset to train on.

Data Augmentation

Efficient Adversarial Training With Data Pruning

no code implementations1 Jul 2022 Maximilian Kaufmann, Yiren Zhao, Ilia Shumailov, Robert Mullins, Nicolas Papernot

In this paper we demonstrate data pruning-a method for increasing adversarial training efficiency through data sub-sampling. We empirically show that data pruning leads to improvements in convergence and reliability of adversarial training, albeit with different levels of utility degradation.

Model Architecture Adaption for Bayesian Neural Networks

no code implementations9 Feb 2022 Duo Wang, Yiren Zhao, Ilia Shumailov, Robert Mullins

Bayesian Neural Networks (BNNs) offer a mathematically grounded framework to quantify the uncertainty of model predictions but come with a prohibitive computation cost for both training and inference.

Uncertainty Quantification

DAdaQuant: Doubly-adaptive quantization for communication-efficient Federated Learning

no code implementations31 Oct 2021 Robert Hönig, Yiren Zhao, Robert Mullins

First, we introduce a time-adaptive quantization algorithm that increases the quantization level as training progresses.

Federated Learning Privacy Preserving +1

FedDrop: Trajectory-weighted Dropout for Efficient Federated Learning

no code implementations29 Sep 2021 Dongping Liao, Xitong Gao, Yiren Zhao, Hao Dai, Li Li, Kafeng Wang, Kejiang Ye, Yang Wang, Cheng-Zhong Xu

Federated learning (FL) enables edge clients to train collaboratively while preserving individual's data privacy.

Federated Learning

Rapid Model Architecture Adaption for Meta-Learning

no code implementations10 Sep 2021 Yiren Zhao, Xitong Gao, Ilia Shumailov, Nicolo Fusi, Robert Mullins

H-Meta-NAS shows a Pareto dominance compared to a variety of NAS and manual baselines in popular few-shot learning benchmarks with various hardware platforms and constraints.

Few-Shot Learning

Markpainting: Adversarial Machine Learning meets Inpainting

1 code implementation1 Jun 2021 David Khachaturov, Ilia Shumailov, Yiren Zhao, Nicolas Papernot, Ross Anderson

Inpainting is a learned interpolation technique that is based on generative modeling and used to populate masked or missing pieces in an image; it has wide applications in picture editing and retouching.

BIG-bench Machine Learning

Nudge Attacks on Point-Cloud DNNs

no code implementations22 Nov 2020 Yiren Zhao, Ilia Shumailov, Robert Mullins, Ross Anderson

The wide adaption of 3D point-cloud data in safety-critical applications such as autonomous driving makes adversarial samples a real threat.

Autonomous Driving

Learned Low Precision Graph Neural Networks

no code implementations19 Sep 2020 Yiren Zhao, Duo Wang, Daniel Bates, Robert Mullins, Mateja Jamnik, Pietro Lio

LPGNAS learns the optimal architecture coupled with the best quantisation strategy for different components in the GNN automatically using back-propagation in a single search round.

Sponge Examples: Energy-Latency Attacks on Neural Networks

2 code implementations5 Jun 2020 Ilia Shumailov, Yiren Zhao, Daniel Bates, Nicolas Papernot, Robert Mullins, Ross Anderson

The high energy costs of neural network training and inference led to the use of acceleration hardware such as GPUs and TPUs.

Autonomous Vehicles

Pay Attention to Features, Transfer Learn Faster CNNs

no code implementations ICLR 2020 Kafeng Wang, Xitong Gao, Yiren Zhao, Xingjian Li, Dejing Dou, Cheng-Zhong Xu

Deep convolutional neural networks are now widely deployed in vision applications, but a limited size of training data can restrict their task performance.

Transfer Learning

Probabilistic Dual Network Architecture Search on Graphs

no code implementations21 Mar 2020 Yiren Zhao, Duo Wang, Xitong Gao, Robert Mullins, Pietro Lio, Mateja Jamnik

We present the first differentiable Network Architecture Search (NAS) for Graph Neural Networks (GNNs).

Towards Certifiable Adversarial Sample Detection

no code implementations20 Feb 2020 Ilia Shumailov, Yiren Zhao, Robert Mullins, Ross Anderson

Convolutional Neural Networks (CNNs) are deployed in more and more classification systems, but adversarial samples can be maliciously crafted to trick them, and are becoming a real threat.

Adversarial Robustness

Blackbox Attacks on Reinforcement Learning Agents Using Approximated Temporal Information

no code implementations6 Sep 2019 Yiren Zhao, Ilia Shumailov, Han Cui, Xitong Gao, Robert Mullins, Ross Anderson

In this work, we show how such samples can be generalised from White-box and Grey-box attacks to a strong Black-box case, where the attacker has no knowledge of the agents, their training parameters and their training methods.

reinforcement-learning Reinforcement Learning +2

Focused Quantization for Sparse CNNs

1 code implementation NeurIPS 2019 Yiren Zhao, Xitong Gao, Daniel Bates, Robert Mullins, Cheng-Zhong Xu

In ResNet-50, we achieved a 18. 08x CR with only 0. 24% loss in top-5 accuracy, outperforming existing compression methods.

Neural Network Compression Quantization

The Taboo Trap: Behavioural Detection of Adversarial Samples

no code implementations18 Nov 2018 Ilia Shumailov, Yiren Zhao, Robert Mullins, Ross Anderson

Most existing detection mechanisms against adversarial attacksimpose significant costs, either by using additional classifiers to spot adversarial samples, or by requiring the DNN to be restructured.

Dynamic Channel Pruning: Feature Boosting and Suppression

2 code implementations ICLR 2019 Xitong Gao, Yiren Zhao, Łukasz Dudziak, Robert Mullins, Cheng-Zhong Xu

Making deep convolutional neural networks more accurate typically comes at the cost of increased computational and memory resources.

Model Compression Network Pruning

Cannot find the paper you are looking for? You can Submit a new open access paper.