Search Results for author: Charles O'Neill

Found 14 papers, 3 papers with code

Sparks of Science: Hypothesis Generation Using Structured Paper Data

no code implementations17 Apr 2025 Charles O'Neill, Tirthankar Ghosal, Roberta Răileanu, Mike Walmsley, Thang Bui, Kevin Schawinski, Ioana Ciucă

We demonstrate that framing hypothesis generation as conditional language modelling, with the model fine-tuned on Bit-Flip-Spark and the Chain-of-Reasoning (and where, at inference, we only provide the Bit), leads to improvements in the overall quality of the hypotheses.

Language Modelling Text Generation

From superposition to sparse codes: interpretable representations in neural networks

no code implementations3 Mar 2025 David Klindt, Charles O'Neill, Patrik Reizinger, Harald Maurer, Nina Miolane

By bridging insights from theoretical neuroscience, representation learning, and interpretability research, we propose an emerging perspective on understanding neural representations in both artificial and biological systems.

compressed sensing Representation Learning

Self-Attention as a Parametric Endofunctor: A Categorical Framework for Transformer Architectures

no code implementations6 Jan 2025 Charles O'Neill

Our results build on and extend recent work on category-theoretic foundations for deep learning, offering deeper insights into the algebraic structure of attention mechanisms.

Position

Compute Optimal Inference and Provable Amortisation Gap in Sparse Autoencoders

no code implementations20 Nov 2024 Charles O'Neill, Alim Gumran, David Klindt

We demonstrate this generalises to SAEs applied to large language models, where more expressive encoders achieve greater interpretability.

compressed sensing Language Modeling +2

Disentangling Dense Embeddings with Sparse Autoencoders

no code implementations1 Aug 2024 Charles O'Neill, Christine Ye, Kartheik Iyer, John F. Wu

Sparse autoencoders (SAEs) have shown promise in extracting interpretable features from complex neural networks.

Astronomy

Sparse Autoencoders Enable Scalable and Reliable Circuit Identification in Language Models

no code implementations21 May 2024 Charles O'Neill, Thang Bui

We propose training sparse autoencoders on carefully designed positive and negative examples, where the model can only correctly predict the next token for the positive examples.

Measuring Sharpness in Grokking

1 code implementation14 Feb 2024 Jack Miller, Patrick Gleeson, Charles O'Neill, Thang Bui, Noam Levi

Neural networks sometimes exhibit grokking, a phenomenon where perfect or near-perfect performance is achieved on a validation set well after the same performance has been obtained on the corresponding training set.

Grokking Beyond Neural Networks: An Empirical Exploration with Model Complexity

1 code implementation26 Oct 2023 Jack Miller, Charles O'Neill, Thang Bui

In some settings neural networks exhibit a phenomenon known as \textit{grokking}, where they achieve perfect or near-perfect accuracy on the validation set long after the same performance has been achieved on the training set.

regression

Adversarial Fine-Tuning of Language Models: An Iterative Optimisation Approach for the Generation and Detection of Problematic Content

no code implementations26 Aug 2023 Charles O'Neill, Jack Miller, Ioana Ciuca, Yuan-Sen Ting, Thang Bui

The performance of our approach is evaluated through classification accuracy on a dataset consisting of problematic prompts not detected by GPT-4, as well as a selection of contentious but unproblematic prompts.

Steering Language Generation: Harnessing Contrastive Expert Guidance and Negative Prompting for Coherent and Diverse Synthetic Data Generation

no code implementations15 Aug 2023 Charles O'Neill, Yuan-Sen Ting, Ioana Ciuca, Jack Miller, Thang Bui

Large Language Models (LLMs) hold immense potential to generate synthetic data of high quality and utility, which has numerous applications from downstream model training to practical data utilisation.

Comment Generation Diversity +2

Rice paddy disease classifications using CNNs

no code implementations15 Mar 2023 Charles O'Neill

Rice is a staple food in the world's diet, and yet huge percentages of crop yields are lost each year to disease.

Eigenvalue initialisation and regularisation for Koopman autoencoders

no code implementations23 Dec 2022 Jack W. Miller, Charles O'Neill, Navid C. Constantinou, Omri Azencot

In addition, we suggest the "eigenloss" penalty scheme that penalises the eigenvalues of the Koopman operator during training.

Inductive Bias

Cannot find the paper you are looking for? You can Submit a new open access paper.