Search Results for author: Max Tegmark

Found 73 papers, 36 papers with code

Towards Understanding Distilled Reasoning Models: A Representational Approach

no code implementations5 Mar 2025 David D. Baek, Max Tegmark

In this paper, we investigate how model distillation impacts the development of reasoning features in large language models (LLMs).

Are Sparse Autoencoders Useful? A Case Study in Sparse Probing

1 code implementation23 Feb 2025 Subhash Kantamneni, Joshua Engels, Senthooran Rajamanoharan, Max Tegmark, Neel Nanda

Sparse autoencoders (SAEs) are a popular method for interpreting concepts represented in large language model (LLM) activations.

Inductive Bias Large Language Model

Harmonic Loss Trains Interpretable AI Models

1 code implementation3 Feb 2025 David D. Baek, Ziming Liu, Riya Tyagi, Max Tegmark

In this paper, we introduce **harmonic loss** as an alternative to the standard cross-entropy loss for training neural networks and large language models (LLMs).

Efficient Neural Network

Language Models Use Trigonometry to Do Addition

no code implementations2 Feb 2025 Subhash Kantamneni, Max Tegmark

By demonstrating that LLMs represent numbers on a helix and manipulate this helix to perform addition, we present the first representation-level explanation of an LLM's mathematical capability.

Language Modeling Language Modelling +2

Low-Rank Adapting Models for Sparse Autoencoders

1 code implementation31 Jan 2025 Matthew Chen, Joshua Engels, Max Tegmark

In these settings, our method reduces the cross entropy loss gap by 30% to 55% when SAEs are inserted during the forward pass.

Language Modeling Language Modelling

Physics of Skill Learning

1 code implementation21 Jan 2025 Ziming Liu, Yizhou Liu, Eric J. Michaud, Jeff Gore, Max Tegmark

We aim to understand physics of skill learning, i. e., how skills are learned in neural networks during training.

Decomposing The Dark Matter of Sparse Autoencoders

1 code implementation18 Oct 2024 Joshua Engels, Logan Riggs, Max Tegmark

Sparse autoencoders (SAEs) are a promising technique for decomposing language model activations into interpretable linear features.

Generalization from Starvation: Hints of Universality in LLM Knowledge Graph Learning

no code implementations10 Oct 2024 David D. Baek, Yuxiao Li, Max Tegmark

Motivated by interpretability and reliability, we investigate how neural networks represent knowledge during graph learning, We find hints of universality, where equivalent representations are learned across a range of model sizes (from $10^2$ to $10^9$ parameters) and contexts (MLP toy models, LLM in-context learning and LLM training).

Graph Learning In-Context Learning

The Geometry of Concepts: Sparse Autoencoder Feature Structure

no code implementations10 Oct 2024 Yuxiao Li, Eric J. Michaud, David D. Baek, Joshua Engels, Xiaoqing Sun, Max Tegmark

Sparse autoencoders have recently produced dictionaries of high-dimensional vectors corresponding to the universe of concepts represented by large language models.

Math

Efficient Dictionary Learning with Switch Sparse Autoencoders

1 code implementation10 Oct 2024 Anish Mudide, Joshua Engels, Eric J. Michaud, Max Tegmark, Christian Schroeder de Witt

We present experiments comparing Switch SAEs with other SAE architectures, and find that Switch SAEs deliver a substantial Pareto improvement in the reconstruction vs. sparsity frontier for a given fixed training compute budget.

Dictionary Learning

KAN 2.0: Kolmogorov-Arnold Networks Meet Science

1 code implementation19 Aug 2024 Ziming Liu, Pingchuan Ma, YiXuan Wang, Wojciech Matusik, Max Tegmark

The synergy is bidirectional: science to KAN (incorporating scientific knowledge into KANs), and KAN to science (extracting scientific insights from KANs).

Kolmogorov-Arnold Networks scientific discovery

The Remarkable Robustness of LLMs: Stages of Inference?

1 code implementation27 Jun 2024 Vedang Lad, Wes Gurnee, Max Tegmark

We find that deleting and swapping interventions retain 72-95\% of the original model's prediction accuracy without fine-tuning, whereas models with more layers exhibit more robustness.

Feature Engineering Prediction

DafnyBench: A Benchmark for Formal Software Verification

1 code implementation12 Jun 2024 Chloe Loughridge, Qinyi Sun, Seth Ahrenbach, Federico Cassano, Chuyue Sun, Ying Sheng, Anish Mudide, Md Rakib Hossain Misu, Nada Amin, Max Tegmark

We introduce DafnyBench, the largest benchmark of its kind for training and evaluating machine learning systems for formal software verification.

Survival of the Fittest Representation: A Case Study with Modular Addition

1 code implementation27 May 2024 Xiaoman Delores Ding, Zifan Carl Guo, Eric J. Michaud, Ziming Liu, Max Tegmark

To investigate this Survival of the Fittest hypothesis, we conduct a case study on neural networks performing modular addition, and find that these networks' multiple circular representations at different Fourier frequencies undergo such competitive dynamics, with only a few circles surviving at the end.

How Do Transformers "Do" Physics? Investigating the Simple Harmonic Oscillator

no code implementations23 May 2024 Subhash Kantamneni, Ziming Liu, Max Tegmark

We develop four criteria for the use of a method within the simple testbed of linear regression, where our method is $y = wx$ and our intermediate is $w$: (1) Can the intermediate be predicted from hidden states?

Not All Language Model Features Are Linear

1 code implementation23 May 2024 Joshua Engels, Eric J. Michaud, Isaac Liao, Wes Gurnee, Max Tegmark

Recent work has proposed that language models perform computation by manipulating one-dimensional representations of concepts ("features") in activation space.

All Language Modeling +2

Towards Guaranteed Safe AI: A Framework for Ensuring Robust and Reliable AI Systems

no code implementations10 May 2024 David "davidad" Dalrymple, Joar Skalse, Yoshua Bengio, Stuart Russell, Max Tegmark, Sanjit Seshia, Steve Omohundro, Christian Szegedy, Ben Goldhaber, Nora Ammann, Alessandro Abate, Joe Halpern, Clark Barrett, Ding Zhao, Tan Zhi-Xuan, Jeannette Wing, Joshua Tenenbaum

Ensuring that AI systems reliably and robustly avoid harmful or dangerous behaviours is a crucial challenge, especially for AI systems with a high degree of autonomy and general intelligence, or systems used in safety-critical contexts.

OptPDE: Discovering Novel Integrable Systems via AI-Human Collaboration

no code implementations7 May 2024 Subhash Kantamneni, Ziming Liu, Max Tegmark

Integrable partial differential equation (PDE) systems are of great interest in natural science, but are exceedingly rare and difficult to discover.

KAN: Kolmogorov-Arnold Networks

25 code implementations30 Apr 2024 Ziming Liu, YiXuan Wang, Sachin Vaidya, Fabian Ruehle, James Halverson, Marin Soljačić, Thomas Y. Hou, Max Tegmark

Inspired by the Kolmogorov-Arnold representation theorem, we propose Kolmogorov-Arnold Networks (KANs) as promising alternatives to Multi-Layer Perceptrons (MLPs).

Kolmogorov-Arnold Networks

GenEFT: Understanding Statics and Dynamics of Model Generalization via Effective Theory

no code implementations8 Feb 2024 David D. Baek, Ziming Liu, Max Tegmark

We present GenEFT: an effective theory framework for shedding light on the statics and dynamics of neural network generalization, and illustrate it with graph learning examples.

Decoder Graph Learning +1

A Resource Model For Neural Scaling Law

no code implementations7 Feb 2024 Jinyeop Song, Ziming Liu, Max Tegmark, Jeff Gore

A task is usually composite hence can be decomposed into many subtasks, which compete for resources (measured by the number of neurons allocated to subtasks).

model

Opening the AI black box: program synthesis via mechanistic interpretability

1 code implementation7 Feb 2024 Eric J. Michaud, Isaac Liao, Vedang Lad, Ziming Liu, Anish Mudide, Chloe Loughridge, Zifan Carl Guo, Tara Rezaei Kheirkhah, Mateja Vukelić, Max Tegmark

We present MIPS, a novel method for program synthesis based on automated mechanistic interpretability of neural networks trained to perform the desired task, auto-distilling the learned algorithm into Python code.

Program Synthesis Symbolic Regression

Generating Interpretable Networks using Hypernetworks

no code implementations5 Dec 2023 Isaac Liao, Ziming Liu, Max Tegmark

The hypernetwork is carefully designed such that it can control network complexity, leading to a diverse family of interpretable algorithms ranked by their complexity.

Systematic Generalization

Growing Brains: Co-emergence of Anatomical and Functional Modularity in Recurrent Neural Networks

no code implementations11 Oct 2023 Ziming Liu, Mikail Khona, Ila R. Fiete, Max Tegmark

Recurrent neural networks (RNNs) trained on compositional tasks can exhibit functional modularity, in which neurons can be clustered by activity similarity and participation in shared computational subtasks.

Clustering

The Geometry of Truth: Emergent Linear Structure in Large Language Model Representations of True/False Datasets

1 code implementation10 Oct 2023 Samuel Marks, Max Tegmark

In this work, we use high-quality datasets of simple true/false statements to study in detail the structure of LLM representations of truth, drawing on three lines of evidence: 1.

Language Modeling Language Modelling +1

Divide-and-Conquer Dynamics in AI-Driven Disempowerment

no code implementations9 Oct 2023 Peter S. Park, Max Tegmark

Myopic members prioritize their future well-being less than their present well-being, and are thus disinclined to solidarily support current victims today at personal cost, even if this is necessary to counter the shared threat of AI-driven disempowerment.

Grokking as Compression: A Nonlinear Complexity Perspective

no code implementations9 Oct 2023 Ziming Liu, Ziqian Zhong, Max Tegmark

To do so, we define linear mapping number (LMN) to measure network complexity, which is a generalized version of linear region number for ReLU networks.

Attribute Memorization +1

A Neural Scaling Law from Lottery Ticket Ensembling

no code implementations3 Oct 2023 Ziming Liu, Max Tegmark

Neural scaling laws (NSL) refer to the phenomenon where model performance improves with scale.

Attribute

Language Models Represent Space and Time

2 code implementations3 Oct 2023 Wes Gurnee, Max Tegmark

The capabilities of large language models (LLMs) have sparked debate over whether such systems just learn an enormous collection of superficial statistics or a set of more coherent and grounded representations that reflect the real world.

Provably safe systems: the only path to controllable AGI

no code implementations5 Sep 2023 Max Tegmark, Steve Omohundro

We describe a path to humanity safely thriving with powerful Artificial General Intelligences (AGIs) by building them to provably satisfy human-specified requirements.

The Clock and the Pizza: Two Stories in Mechanistic Explanation of Neural Networks

2 code implementations NeurIPS 2023 Ziqian Zhong, Ziming Liu, Max Tegmark, Jacob Andreas

Do neural networks, trained on well-understood algorithmic tasks, reliably rediscover known algorithms for solving those tasks?

Diversity

Discovering New Interpretable Conservation Laws as Sparse Invariants

1 code implementation31 May 2023 Ziming Liu, Patrick Obin Sturm, Saketh Bharadwaj, Sam Silva, Max Tegmark

Discovering conservation laws for a given dynamical system is important but challenging.

Seeing is Believing: Brain-Inspired Modular Training for Mechanistic Interpretability

1 code implementation4 May 2023 Ziming Liu, Eric Gan, Max Tegmark

We introduce Brain-Inspired Modular Training (BIMT), a method for making neural networks more modular and interpretable.

GenPhys: From Physical Processes to Generative Models

no code implementations5 Apr 2023 Ziming Liu, Di Luo, Yilun Xu, Tommi Jaakkola, Max Tegmark

We introduce a general family, Generative Models from Physical Processes (GenPhys), where we translate partial differential equations (PDEs) describing physical processes to generative models.

The Quantization Model of Neural Scaling

1 code implementation NeurIPS 2023 Eric J. Michaud, Ziming Liu, Uzay Girit, Max Tegmark

We tentatively find that the frequency at which these quanta are used in the training distribution roughly follows a power law corresponding with the empirical scaling exponent for language models, a prediction of our theory.

Language Modeling Language Modelling +2

Precision Machine Learning

1 code implementation24 Oct 2022 Eric J. Michaud, Ziming Liu, Max Tegmark

We explore unique considerations involved in fitting ML models to data with very high precision, as is often required for science applications.

Omnigrok: Grokking Beyond Algorithmic Data

2 code implementations3 Oct 2022 Ziming Liu, Eric J. Michaud, Max Tegmark

Grokking, the unusual phenomenon for algorithmic datasets where generalization happens long after overfitting the training data, has remained elusive.

Attribute Representation Learning

Poisson Flow Generative Models

1 code implementation22 Sep 2022 Yilun Xu, Ziming Liu, Max Tegmark, Tommi Jaakkola

We interpret the data points as electrical charges on the $z=0$ hyperplane in a space augmented with an additional dimension $z$, generating a high-dimensional electric field (the gradient of the solution to Poisson equation).

Image Generation

AI Poincaré 2.0: Machine Learning Conservation Laws from Differential Equations

no code implementations23 Mar 2022 Ziming Liu, Varun Madhavan, Max Tegmark

We present a machine learning algorithm that discovers conservation laws from differential equations, both numerically (parametrized as neural networks) and symbolically, ensuring their functional independence (a non-linear generalization of linear independence).

BIG-bench Machine Learning

Fault-Tolerant Neural Networks from Biological Error Correction Codes

no code implementations25 Feb 2022 Alexander Zlokapa, Andrew K. Tan, John M. Martyn, Ila R. Fiete, Max Tegmark, Isaac L. Chuang

It has been an open question in deep learning if fault-tolerant computation is possible: can arbitrarily reliable computation be achieved using only unreliable neurons?

Open-Ended Question Answering

Machine-learning hidden symmetries

no code implementations20 Sep 2021 Ziming Liu, Max Tegmark

We present an automated method for finding hidden symmetries, defined as symmetries that become manifest only in a new coordinate system that must be discovered.

BIG-bench Machine Learning

AI Poincaré: Machine Learning Conservation Laws from Trajectories

no code implementations9 Nov 2020 Ziming Liu, Max Tegmark

We present AI Poincar\'e, a machine learning algorithm for auto-discovering conserved quantities using trajectory data from unknown dynamical systems.

BIG-bench Machine Learning

AI Feynman 2.0: Pareto-optimal symbolic regression exploiting graph modularity

2 code implementations NeurIPS 2020 Silviu-Marian Udrescu, Andrew Tan, Jiahai Feng, Orisvaldo Neto, Tailin Wu, Max Tegmark

We present an improved method for symbolic regression that seeks to fit data to formulas that are Pareto-optimal, in the sense of having the best accuracy for a given complexity.

regression Symbolic Regression +1

Symbolic Pregression: Discovering Physical Laws from Distorted Video

no code implementations19 May 2020 Silviu-Marian Udrescu, Max Tegmark

We present a method for unsupervised learning of equations of motion for objects in raw and optionally distorted unlabeled video.

regression Symbolic Regression

Foreground modelling via Gaussian process regression: an application to HERA data

no code implementations13 Apr 2020 Abhik Ghosh, Florent Mertens, Gianni Bernardi, Mário G. Santos, Nicholas S. Kern, Christopher L. Carilli, Trienko L. Grobler, Léon V. E. Koopmans, Daniel C. Jacobs, Adrian Liu, Aaron R. Parsons, Miguel F. Morales, James E. Aguirre, Joshua S. Dillon, Bryna J. Hazelton, Oleg M. Smirnov, Bharat K. Gehlot, Siyanda Matika, Paul Alexander, Zaki S. Ali, Adam P. Beardsley, Roshan K. Benefo, Tashalee S. Billings, Judd D. Bowman, Richard F. Bradley, Carina Cheng, Paul M. Chichura, David R. DeBoer, Eloy de Lera Acedo, Aaron Ewall-Wice, Gcobisa Fadana, Nicolas Fagnoni, Austin F. Fortino, Randall Fritz, Steve R. Furlanetto, Samavarti Gallardo, Brian Glendenning, Deepthi Gorthi, Bradley Greig, Jasper Grobbelaar, Jack Hickish, Alec Josaitis, Austin Julius, Amy S. Igarashi, MacCalvin Kariseb, Saul A. Kohn, Matthew Kolopanis, Telalo Lekalake, Anita Loots, David MacMahon, Lourence Malan, Cresshim Malgas, Matthys Maree, Zachary E. Martinot, Nathan Mathison, Eunice Matsetela, Andrei Mesinger, Abraham R. Neben, Bojan Nikolic, Chuneeta D. Nunhokee, Nipanjana Patra, Samantha Pieterse, Nima Razavi-Ghods, Jon Ringuette, James Robnett, Kathryn Rosie, Raddwine Sell, Craig Smith, Angelo Syce, Max Tegmark, Nithyanandan Thyagarajan, Peter K. G. Williams, Haoxuan Zheng

The key challenge in the observation of the redshifted 21-cm signal from cosmic reionization is its separation from the much brighter foreground emission.

Cosmology and Nongalactic Astrophysics

Pareto-optimal data compression for binary classification tasks

1 code implementation23 Aug 2019 Max Tegmark, Tailin Wu

The goal of lossy data compression is to reduce the storage cost of a data set $X$ while retaining as much information as possible about something ($Y$) that you care about.

Binary Classification Classification +5

Learnability for the Information Bottleneck

no code implementations ICLR Workshop LLD 2019 Tailin Wu, Ian Fischer, Isaac L. Chuang, Max Tegmark

However, in practice, not only is $\beta$ chosen empirically without theoretical guidance, there is also a lack of theoretical understanding between $\beta$, learnability, the intrinsic nature of the dataset and model capacity.

Representation Learning

AI Feynman: a Physics-Inspired Method for Symbolic Regression

2 code implementations27 May 2019 Silviu-Marian Udrescu, Max Tegmark

A core challenge for both physics and artificial intellicence (AI) is symbolic regression: finding a symbolic expression that matches data from an unknown function.

regression Symbolic Regression

Latent Representations of Dynamical Systems: When Two is Better Than One

no code implementations9 Feb 2019 Max Tegmark

A popular approach for predicting the future of dynamical systems involves mapping them into a lower-dimensional "latent space" where prediction is easier.

Vocal Bursts Valence Prediction

Toward an AI Physicist for Unsupervised Learning

1 code implementation24 Oct 2018 Tailin Wu, Max Tegmark

We investigate opportunities and challenges for improving unsupervised machine learning using four common strategies with a long history in physics: divide-and-conquer, Occam's razor, unification and lifelong learning.

Meta-learning autoencoders for few-shot prediction

1 code implementation26 Jul 2018 Tailin Wu, John Peurifoy, Isaac L. Chuang, Max Tegmark

Compared to humans, machine learning models generally require significantly more training examples and fail to extrapolate from experience to solve previously unseen challenges.

Meta-Learning Prediction

Nanophotonic Particle Simulation and Inverse Design Using Artificial Neural Networks

1 code implementation18 Oct 2017 John Peurifoy, Yichen Shen, Li Jing, Yi Yang, Fidel Cano-Renteria, Brendan Delacy, Max Tegmark, John D. Joannopoulos, Marin Soljacic

We propose a method to use artificial neural networks to approximate light scattering by multilayer nanoparticles.

Computational Physics Applied Physics Optics

Gated Orthogonal Recurrent Units: On Learning to Forget

1 code implementation8 Jun 2017 Li Jing, Caglar Gulcehre, John Peurifoy, Yichen Shen, Max Tegmark, Marin Soljačić, Yoshua Bengio

We present a novel recurrent neural network (RNN) based model that combines the remembering ability of unitary RNNs with the ability of gated RNNs to effectively forget redundant/irrelevant information in its memory.

Ranked #7 on Question Answering on bAbi (Accuracy (trained on 1k) metric)

Denoising Question Answering

The power of deeper networks for expressing natural functions

no code implementations ICLR 2018 David Rolnick, Max Tegmark

It is well-known that neural networks are universal approximators, but that deeper networks tend in practice to be more powerful than shallower ones.

Tunable Efficient Unitary Neural Networks (EUNN) and their application to RNNs

4 code implementations ICML 2017 Li Jing, Yichen Shen, Tena Dubček, John Peurifoy, Scott Skirlo, Yann Lecun, Max Tegmark, Marin Soljačić

Using unitary (instead of general) matrices in artificial neural networks (ANNs) is a promising way to solve the gradient explosion/vanishing problem, as well as to enable ANNs to learn long-term correlations in the data.

Permuted-MNIST

Why does deep and cheap learning work so well?

no code implementations29 Aug 2016 Henry W. Lin, Max Tegmark, David Rolnick

We show how the success of deep learning could depend not only on mathematics but also on physics: although well-known mathematical theorems guarantee that neural networks can approximate arbitrary functions well, the class of functions of practical interest can frequently be approximated through "cheap learning" with exponentially fewer parameters than generic ones.

Criticality in Formal Languages and Statistical Physics

no code implementations21 Jun 2016 Henry W. Lin, Max Tegmark

We show that the mutual information between two symbols, as a function of the number of symbols between the two, decays exponentially in any probabilistic regular grammar, but can decay like a power law for a context-free grammar.

An Improved Model of Diffuse Galactic Radio Emission from 10 MHz to 5 THz

3 code implementations16 May 2016 Haoxuan Zheng, Max Tegmark, Joshua S. Dillon, Doyeon A. Kim, Adrian Liu, Abraham Neben, Justin Jonas, Patricia Reich, Wolfgang Reich

We present an improved Global Sky Model (GSM) of diffuse galactic radio emission from 10 MHz to 5 THz, whose uses include foreground modeling for CMB and 21 cm cosmology.

Cosmology and Nongalactic Astrophysics Astrophysics of Galaxies Instrumentation and Methods for Astrophysics

Research Priorities for Robust and Beneficial Artificial Intelligence

no code implementations10 Feb 2016 Stuart Russell, Daniel Dewey, Max Tegmark

Success in the quest for artificial intelligence has the potential to bring unprecedented benefits to humanity, and it is therefore worthwhile to investigate how to maximize these benefits while avoiding potential pitfalls.

Friendly Artificial Intelligence: the Physics Challenge

no code implementations2 Sep 2014 Max Tegmark

Relentless progress in artificial intelligence (AI) is increasingly raising concerns that machines will replace humans on the job market, and perhaps altogether.

A model of diffuse Galactic Radio Emission from 10 MHz to 100 GHz

2 code implementations12 Feb 2008 Angelica de Oliveira-Costa, Max Tegmark, B. M. Gaensler, Justin Jonas, T. L. Landecker, Patricia Reich

Understanding diffuse Galactic radio emission is interesting both in its own right and for minimizing foreground contamination of cosmological measurements.

Optical Character Recognition Optical Character Recognition (OCR)

Separating the Early Universe from the Late Universe: cosmological parameter estimation beyond the black box

1 code implementation2 Jul 2002 Max Tegmark, Matias Zaldarriaga

We present a method for measuring the cosmic matter budget without assumptions about speculative Early Universe physics, and for measuring the primordial power spectrum P*(k) non-parametrically, either by combining CMB and LSS information or by using CMB polarization.

Clustering

The importance of quantum decoherence in brain processes

no code implementations5 Jul 1999 Max Tegmark

Based on a calculation of neural decoherence rates, we argue that that the degrees of freedom of the human brain that relate to cognitive processes should be thought of as a classical rather than quantum system, i. e., that there is nothing fundamentally wrong with the current classical approach to neural network simulations.

On the dimensionality of spacetime

no code implementations25 Feb 1997 Max Tegmark

Some superstring theories have more than one effective low-energy limit, corresponding to classical spacetimes with different dimensionalities.

General Relativity and Quantum Cosmology High Energy Physics - Phenomenology High Energy Physics - Theory

Cannot find the paper you are looking for? You can Submit a new open access paper.