Search Results for author: Christoph Ortner

Found 10 papers, 5 papers with code

Equivariant Matrix Function Neural Networks

no code implementations16 Oct 2023 Ilyes Batatia, Lars L. Schaaf, Huajie Chen, Gábor Csányi, Christoph Ortner, Felix A. Faber

Graph Neural Networks (GNNs), especially message-passing neural networks (MPNNs), have emerged as powerful architectures for learning on graphs in diverse applications.

Graph Regression

Retrieval of Boost Invariant Symbolic Observables via Feature Importance

no code implementations23 Jun 2023 Jose M Munoz, Ilyes Batatia, Christoph Ortner, Francesco Romeo

Deep learning approaches for jet tagging in high-energy physics are characterized as black boxes that process a large amount of information from which it is difficult to extract key distinctive observables.

Feature Importance Jet Tagging +1

Hyperactive Learning (HAL) for Data-Driven Interatomic Potentials

no code implementations9 Oct 2022 Cas van der Oord, Matthias Sachs, Dávid Péter Kovács, Christoph Ortner, Gábor Csányi

Data-driven interatomic potentials have emerged as a powerful class of surrogate models for {\it ab initio} potential energy surfaces that are able to reliably predict macroscopic properties with experimental accuracy.

Tensor-reduced atomic density representations

no code implementations2 Oct 2022 James P. Darby, Dávid P. Kovács, Ilyes Batatia, Miguel A. Caro, Gus L. W. Hart, Christoph Ortner, Gábor Csányi

Density based representations of atomic environments that are invariant under Euclidean symmetries have become a widely used tool in the machine learning of interatomic potentials, broader data-driven atomistic modelling and the visualisation and analysis of materials datasets. The standard mechanism used to incorporate chemical element information is to create separate densities for each element and form tensor products between them.

BIP: Boost Invariant Polynomials for Efficient Jet Tagging

1 code implementation17 Jul 2022 Jose M Munoz, Ilyes Batatia, Christoph Ortner

Deep Learning approaches are becoming the go-to methods for data analysis in High Energy Physics (HEP).

Computational Efficiency Jet Tagging

MACE: Higher Order Equivariant Message Passing Neural Networks for Fast and Accurate Force Fields

2 code implementations15 Jun 2022 Ilyes Batatia, Dávid Péter Kovács, Gregor N. C. Simm, Christoph Ortner, Gábor Csányi

In particular, we show that using four-body messages reduces the required number of message passing iterations to just two, resulting in a fast and highly parallelizable model, reaching or exceeding state-of-the-art accuracy on the rMD17, 3BPA, and AcAc benchmark tasks.

The Design Space of E(3)-Equivariant Atom-Centered Interatomic Potentials

2 code implementations13 May 2022 Ilyes Batatia, Simon Batzner, Dávid Péter Kovács, Albert Musaelian, Gregor N. C. Simm, Ralf Drautz, Christoph Ortner, Boris Kozinsky, Gábor Csányi

The rapid progress of machine learning interatomic potentials over the past couple of years produced a number of new architectures.

Physics-inspired structural representations for molecules and materials

no code implementations12 Jan 2021 Felix Musil, Andrea Grisafi, Albert P. Bartók, Christoph Ortner, Gábor Csányi, Michele Ceriotti

The first step in the construction of a regression model or a data-driven analysis, aiming to predict or elucidate the relationship between the atomic scale structure of matter and its properties, involves transforming the Cartesian coordinates of the atoms into a suitable representation.

Chemical Physics

Sensitivity and Dimensionality of Atomic Environment Representations used for Machine Learning Interatomic Potentials

7 code implementations2 Jun 2020 Berk Onat, Christoph Ortner, James R. Kermode

Faithfully representing chemical environments is essential for describing materials and molecules with machine learning approaches.

Materials Science Chemical Physics

Cannot find the paper you are looking for? You can Submit a new open access paper.