1 code implementation • EMNLP (BlackboxNLP) 2020 • David Yenicelik, Florian Schmidt, Yannic Kilcher
The recent paradigm shift to contextual word embeddings has seen tremendous success across a wide range of down-stream tasks.
1 code implementation • 14 Apr 2023 • Andreas Köpf, Yannic Kilcher, Dimitri von Rütte, Sotiris Anagnostidis, Zhi-Rui Tam, Keith Stevens, Abdullah Barhoum, Nguyen Minh Duc, Oliver Stanley, Richárd Nagyfi, Shahul ES, Sameer Suri, David Glushkov, Arnav Dantuluri, Andrew Maguire, Christoph Schuhmann, Huu Nguyen, Alexander Mattick
In an effort to democratize research on large-scale alignment, we release OpenAssistant Conversations, a human-generated, human-annotated assistant-style conversation corpus consisting of 161, 443 messages in 35 different languages, annotated with 461, 292 quality ratings, resulting in over 10, 000 complete and fully annotated conversation trees.
2 code implementations • 26 Jan 2022 • Dimitri von Rütte, Luca Biggio, Yannic Kilcher, Thomas Hofmann
Generating music with deep neural networks has been an area of active research in recent years.
no code implementations • 1 Sep 2021 • Leonard Adolphs, Benjamin Boerschinger, Christian Buck, Michelle Chen Huebscher, Massimiliano Ciaramita, Lasse Espeholt, Thomas Hofmann, Yannic Kilcher, Sascha Rothe, Pier Giuseppe Sessa, Lierni Sestorain Saralegui
This paper presents first successful steps in designing search agents that learn meta-strategies for iterative query refinement in information-seeking tasks.
no code implementations • 23 Mar 2021 • Paulina Grnarova, Yannic Kilcher, Kfir Y. Levy, Aurelien Lucchi, Thomas Hofmann
Among known problems experienced by practitioners is the lack of convergence guarantees or convergence to a non-optimum cycle.
1 code implementation • 5 Feb 2021 • Surya Kant Sahu, Abhinav Java, Arshad Shaikh, Yannic Kilcher
To that end, we first define a metric, MLH (Model Enthalpy), that measures the closeness of a set of numbers to Benford's Law and we show empirically that it is a strong predictor of Validation Accuracy.
no code implementations • 11 Nov 2019 • Benjamin Borschinger, Jordan Boyd-Graber, Christian Buck, Jannis Bulian, Massimiliano Ciaramita, Michelle Chen Huebscher, Wojciech Gajewski, Yannic Kilcher, Rodrigo Nogueira, Lierni Sestorain Saralegu
We investigate a framework for machine reading, inspired by real world information-seeking problems, where a meta question answering system interacts with a black box environment.
no code implementations • 25 Sep 2019 • Kevin Roth, Yannic Kilcher, Thomas Hofmann
We establish a theoretical link between adversarial training and operator norm regularization for deep neural networks.
no code implementations • NeurIPS 2020 • Kevin Roth, Yannic Kilcher, Thomas Hofmann
We establish a theoretical link between adversarial training and operator norm regularization for deep neural networks.
no code implementations • ICLR 2019 • Yannic Kilcher, Gary Bécigneul, Thomas Hofmann
We develop our method for fully-connected as well as convolutional layers.
1 code implementation • 13 Feb 2019 • Kevin Roth, Yannic Kilcher, Thomas Hofmann
We investigate conditions under which test statistics exist that can reliably detect examples, which have been adversarially manipulated in a white-box attack.
no code implementations • 15 Nov 2017 • Yannic Kilcher, Thomas Hofmann
Black-Box attacks on machine learning models occur when an attacker, despite having no access to the inner workings of a model, can successfully craft an attack by means of model theft.
no code implementations • ICLR 2018 • Yannic Kilcher, Aurelien Lucchi, Thomas Hofmann
In implicit models, one often interpolates between sampled points in latent space.
no code implementations • ICLR 2018 • Yannic Kilcher, Gary Becigneul, Thomas Hofmann
It is commonly agreed that the use of relevant invariances as a good statistical bias is important in machine-learning.
no code implementations • ICLR 2018 • Yannic Kilcher, Aurelien Lucchi, Thomas Hofmann
We consider the problem of training generative models with deep neural networks as generators, i. e. to map latent codes to data points.
no code implementations • 28 Jul 2017 • Yannic Kilcher, Aurélien Lucchi, Thomas Hofmann
We consider the problem of training generative models with deep neural networks as generators, i. e. to map latent codes to data points.
no code implementations • NeurIPS 2016 • Gabriel Krummenacher, Brian McWilliams, Yannic Kilcher, Joachim M. Buhmann, Nicolai Meinshausen
We show that the regret of Ada-LR is close to the regret of full-matrix AdaGrad which can have an up-to exponentially smaller dependence on the dimension than the diagonal variant.