no code implementations • 5 Dec 2024 • Bernd Prach, Christoph H. Lampert
We come to the conclusion that on this datasets, the limitation of current robust models also lies in the generalization, and that they require a lot of data to do well on the test set.
no code implementations • 17 Jul 2024 • Nikita P. Kalinin, Simone Bombari, Hossein Zakerinia, Christoph H. Lampert
We study the Kolmogorov-Arnold Network (KAN), recently proposed as an alternative to the classical Multilayer Perceptron (MLP), in the application for differentially private model training.
1 code implementation • 11 Mar 2024 • Egor Zverev, Sahar Abdelnabi, Soroush Tabesh, Mario Fritz, Christoph H. Lampert
We also present a new dataset, SEP, that allows estimating the measure for real-world models.
1 code implementation • 6 Feb 2024 • Hossein Zakerinia, Amin Behjati, Christoph H. Lampert
We introduce a new framework for studying meta-learning methods using PAC-Bayesian theory.
1 code implementation • CVPR 2024 • Bernd Prach, Fabio Brau, Giorgio Buttazzo, Christoph H. Lampert
The robustness of neural networks against input perturbations with bounded magnitude represents a serious concern in the deployment of deep learning models in safety-critical systems.
no code implementations • 11 Dec 2023 • Paniz Halvachi, Alexandra Peste, Dan Alistarh, Christoph H. Lampert
We present ELSA, a practical solution for creating deep networks that can easily be deployed at different levels of sparsity.
1 code implementation • 28 Nov 2023 • Bernd Prach, Fabio Brau, Giorgio Buttazzo, Christoph H. Lampert
The robustness of neural networks against input perturbations with bounded magnitude represents a serious concern in the deployment of deep learning models in safety-critical systems.
no code implementations • 20 Nov 2023 • Eli Verwimp, Rahaf Aljundi, Shai Ben-David, Matthias Bethge, Andrea Cossu, Alexander Gepperth, Tyler L. Hayes, Eyke Hüllermeier, Christopher Kanan, Dhireesha Kudithipudi, Christoph H. Lampert, Martin Mundt, Razvan Pascanu, Adrian Popescu, Andreas S. Tolias, Joost Van de Weijer, Bing Liu, Vincenzo Lomonaco, Tinne Tuytelaars, Gido M. van de Ven
Continual learning is a subfield of machine learning, which aims to allow machine learning models to continuously learn on new data, by accumulating knowledge without forgetting what was learned in the past.
1 code implementation • 10 Nov 2023 • Bernd Prach, Christoph H. Lampert
A crucial property for achieving secure, trustworthy and interpretable deep learning systems is their robustness: small changes to a system's inputs should not result in large changes to its outputs.
2 code implementations • 8 Jun 2023 • Jonathan Scott, Hossein Zakerinia, Christoph H. Lampert
We present PeFLL, a new personalized federated learning algorithm that improves over the state-of-the-art in three aspects: 1) it produces more accurate models, especially in the low-data regime, and not only for clients present during its training phase, but also for any that may emerge in the future; 2) it reduces the amount of on-client computation and client-server communication by providing future clients with ready-to-use personalized models that require no additional finetuning or optimization; 3) it comes with theoretical guarantees that establish generalization from the observed clients to future ones.
1 code implementation • 14 Mar 2023 • Kateryna Lutsai, Christoph H. Lampert
This research is aimed to solve the tweet/user geolocation prediction task and provide a flexible methodology for the geotagging of textual big data.
1 code implementation • 12 Oct 2022 • Jonathan Scott, Michelle Yeo, Christoph H. Lampert
We present Cross-Client Label Propagation(XCLP), a new method for transductive federated learning.
no code implementations • 29 Aug 2022 • Peter Súkeník, Christoph H. Lampert
Modern machine learning tasks often require considering not just one but multiple objectives.
1 code implementation • 5 Aug 2022 • Bernd Prach, Christoph H. Lampert
In this work, we propose a new technique for constructing such Lipschitz networks that has a number of desirable properties: it can be applied to any linear network layer (fully-connected or convolutional), it provides formal guarantees on the Lipschitz constant, it is easy to implement and efficient to run, and it can be combined with any training objective and optimization method.
1 code implementation • 28 Jul 2022 • Alexandra Peste, Adrian Vladu, Eldar Kurtic, Christoph H. Lampert, Dan Alistarh
In this work we propose a new compression-aware minimizer dubbed CrAM that modifies the optimization step in a principled way, in order to produce models whose local loss behavior is stable under compression operations such as pruning.
1 code implementation • 10 Jun 2022 • Paulina Tomaszewska, Christoph H. Lampert
We introduce LIMES, a new method for learning with non-stationary streaming data, inspired by the recent success of meta-learning.
no code implementations • 8 Jul 2021 • Alexandra Peste, Dan Alistarh, Christoph H. Lampert
The availability of large amounts of user-provided data has been key to the success of machine learning for many real-world tasks.
1 code implementation • 22 Jun 2021 • Eugenia Iofinova, Nikola Konstantinov, Christoph H. Lampert
In this work we address the problem of fair learning from unreliable training data in the robust multisource setting, where the available training data comes from multiple sources, a fraction of which might not be representative of the true data distribution.
no code implementations • 16 Jun 2021 • Paul Henderson, Christoph H. Lampert, Bernd Bickel
Our goal in this work is to generate realistic videos given just one initial frame as input.
no code implementations • 27 May 2021 • Mary Phuong, Christoph H. Lampert
Knowledge distillation, i. e., one classifier being trained on the outputs of another classifier, is an empirically very successful technique for knowledge transfer between classifiers.
no code implementations • 11 Feb 2021 • Nikola Konstantinov, Christoph H. Lampert
Addressing fairness concerns about machine learning models is a crucial step towards their long-term adoption in real-world automated systems.
no code implementations • 11 Feb 2021 • Nikola Konstantinov, Christoph H. Lampert
Given the abundance of applications of ranking in recent years, addressing fairness concerns around automated ranking systems becomes necessary for increasing the trust among end-users.
no code implementations • 27 Aug 2020 • Amelie Royer, Christoph H. Lampert
Fine-tuning is a popular way of exploiting knowledge contained in a pre-trained convolutional network for a new visual recognition task.
1 code implementation • NeurIPS 2020 • Paul Henderson, Christoph H. Lampert
A natural approach to generative modeling of videos is to represent them as a composition of moving objects.
no code implementations • ICLR 2020 • Mary Phuong, Christoph H. Lampert
We address the following question: How redundant is the parameterisation of ReLU networks?
1 code implementation • 27 Apr 2020 • Amelie Royer, Christoph H. Lampert
State-of-the-art detection systems are generally evaluated on their ability to exhaustively retrieve objects densely distributed in the image, across a wide variety of appearances and semantic categories.
1 code implementation • CVPR 2020 • Paul Henderson, Vagia Tsiminaki, Christoph H. Lampert
Thus, it learns to generate meshes that when rendered, produce images similar to those in its training set.
no code implementations • 1 Apr 2020 • Titas Anciukevicius, Christoph H. Lampert, Paul Henderson
We present a generative model of images that explicitly reasons over the set of objects they show.
no code implementations • ICML 2020 • Nikola Konstantinov, Elias Frantar, Dan Alistarh, Christoph H. Lampert
We study the problem of learning from multiple untrusted data sources, a scenario of increasing practical relevance given the recent emergence of crowdsourcing and collaborative learning paradigms.
no code implementations • 7 Dec 2018 • Ehsan Pajouheshgar, Christoph H. Lampert
We introduce a spatio-temporal convolutional neural network model for trajectory forecasting from visual sources.
no code implementations • 5 Jul 2018 • Alexander Kolesnikov, Alina Kuznetsova, Christoph H. Lampert, Vittorio Ferrari
We propose a new model for detecting visual relationships, such as "person riding motorcycle" or "bottle on table".
1 code implementation • ICML 2018 • Subham S. Sahoo, Christoph H. Lampert, Georg Martius
We present an approach to identify concise equations from data using a shallow neural network approach.
1 code implementation • 11 Apr 2018 • Rémy Sun, Christoph H. Lampert
Computer vision systems for automatic image categorization have become accurate and reliable enough that they can run continuously for days or even years as components of real-world commercial applications.
10 code implementations • 3 Jul 2017 • Yongqin Xian, Christoph H. Lampert, Bernt Schiele, Zeynep Akata
Due to the importance of zero-shot learning, i. e. classifying images where there is a lack of labeled training data, the number of proposed approaches has recently increased steadily.
1 code implementation • 11 May 2017 • Amelie Royer, Alexander Kolesnikov, Christoph H. Lampert
We develop a probabilistic technique for colorizing grayscale natural images.
no code implementations • ICML 2018 • Ilja Kuzborskij, Christoph H. Lampert
We establish a data-dependent notion of algorithmic stability for Stochastic Gradient Descent (SGD), and employ it to develop novel generalization bounds.
no code implementations • ICML 2017 • Alexander Kolesnikov, Christoph H. Lampert
We study probabilistic models of natural images and extend the autoregressive family of PixelCNN architectures by incorporating auxiliary variables.
Ranked #17 on Image Generation on ImageNet 64x64 (Bits per dim metric)
10 code implementations • CVPR 2017 • Sylvestre-Alvise Rebuffi, Alexander Kolesnikov, Georg Sperl, Christoph H. Lampert
A major open problem on the road to artificial intelligence is the development of incrementally learning systems that learn about more and more concepts over time from a stream of data.
Ranked #2 on Incremental Learning on ImageNet100 - 10 steps (# M Params metric)
1 code implementation • 10 Oct 2016 • Georg Martius, Christoph H. Lampert
In classical machine learning, regression is treated as a black box process of identifying a suitable function from a hypothesis set without attempting to gain insight into the mechanism connecting inputs and outputs.
no code implementations • 18 May 2016 • Alexander Kolesnikov, Christoph H. Lampert
Weakly-supervised object localization methods tend to fail for object classes that consistently co-occur with the same background elements, e. g. trains on tracks.
2 code implementations • 19 Mar 2016 • Alexander Kolesnikov, Christoph H. Lampert
We introduce a new loss function for the weakly-supervised training of semantic image segmentation models based on three guiding principles: to seed with weak localization cues, to expand objects based on the information about which classes can occur in an image, and to constrain the segmentations to coincide with object boundaries.
no code implementations • ICML 2017 • Anastasia Pentina, Christoph H. Lampert
In contrast to previous work, which required that annotated training data is available for all tasks, we consider a new setting, in which for some tasks, potentially most of them, only unlabeled training data is provided.
no code implementations • NeurIPS 2015 • Anastasia Pentina, Christoph H. Lampert
In the first case we prove a PAC-Bayesian theorem, which can be seen as a direct generalization of the analogous previous result for the i. i. d.
no code implementations • 9 Oct 2015 • Alexander Zimin, Christoph H. Lampert
In particular, we aim at learning predictors that minimize the conditional risk for a stochastic process, i. e. the expected loss of the predictor on the next point conditioned on the set of training samples observed so far.
no code implementations • CVPR 2015 • Amelie Royer, Christoph H. Lampert
Experiments on the ILSVRC2010 and ILSVRC2012 datasets show that adapting object classification systems at prediction time can significantly reduce their error rate, even with no additional human feedback.
no code implementations • 28 Apr 2015 • Alexander Kolesnikov, Christoph H. Lampert
In this work, we present a Gaussian process (GP) based technique for simultaneously identifying which images of a training set have unreliable annotation and learning a segmentation model in which the negative effect of these images is suppressed.
no code implementations • CVPR 2015 • Anastasia Pentina, Viktoriia Sharmanska, Christoph H. Lampert
Sharing information between multiple tasks enables algorithms to achieve good generalization performance even from small amounts of training data.
no code implementations • 1 Oct 2014 • Viktoriia Sharmanska, Novi Quadrianto, Christoph H. Lampert
We interpret these methods as learning easiness and hardness of the objects in the privileged space and then transferring this knowledge to train a better classifier in the original space.
no code implementations • CVPR 2015 • Neel Shah, Vladimir Kolmogorov, Christoph H. Lampert
Structural support vector machines (SSVMs) are amongst the best performing models for structured computer vision tasks, such as semantic image segmentation or human pose estimation.
no code implementations • NeurIPS 2014 • Daniel Hernández-Lobato, Viktoriia Sharmanska, Kristian Kersting, Christoph H. Lampert, Novi Quadrianto
That is, in contrast to the standard GPC setting, the latent function is not just a nuisance but a feature: it becomes a natural measure of confidence about the training data by modulating the slope of the GPC sigmoid likelihood function.
no code implementations • CVPR 2015 • Christoph H. Lampert
Our main contribution is a method for predicting the next step of the time-varying distribution from a given sequence of sample sets from earlier time steps.
no code implementations • CVPR 2014 • Vladyslav Sydorov, Mayu Sakurada, Christoph H. Lampert
Fisher Kernels and Deep Learning were two developments with significant impact on large-scale object categorization in the last years.
no code implementations • 27 Mar 2014 • Alexander Kolesnikov, Matthieu Guillaumin, Vittorio Ferrari, Christoph H. Lampert
It is inspired by existing closed-form expressions for the maximum likelihood parameters of a generative graphical model with tree topology.
no code implementations • 12 Nov 2013 • Anastasia Pentina, Christoph H. Lampert
Transfer learning has received a lot of attention in the machine learning community over the last years, and several effective algorithms have been developed.
no code implementations • IEEE Transactions on Pattern Analysis and Machine Intelligence 2013 • Christoph H. Lampert, Hannes Nickisch, Stefan Harmeling
To tackle the problem, we introduce attribute-based classification: Objects are identified based on a high-level description that is phrased in terms of semantic attributes, such as the object’s color or shape.
no code implementations • NeurIPS 2012 • Christoph H. Lampert
MMP is typically performed as a two-stage procedure: one estimates each variable's marginal probability and then forms a prediction from the states of maximal probability.
no code implementations • NeurIPS 2011 • Christoph H. Lampert
We study multi-label prediction for structured output spaces, a problem that occurs, for example, in object detection in images, secondary structure prediction in computational biology, and graph matching with symmetries.