Search Results for author: Sebastian Stabinger

Found 15 papers, 7 papers with code

ANLS* -- A Universal Document Processing Metric for Generative Large Language Models

1 code implementation6 Feb 2024 David Peer, Philemon Schöpf, Volckmar Nebendahl, Alexander Rietzler, Sebastian Stabinger

However, evaluating GLLMs presents a challenge as the binary true or false evaluation used for discriminative models is not applicable to the predictions made by GLLMs.

Document Classification

Improving the Trainability of Deep Neural Networks through Layerwise Batch-Entropy Regularization

2 code implementations1 Aug 2022 David Peer, Bart Keulen, Sebastian Stabinger, Justus Piater, Antonio Rodríguez-Sánchez

We show empirically that we can therefore train a "vanilla" fully connected network and convolutional neural network -- no skip connections, batch normalization, dropout, or any other architectural tweak -- with 500 layers by simply adding the batch-entropy regularization term to the loss function.

Greedy-layer Pruning: Speeding up Transformer Models for Natural Language Processing

1 code implementation31 May 2021 David Peer, Sebastian Stabinger, Stefan Engl, Antonio Rodriguez-Sanchez

Knowledge distillation maintains high performance and reaches high compression rates, nevertheless, the size of the student model is fixed after pre-training and can not be changed individually for a given downstream task and use-case to reach a desired performance/speedup ratio.

Knowledge Distillation Unsupervised Pre-training

Auto-tuning of Deep Neural Networks by Conflicting Layer Removal

1 code implementation7 Mar 2021 David Peer, Sebastian Stabinger, Antonio Rodriguez-Sanchez

In the worst-case scenario, we prove that such a layer could lead to a network that cannot be trained at all.

Neural Architecture Search

Arguments for the Unsuitability of Convolutional Neural Networks for Non--Local Tasks

no code implementations23 Feb 2021 Sebastian Stabinger, David Peer, Antonio Rodríguez-Sánchez

Convolutional neural networks have established themselves over the past years as the state of the art method for image classification, and for many datasets, they even surpass humans in categorizing images.

Image Classification

Conflicting Bundles: Adapting Architectures Towards the Improved Training of Deep Neural Networks

1 code implementation5 Nov 2020 David Peer, Sebastian Stabinger, Antonio Rodriguez-Sanchez

In this paper, we introduce a novel theory and metric to identify layers that decrease the test accuracy of the trained models, this identification is done as early as at the beginning of training.

Evaluating the Progress of Deep Learning for Visual Relational Concepts

no code implementations29 Jan 2020 Sebastian Stabinger, Peer David, Justus Piater, Antonio Rodríguez-Sánchez

Convolutional Neural Networks (CNNs) have become the state of the art method for image classification in the last ten years.

Classification General Classification +2

Adapt or Get Left Behind: Domain Adaptation through BERT Language Model Finetuning for Aspect-Target Sentiment Classification

3 code implementations LREC 2020 Alexander Rietzler, Sebastian Stabinger, Paul Opitz, Stefan Engl

Aspect-Target Sentiment Classification (ATSC) is a subtask of Aspect-Based Sentiment Analysis (ABSA), which has many applications e. g. in e-commerce, where data and insights from reviews can be leveraged to create value for businesses and customers.

Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +5

Limitation of capsule networks

no code implementations21 May 2019 David Peer, Sebastian Stabinger, Antonio Rodriguez-Sanchez

A recently proposed method in deep learning groups multiple neurons to capsules such that each capsule represents an object or part of an object.

Evaluating CNNs on the Gestalt Principle of Closure

no code implementations30 Mar 2019 Gregor Ehrensperger, Sebastian Stabinger, Antonio Rodríguez Sánchez

Deep convolutional neural networks (CNNs) are widely known for their outstanding performance in classification and regression tasks over high-dimensional data.

valid

Increasing the adversarial robustness and explainability of capsule networks with $γ$-capsules

1 code implementation23 Dec 2018 David Peer, Sebastian Stabinger, Antonio Rodriguez-Sanchez

In this paper we introduce a new inductive bias for capsule networks and call networks that use this prior $\gamma$-capsule networks.

Adversarial Robustness Inductive Bias

Guided Labeling using Convolutional Neural Networks

no code implementations6 Dec 2017 Sebastian Stabinger, Antonio Rodriguez-Sanchez

Over the last couple of years, deep learning and especially convolutional neural networks have become one of the work horses of computer vision.

Evaluation of Deep Learning on an Abstract Image Classification Dataset

no code implementations25 Aug 2017 Sebastian Stabinger, Antonio Rodriguez-Sanchez

Convolutional Neural Networks have become state of the art methods for image classification over the last couple of years.

Classification General Classification +1

25 years of CNNs: Can we compare to human abstraction capabilities?

no code implementations28 Jul 2016 Sebastian Stabinger, Antonio Rodríguez-Sánchez, Justus Piater

We try to determine the progress made by convolutional neural networks over the past 25 years in classifying images into abstractc lasses.

Learning Abstract Classes using Deep Learning

no code implementations17 Jun 2016 Sebastian Stabinger, Antonio Rodriguez-Sanchez, Justus Piater

Humans are generally good at learning abstract concepts about objects and scenes (e. g.\ spatial orientation, relative sizes, etc.).

Cannot find the paper you are looking for? You can Submit a new open access paper.