no code implementations • 12 Sep 2016 • Muhammet Bastan, S. Saqib Bukhari, Thomas M. Breuel
This way, the strong edges are used to recover weak or missing edges by considering the local edge structures, instead of blindly linking them if gradient magnitudes are above some threshold.
no code implementations • 13 Nov 2015 • Federico Raue, Andreas Dengel, Thomas M. Breuel, Marcus Liwicki
We evaluated the proposed extension in the following scenarios: missing elements in one modality (visual or audio) and missing elements in both modalities (visual and sound).
no code implementations • 12 Aug 2015 • Thomas M. Breuel
The paper introduces a biologically and evolutionarily plausible neural architecture that allows a single group of neurons, or an entire cortical pathway, to be dynamically reconfigured to perform multiple, potentially very different computations.
no code implementations • 12 Aug 2015 • Thomas M. Breuel
Neural networks are usually trained by some form of stochastic gradient descent (SGD)).
1 code implementation • 12 Aug 2015 • Thomas M. Breuel
The performance of neural network classifiers is determined by a number of hyperparameters, including learning rate, batch size, and depth.
1 code implementation • 11 Aug 2015 • Thomas M. Breuel
LSTM (Long Short-Term Memory) recurrent neural networks have been highly successful in a number of application areas.
no code implementations • CVPR 2015 • Wonmin Byeon, Thomas M. Breuel, Federico Raue, Marcus Liwicki
This paper addresses the problem of pixel-level segmentation and classification of scene images with an entirely learning-based approach using Long Short Term Memory (LSTM) recurrent neural networks, which are commonly used for sequence classification.