Search Results for author: Tapani Raiko

Found 15 papers, 9 papers with code

A Character-Word Compositional Neural Language Model for Finnish

1 code implementation10 Dec 2016 Matti Lankinen, Hannes Heikinheimo, Pyry Takala, Tapani Raiko, Juha Karhunen

Inspired by recent research, we explore ways to model the highly morphological Finnish language at the level of characters while maintaining the performance of word-level models.

Language Modelling

Semi-Supervised Domain Adaptation for Weakly Labeled Semantic Video Object Segmentation

no code implementations7 Jun 2016 Huiling Wang, Tapani Raiko, Lasse Lensu, Tinghuai Wang, Juha Karhunen

We propose a semi-supervised approach to adapting CNN image recognition model trained from labeled image data to the target domain exploiting both semantic evidence learned from CNN, and the intrinsic structures of video data.

Domain Adaptation Segmentation +5

Bidirectional Recurrent Neural Networks as Generative Models

no code implementations NeurIPS 2015 Mathias Berglund, Tapani Raiko, Mikko Honkala, Leo Kärkkäinen, Akos Vetek, Juha T. Karhunen

Although unidirectional RNNs have recently been trained successfully to model such time series, inference in the negative time direction is non-trivial.

Bayesian Inference Time Series +1

Scalable Gradient-Based Tuning of Continuous Regularization Hyperparameters

1 code implementation20 Nov 2015 Jelena Luketina, Mathias Berglund, Klaus Greff, Tapani Raiko

Hyperparameter selection generally relies on running multiple full training trials, with selection based on validation set performance.

Hyperparameter Optimization

DopeLearning: A Computational Approach to Rap Lyrics Generation

1 code implementation18 May 2015 Eric Malmi, Pyry Takala, Hannu Toivonen, Tapani Raiko, Aristides Gionis

First, we develop a prediction model to identify the next line of existing lyrics from a set of candidate next lines.

Lateral Connections in Denoising Autoencoders Support Supervised Learning

1 code implementation30 Apr 2015 Antti Rasmus, Harri Valpola, Tapani Raiko

We show how a deep denoising autoencoder with lateral connections can be used as an auxiliary unsupervised learning task to support supervised learning.

Denoising General Classification

Denoising autoencoder with modulated lateral connections learns invariant representations of natural images

1 code implementation22 Dec 2014 Antti Rasmus, Tapani Raiko, Harri Valpola

Suitable lateral connections between encoder and decoder are shown to allow higher layers of a denoising autoencoder (dAE) to focus on invariant representations.

Denoising

Iterative Neural Autoregressive Distribution Estimator NADE-k

1 code implementation NeurIPS 2014 Tapani Raiko, Yao Li, Kyunghyun Cho, Yoshua Bengio

Training of the neural autoregressive density estimator (NADE) can be viewed as doing one step of probabilistic inference on missing values in data.

Density Estimation Image Generation +1

Linear State-Space Model with Time-Varying Dynamics

no code implementations2 Oct 2014 Jaakko Luttinen, Tapani Raiko, Alexander Ilin

The time dependency is obtained by forming the state dynamics matrix as a time-varying linear combination of a set of matrices.

Techniques for Learning Binary Stochastic Feedforward Neural Networks

no code implementations11 Jun 2014 Tapani Raiko, Mathias Berglund, Guillaume Alain, Laurent Dinh

Our experiments confirm that training stochastic networks is difficult and show that the proposed two estimators perform favorably among all the five known estimators.

Structured Prediction

Iterative Neural Autoregressive Distribution Estimator (NADE-k)

1 code implementation5 Jun 2014 Tapani Raiko, Li Yao, Kyunghyun Cho, Yoshua Bengio

Training of the neural autoregressive density estimator (NADE) can be viewed as doing one step of probabilistic inference on missing values in data.

Density Estimation Image Generation +1

Stochastic Gradient Estimate Variance in Contrastive Divergence and Persistent Contrastive Divergence

no code implementations20 Dec 2013 Mathias Berglund, Tapani Raiko

Contrastive Divergence (CD) and Persistent Contrastive Divergence (PCD) are popular methods for training the weights of Restricted Boltzmann Machines.

Cannot find the paper you are looking for? You can Submit a new open access paper.