1 code implementation • 18 Jul 2024 • Guoli Yin, Haoping Bai, Shuang Ma, Feng Nan, Yanchao Sun, Zhaoyang Xu, Shen Ma, Jiarui Lu, Xiang Kong, Aonan Zhang, Dian Ang Yap, Yizhe Zhang, Karsten Ahnert, Vik Kamath, Mathias Berglund, Dominic Walsh, Tobias Gindele, Juergen Wiest, Zhengfeng Lai, Xiaoming Wang, Jiulong Shan, Meng Cao, Ruoming Pang, ZiRui Wang
Ultimately, MMAU not only sheds light on the capabilities and limitations of LLM agents but also enhances the interpretability of their performance.
no code implementations • NeurIPS 2019 • Rinu Boney, Norman Di Palo, Mathias Berglund, Alexander Ilin, Juho Kannala, Antti Rasmus, Harri Valpola
Trajectory optimization using a learned model of the environment is one of the core elements of model-based reinforcement learning.
2 code implementations • NeurIPS 2016 • Klaus Greff, Antti Rasmus, Mathias Berglund, Tele Hotloo Hao, Jürgen Schmidhuber, Harri Valpola
We present a framework for efficient perceptual inference that explicitly reasons about the segmentation of its inputs and features.
no code implementations • NeurIPS 2015 • Mathias Berglund, Tapani Raiko, Mikko Honkala, Leo Kärkkäinen, Akos Vetek, Juha T. Karhunen
Although unidirectional RNNs have recently been trained successfully to model such time series, inference in the negative time direction is non-trivial.
1 code implementation • 20 Nov 2015 • Jelena Luketina, Mathias Berglund, Klaus Greff, Tapani Raiko
Hyperparameter selection generally relies on running multiple full training trials, with selection based on validation set performance.
10 code implementations • NeurIPS 2015 • Antti Rasmus, Harri Valpola, Mikko Honkala, Mathias Berglund, Tapani Raiko
We combine supervised learning with unsupervised learning in deep neural networks.
no code implementations • NeurIPS 2015 • Mathias Berglund, Tapani Raiko, Mikko Honkala, Leo Kärkkäinen, Akos Vetek, Juha Karhunen
Although unidirectional RNNs have recently been trained successfully to model such time series, inference in the negative time direction is non-trivial.
no code implementations • 11 Jun 2014 • Tapani Raiko, Mathias Berglund, Guillaume Alain, Laurent Dinh
Our experiments confirm that training stochastic networks is difficult and show that the proposed two estimators perform favorably among all the five known estimators.
no code implementations • 20 Dec 2013 • Mathias Berglund, Tapani Raiko
Contrastive Divergence (CD) and Persistent Contrastive Divergence (PCD) are popular methods for training the weights of Restricted Boltzmann Machines.