Search Results for author: Andis Draguns

Found 6 papers, 3 papers with code

Limitations of Agents Simulated by Predictive Models

no code implementations8 Feb 2024 Raymond Douglas, Jacek Karwowski, Chan Bae, Andis Draguns, Victoria Krakovna

Prior work has shown theoretically that models fail to imitate agents that generated the training data if the agents relied on hidden observations: the hidden observations act as confounding variables, and the models treat actions they generate as evidence for nonexistent observations.

Mitigating the Problem of Strong Priors in LMs with Context Extrapolation

no code implementations31 Jan 2024 Raymond Douglas, Andis Draguns, Tomáš Gavenčiak

We develop a new technique for mitigating the problem of strong priors: we take the original set of instructions, produce a weakened version of the original prompt that is even more susceptible to the strong priors problem, and then extrapolate the continuation away from the weakened prompt.

Instruction Following

Unsupervised Training for Neural TSP Solver

no code implementations27 Jul 2022 Elīza Gaile, Andis Draguns, Emīls Ozoliņš, Kārlis Freivalds

We use our loss function with a Graph Neural Network and design controlled experiments on both Euclidean and asymmetric TSP.

reinforcement-learning Reinforcement Learning (RL)

Gates Are Not What You Need in RNNs

1 code implementation1 Aug 2021 Ronalds Zakovskis, Andis Draguns, Eliza Gaile, Emils Ozolins, Karlis Freivalds

In this paper, we propose a new recurrent cell called Residual Recurrent Unit (RRU) which beats traditional cells and does not employ a single gate.

Language Modelling Music Modeling +1

Goal-Aware Neural SAT Solver

1 code implementation14 Jun 2021 Emils Ozolins, Karlis Freivalds, Andis Draguns, Eliza Gaile, Ronalds Zakovskis, Sergejs Kozlovics

To demonstrate the capabilities of the query mechanism, we formulate an unsupervised (not depending on labels) loss function for Boolean Satisfiability Problem (SAT) and theoretically show that it allows the network to extract rich information about the problem.

Residual Shuffle-Exchange Networks for Fast Processing of Long Sequences

2 code implementations6 Apr 2020 Andis Draguns, Emīls Ozoliņš, Agris Šostaks, Matīss Apinis, Kārlis Freivalds

Attention is a commonly used mechanism in sequence processing, but it is of O(n^2) complexity which prevents its application to long sequences.

LAMBADA Language Modelling +1

Cannot find the paper you are looking for? You can Submit a new open access paper.