1 code implementation • CLASP 2022 • Lovisa Hagström, Tobias Norlund, Richard Johansson
This is a setting in which we fuse language with information from the math modality and strive to replicate some fusion methods from the vision-and-language domain.
1 code implementation • 2 Nov 2023 • Lovisa Hagström, Denitsa Saynova, Tobias Norlund, Moa Johansson, Richard Johansson
In this work, we identify potential causes of inconsistency and evaluate the effectiveness of two mitigation strategies: up-scaling and augmenting the LM with a retrieval corpus.
1 code implementation • 25 May 2023 • Ehsan Doostmohammadi, Tobias Norlund, Marco Kuhlmann, Richard Johansson
Inspired by this, we replace the semantic retrieval in Retro with a surface-level method based on BM25, obtaining a significant reduction in perplexity.
no code implementations • 23 Feb 2023 • Tobias Norlund, Ehsan Doostmohammadi, Richard Johansson, Marco Kuhlmann
Recent work on the Retrieval-Enhanced Transformer (RETRO) model has shown that off-loading memory from trainable weights to a retrieval database can significantly improve language modeling and match the performance of non-retrieval models that are an order of magnitude larger in size.
no code implementations • EMNLP (BlackboxNLP) 2021 • Tobias Norlund, Lovisa Hagström, Richard Johansson
We find that our method can successfully be used to measure visual knowledge transfer capabilities in models and that our novel model architecture shows promising results for leveraging multimodal knowledge in a unimodal setting.
1 code implementation • NoDaLiDa 2021 • Tobias Norlund, Agnes Stenbom
We present on-going work of evaluating the, to our knowledge, first large generative language model trained to converse in Swedish, using data from the online discussion forum Flashback.