Search Results for author: Tobias Norlund

Found 7 papers, 4 papers with code

Can We Use Small Models to Investigate Multimodal Fusion Methods?

1 code implementation CLASP 2022 Lovisa Hagström, Tobias Norlund, Richard Johansson

This is a setting in which we fuse language with information from the math modality and strive to replicate some fusion methods from the vision-and-language domain.

Math

The Effect of Scaling, Retrieval Augmentation and Form on the Factual Consistency of Language Models

1 code implementation2 Nov 2023 Lovisa Hagström, Denitsa Saynova, Tobias Norlund, Moa Johansson, Richard Johansson

In this work, we identify potential causes of inconsistency and evaluate the effectiveness of two mitigation strategies: up-scaling and augmenting the LM with a retrieval corpus.

Retrieval

Surface-Based Retrieval Reduces Perplexity of Retrieval-Augmented Language Models

1 code implementation25 May 2023 Ehsan Doostmohammadi, Tobias Norlund, Marco Kuhlmann, Richard Johansson

Inspired by this, we replace the semantic retrieval in Retro with a surface-level method based on BM25, obtaining a significant reduction in perplexity.

Re-Ranking Retrieval +1

On the Generalization Ability of Retrieval-Enhanced Transformers

no code implementations23 Feb 2023 Tobias Norlund, Ehsan Doostmohammadi, Richard Johansson, Marco Kuhlmann

Recent work on the Retrieval-Enhanced Transformer (RETRO) model has shown that off-loading memory from trainable weights to a retrieval database can significantly improve language modeling and match the performance of non-retrieval models that are an order of magnitude larger in size.

Language Modelling Retrieval

Transferring Knowledge from Vision to Language: How to Achieve it and how to Measure it?

no code implementations EMNLP (BlackboxNLP) 2021 Tobias Norlund, Lovisa Hagström, Richard Johansson

We find that our method can successfully be used to measure visual knowledge transfer capabilities in models and that our novel model architecture shows promising results for leveraging multimodal knowledge in a unimodal setting.

Hallucination Transfer Learning

Building a Swedish Open-Domain Conversational Language Model

1 code implementation NoDaLiDa 2021 Tobias Norlund, Agnes Stenbom

We present on-going work of evaluating the, to our knowledge, first large generative language model trained to converse in Swedish, using data from the online discussion forum Flashback.

Language Modelling

Cannot find the paper you are looking for? You can Submit a new open access paper.