no code implementations • EMNLP (Louhi) 2020 • Hanna Berg, Aron Henriksson, Hercules Dalianis
The impact of de-identification on data quality and, in particular, utility for developing models for downstream tasks has been more thoroughly studied for structured data than for unstructured text.
no code implementations • RANLP 2021 • Anastasios Lamproudis, Aron Henriksson, Hercules Dalianis
The use of pretrained language models, fine-tuned to perform a specific downstream task, has become widespread in NLP.
no code implementations • LREC 2022 • Anastasios Lamproudis, Aron Henriksson, Hercules Dalianis
Here, an empirical investigation is carried out in which various strategies for adapting a generic language model to the clinical domain are compared to pretraining a pure clinical language model.
no code implementations • LREC 2022 • Thomas Vakili, Anastasios Lamproudis, Aron Henriksson, Hercules Dalianis
The impact of the de-identification techniques is assessed by training and evaluating the models using six clinical downstream tasks.
no code implementations • 25 Mar 2025 • Korbinian Randl, John Pavlopoulos, Aron Henriksson, Tony Lindgren, Juli Bakagianni
In this challenge, we explored text-based food hazard prediction with long tail distributed classes.
no code implementations • 20 Feb 2025 • Thomas Vakili, Aron Henriksson, Hercules Dalianis
In this study, we domain-adapt LLMs to the clinical domain and generate synthetic clinical texts that are machine-annotated with tags for personally identifiable information using capable encoder-based NER models.
1 code implementation • 19 Jul 2024 • Korbinian Randl, John Pavlopoulos, Aron Henriksson, Tony Lindgren
This paper investigates the reliability of explanations generated by large language models (LLMs) when prompted to explain their previous output.
1 code implementation • 18 Mar 2024 • Korbinian Randl, John Pavlopoulos, Aron Henriksson, Tony Lindgren
Contaminated or adulterated food poses a substantial risk to human health.