1 code implementation • nlppower (ACL) 2022 • Kathrin Blagec, Georg Dorffner, Milad Moradi, Simon Ott, Matthias Samwald
Our results suggest that the large majority of natural language processing metrics currently used have properties that may result in an inadequate reflection of a models' performance.
no code implementations • 9 Mar 2022 • Simon Ott, Adriano Barbosa-Silva, Kathrin Blagec, Jan Brauner, Matthias Samwald
Benchmarks are crucial to measuring and steering progress in artificial intelligence (AI).
no code implementations • 18 Jan 2022 • Kathrin Blagec, Jakob Kraiger, Wolfgang Frühwirt, Matthias Samwald
Furthermore, there is a lack of systematized meta-information that allows clinical AI researchers to quickly determine accessibility, scope, content and other characteristics of datasets and benchmark datasets relevant to the clinical domain.
1 code implementation • 4 Oct 2021 • Kathrin Blagec, Adriano Barbosa-Silva, Simon Ott, Matthias Samwald
Research in artificial intelligence (AI) is addressing a growing number of tasks through a rapidly growing number of models and methodologies.
1 code implementation • 1 Oct 2021 • Kathrin Blagec, Hong Xu, Asan Agibetov, Matthias Samwald
BACKGROUND: In this study, we investigated the efficacy of current state-of-the-art neural sentence embedding models for semantic similarity estimation of sentences from biomedical literature.
Ranked #1 on Sentence Embeddings For Biomedical Texts on BIOSSES
1 code implementation • 6 Sep 2021 • Milad Moradi, Kathrin Blagec, Florian Haberl, Matthias Samwald
However, in-domain pretraining seems not to be sufficient; novel pretraining and few-shot learning strategies are required in the biomedical NLP domain.
1 code implementation • 27 Aug 2021 • Milad Moradi, Kathrin Blagec, Matthias Samwald
The proposed perturbation methods can be used in performance evaluation tests to assess how robustly clinical NLP models can operate on noisy data, in real-world settings.
no code implementations • 6 Aug 2020 • Kathrin Blagec, Georg Dorffner, Milad Moradi, Matthias Samwald
Our results suggest that the large majority of metrics currently used have properties that may result in an inadequate reflection of a models' performance.