1 code implementation • ACL 2022 • Fredrik Carlsson, Joey Öhman, Fangyu Liu, Severine Verlinden, Joakim Nivre, Magnus Sahlgren
We propose a resource-efficient method for converting a pre-trained CLM into this architecture, and demonstrate its potential on various experiments, including the novel task of contextualized word inclusion.
no code implementations • WNUT (ACL) 2021 • Thomas Clark, Costanza Conforti, Fangyu Liu, Zaiqiao Meng, Ehsan Shareghi, Nigel Collier
Stance detection (SD) entails classifying the sentiment of a text towards a given target, and is a relevant sub-task for opinion mining and social media analysis.
no code implementations • 23 Mar 2023 • Fangyu Liu, Qianchu Liu, Shruthi Bannur, Fernando Pérez-García, Naoto Usuyama, Sheng Zhang, Tristan Naumann, Aditya Nori, Hoifung Poon, Javier Alvarez-Valle, Ozan Oktay, Stephanie L. Hyland
We evaluate DoT5 on the biomedical domain and the resource-lean subdomain of radiology, focusing on NLI, text summarisation and embedding learning.
no code implementations • 17 Jan 2023 • Junjie Zhou, Yongping Xiong, Chinwai Chiu, Fangyu Liu, Xiangyang Gong
In this paper, we propose the Size-Aware Transformer (SAT) that can tailor effective receptive fields for objects of different sizes.
Ranked #2 on
Semantic Segmentation
on S3DIS Area5
1 code implementation • 20 Dec 2022 • Fangyu Liu, Julian Martin Eisenschlos, Francesco Piccinno, Syrine Krichene, Chenxi Pang, Kenton Lee, Mandar Joshi, Wenhu Chen, Nigel Collier, Yasemin Altun
Compared with a SOTA model finetuned on more than >28k data points, DePlot+LLM with just one-shot prompting achieves a 24. 0% improvement over finetuned SOTA on human-written queries from the task of chart QA.
Ranked #1 on
Chart Question Answering
on ChartQA
no code implementations • 20 Dec 2022 • Tianqing Fang, Wenxuan Zhou, Fangyu Liu, Hongming Zhang, Yangqiu Song, Muhao Chen
However, data augmentation may introduce noisy data that impairs training.
1 code implementation • 19 Dec 2022 • Fangyu Liu, Francesco Piccinno, Syrine Krichene, Chenxi Pang, Kenton Lee, Mandar Joshi, Yasemin Altun, Nigel Collier, Julian Martin Eisenschlos
Visual language data such as plots, charts, and infographics are ubiquitous in the human world.
Ranked #1 on
Visual Question Answering (VQA)
on PlotQA-D1
1 code implementation • 7 Nov 2022 • Songbo Hu, Ivan Vulić, Fangyu Liu, Anna Korhonen
At training, the high-scoring partition comprises all generated responses whose similarity to the gold response is higher than the similarity of the greedy response to the gold response.
1 code implementation • 30 Oct 2022 • Yaoyiran Li, Fangyu Liu, Ivan Vulić, Anna Korhonen
This crucial step is done via 1) creating a word similarity dataset, comprising positive word pairs (i. e., true translations) and hard negative pairs induced from the original CLWE space, and then 2) fine-tuning an mPLM (e. g., mBERT or XLM-R) in a cross-encoder manner to predict the similarity scores.
Bilingual Lexicon Induction
Cross-Lingual Word Embeddings
+7
1 code implementation • 7 Oct 2022 • Kenton Lee, Mandar Joshi, Iulia Turc, Hexiang Hu, Fangyu Liu, Julian Eisenschlos, Urvashi Khandelwal, Peter Shaw, Ming-Wei Chang, Kristina Toutanova
Visually-situated language is ubiquitous -- sources range from textbooks with diagrams to web pages with images and tables, to mobile apps with buttons and forms.
Ranked #3 on
Chart Question Answering
on ChartQA
1 code implementation • 29 Sep 2022 • Aleksander Ficek, Fangyu Liu, Nigel Collier
Being able to train Named Entity Recognition (NER) models for emerging topics is crucial for many real-world applications especially in the medical domain where new topics are continuously evolving out of the scope of existing models and datasets.
no code implementations • 26 Sep 2022 • Fangyu Liu, Julian Martin Eisenschlos, Jeremy R. Cole, Nigel Collier
Language models (LMs) trained on raw texts have no direct access to the physical world.
no code implementations • 25 Sep 2022 • Julian Martin Eisenschlos, Jeremy R. Cole, Fangyu Liu, William W. Cohen
We introduce a new in-context learning paradigm to measure Large Language Models' (LLMs) ability to learn novel words during inference.
no code implementations • 25 Aug 2022 • Nigel H. Collier, Fangyu Liu, Ehsan Shareghi
Recent advancements in Large Language Models (LLMs) harness linguistic associations in vast natural language data for practical applications.
1 code implementation • 29 Jun 2022 • Jose Camacho-Collados, Kiamehr Rezaee, Talayeh Riahi, Asahi Ushio, Daniel Loureiro, Dimosthenis Antypas, Joanne Boisson, Luis Espinosa-Anke, Fangyu Liu, Eugenio Martínez-Cámara, Gonzalo Medina, Thomas Buhrmann, Leonardo Neves, Francesco Barbieri
In this paper we present TweetNLP, an integrated platform for Natural Language Processing (NLP) in social media.
1 code implementation • 5 May 2022 • Yixuan Su, Tian Lan, Yahui Liu, Fangyu Liu, Dani Yogatama, Yan Wang, Lingpeng Kong, Nigel Collier
MAGIC is a flexible framework and is theoretically compatible with any text generation tasks that incorporate image grounding.
2 code implementations • 30 Apr 2022 • Fangyu Liu, Guy Emerson, Nigel Collier
Spatial relations are a basic part of human cognition.
Ranked #1 on
Visual Reasoning
on VSR
no code implementations • 30 Apr 2022 • Ivan Vulić, Goran Glavaš, Fangyu Liu, Nigel Collier, Edoardo Maria Ponti, Anna Korhonen
In this work, we probe SEs for the amount of cross-lingual lexical knowledge stored in their parameters, and compare them against the original multilingual LMs.
no code implementations • 18 Apr 2022 • Xun Wang, Bingqing Ke, Xuanping Li, Fangyu Liu, Mingyu Zhang, Xiao Liang, Qiushi Xiao, Cheng Luo, Yue Yu
This modality imbalanceresults from a) modality gap: the relevance between a query and a video text is much easier to learn as the query is also a piece of text, with the same modality as the video text; b) data bias: most training samples can be solved solely by text matching.
1 code implementation • ACL 2022 • Yaoyiran Li, Fangyu Liu, Nigel Collier, Anna Korhonen, Ivan Vulić
At Stage C1, we propose to refine standard cross-lingual linear maps between static word embeddings (WEs) via a contrastive learning objective; we also show how to integrate it into the self-learning procedure for even more refined cross-lingual maps.
1 code implementation • 16 Feb 2022 • Guanzheng Chen, Fangyu Liu, Zaiqiao Meng, Shangsong Liang
Parameter-Efficient Tuning (PETuning) methods have been deemed by many as the new paradigm for using pretrained language models (PLMs).
2 code implementations • 27 Jan 2022 • Emanuele Bugliarello, Fangyu Liu, Jonas Pfeiffer, Siva Reddy, Desmond Elliott, Edoardo Maria Ponti, Ivan Vulić
Our benchmark enables the evaluation of multilingual multimodal models for transfer learning, not only in a zero-shot setting, but also in newly defined few-shot learning setups.
no code implementations • 16 Dec 2021 • Wenxuan Zhou, Fangyu Liu, huan zhang, Muhao Chen
Deep neural networks are often overparameterized and may not easily achieve model generalization.
2 code implementations • Findings (NAACL) 2022 • Yixuan Su, Fangyu Liu, Zaiqiao Meng, Tian Lan, Lei Shu, Ehsan Shareghi, Nigel Collier
Masked language models (MLMs) such as BERT and RoBERTa have revolutionized the field of Natural Language Understanding in the past few years.
1 code implementation • ACL 2022 • Wenxuan Zhou, Fangyu Liu, Ivan Vulić, Nigel Collier, Muhao Chen
To achieve this, it is crucial to represent multilingual knowledge in a shared/unified space.
1 code implementation • ACL 2022 • Zaiqiao Meng, Fangyu Liu, Ehsan Shareghi, Yixuan Su, Charlotte Collins, Nigel Collier
To catalyse the research in this direction, we release a well-curated biomedical knowledge probing benchmark, MedLAMA, which is constructed based on the Unified Medical Language System (UMLS) Metathesaurus.
2 code implementations • EMNLP 2021 • Fangyu Liu, Emanuele Bugliarello, Edoardo Maria Ponti, Siva Reddy, Nigel Collier, Desmond Elliott
The design of widespread vision-and-language datasets and pre-trained encoders directly adopts, or draws inspiration from, the concepts and images of ImageNet.
Ranked #1 on
Zero-Shot Cross-Lingual Transfer
on MaRVL
1 code implementation • ICLR 2022 • Fangyu Liu, Yunlong Jiao, Jordan Massiah, Emine Yilmaz, Serhii Havrylov
Predominantly, two formulations are used for sentence-pair tasks: bi-encoders and cross-encoders.
Ranked #1 on
Semantic Textual Similarity
on STS16
1 code implementation • CoNLL (EMNLP) 2021 • Qianchu Liu, Fangyu Liu, Nigel Collier, Anna Korhonen, Ivan Vulić
Recent work indicated that pretrained language models (PLMs) such as BERT and RoBERTa can be transformed into effective sentence and word encoders even via simple self-supervised techniques.
1 code implementation • EMNLP 2021 • Zaiqiao Meng, Fangyu Liu, Thomas Hikaru Clark, Ehsan Shareghi, Nigel Collier
Infusing factual knowledge into pre-trained models is fundamental for many knowledge-intensive tasks.
1 code implementation • ACL 2021 • Fangyu Liu, Ivan Vulić, Anna Korhonen, Nigel Collier
To this end, we propose and evaluate a series of cross-lingual transfer methods for the XL-BEL task, and demonstrate that general-domain bitext helps propagate the available English knowledge to languages with little to no in-domain data.
1 code implementation • EMNLP 2021 • Wenxuan Zhou, Fangyu Liu, Muhao Chen
Pretrained Transformers achieve remarkable performance when training and test data are from the same distribution.
1 code implementation • EMNLP 2021 • Fangyu Liu, Ivan Vulić, Anna Korhonen, Nigel Collier
In this work, we demonstrate that it is possible to turn MLMs into effective universal lexical and sentence encoders even without any additional data and without any supervision.
Ranked #9 on
Semantic Textual Similarity
on STS16
Contrastive Learning
Cross-Lingual Semantic Textual Similarity
+4
1 code implementation • 1 Nov 2020 • Jia-Hong Huang, Chao-Han Huck Yang, Fangyu Liu, Meng Tian, Yi-Chieh Liu, Ting-Wei Wu, I-Hung Lin, Kang Wang, Hiromasa Morikawa, Hernghua Chang, Jesper Tegner, Marcel Worring
To train and validate the effectiveness of our DNN-based module, we propose a large-scale retinal disease image dataset.
1 code implementation • NAACL 2021 • Fangyu Liu, Ehsan Shareghi, Zaiqiao Meng, Marco Basaldella, Nigel Collier
Despite the widespread success of self-supervised learning via masked language models (MLM), accurately capturing fine-grained semantic relationships in the biomedical domain remains a challenge.
1 code implementation • EMNLP 2020 • Marco Basaldella, Fangyu Liu, Ehsan Shareghi, Nigel Collier
Whilst there has been growing progress in Entity Linking (EL) for general language, existing datasets fail to address the complex nature of health terminology in layman's language.
2 code implementations • 28 Sep 2020 • Fangyu Liu, Muhao Chen, Dan Roth, Nigel Collier
This work studies the use of visual semantic representations to align entities in heterogeneous knowledge graphs (KGs).
Ranked #11 on
Entity Alignment
on dbp15k ja-en
(using extra training data)
no code implementations • 23 Apr 2020 • Fangyu Liu, Rémi Lebret, Didier Orel, Philippe Sordet, Karl Aberer
The system fuses multiple textual sources extracted from news articles and accepts multilingual inputs.
1 code implementation • 22 Nov 2019 • Fangyu Liu, Rongtian Ye, Xun Wang, Shuaipeng Li
The hubness problem widely exists in high-dimensional embedding space and is a fundamental source of error for cross-modal matching tasks.
no code implementations • ACL 2019 • Fangyu Liu, Rongtian Ye
We review the current schemes of text-image matching models and propose improvements for both training and inference.
1 code implementation • 16 Aug 2018 • C. -H. Huck Yang, Fangyu Liu, Jia-Hong Huang, Meng Tian, Hiromasa Morikawa, I-Hung Lin, Yi-Chieh Liu, Hao-Hsiang Yang, Jesper Tegner
Automatic clinical diagnosis of retinal diseases has emerged as a promising approach to facilitate discovery in areas with limited access to specialists.
no code implementations • 5 Aug 2018 • Rongtian Ye, Fangyu Liu, Liqiang Zhang
Standard 3D convolution operations require much larger amounts of memory and computation cost than 2D convolution operations.
1 code implementation • 17 Jun 2018 • C. -H. Huck Yang, Jia-Hong Huang, Fangyu Liu, Fang-Yi Chiu, Mengya Gao, Weifeng Lyu, I-Hung Lin M. D., Jesper Tegner
Automatic clinical diagnosis of retinal diseases has emerged as a promising approach to facilitate discovery in areas with limited access to specialists.
no code implementations • ICCV 2017 • Fangyu Liu, Shuaipeng Li, Liqiang Zhang, Chenghu Zhou, Rongtian Ye, Yuebin Wang, Jiwen Lu
Our method provides an automatic process that maps the raw data to the classification results.