Search Results for author: Yuki Asano

Found 6 papers, 0 papers with code

Think Big, Generate Quick: LLM-to-SLM for Fast Autoregressive Decoding

no code implementations26 Feb 2024 Benjamin Bergner, Andrii Skliar, Amelie Royer, Tijmen Blankevoort, Yuki Asano, Babak Ehteshami Bejnordi

We investigate the combination of encoder-decoder LLMs with both encoder-decoder and decoder-only SLMs from different model families and only require fine-tuning of the SLM.

Instruction Following Language Modelling +1

Auto-Vocabulary Semantic Segmentation

no code implementations7 Dec 2023 Osman Ülger, Maksymilian Kulicki, Yuki Asano, Martin R. Oswald

Open-Vocabulary Segmentation (OVS) methods are capable of performing semantic segmentation without relying on a fixed vocabulary, and in some cases, they operate without the need for training or fine-tuning.

Language Modelling Large Language Model +3

Federated Fine-Tuning of Foundation Models via Probabilistic Masking

no code implementations29 Nov 2023 Vasileios Tsouvalas, Yuki Asano, Aaqib Saeed

Foundation Models (FMs) have revolutionized machine learning with their adaptability and high performance across tasks; yet, their integration into Federated Learning (FL) is challenging due to substantial communication overhead from their extensive parameterization.

Federated Learning

Support-set bottlenecks for video-text representation learning

no code implementations ICLR 2021 Mandela Patrick, Po-Yao Huang, Yuki Asano, Florian Metze, Alexander Hauptmann, João Henriques, Andrea Vedaldi

The dominant paradigm for learning video-text representations -- noise contrastive learning -- increases the similarity of the representations of pairs of samples that are known to be related, such as text and video from the same sample, and pushes away the representations of all other pairs.

Contrastive Learning Representation Learning +3

Multi-modal Self-Supervision from Generalized Data Transformations

no code implementations28 Sep 2020 Mandela Patrick, Yuki Asano, Polina Kuznetsova, Ruth Fong, Joao F. Henriques, Geoffrey Zweig, Andrea Vedaldi

In this paper, we show that, for videos, the answer is more complex, and that better results can be obtained by accounting for the interplay between invariance, distinctiveness, multiple modalities and time.

Audio Classification Retrieval +1

Cannot find the paper you are looking for? You can Submit a new open access paper.