no code implementations • FL4NLP (ACL) 2022 • Luke Melas-Kyriazi, Franklyn Wang
Federated learning is a rapidly growing area of research, holding the promise of privacy-preserving distributed training on edge devices.
no code implementations • 12 Sep 2024 • Runjia Li, Junlin Han, Luke Melas-Kyriazi, Chunyi Sun, Zhaochong An, Zhongrui Gui, Shuyang Sun, Philip Torr, Tomas Jakab
Existing SDS methods often struggle with this generation task due to a limited understanding of part-level semantics in text-to-image diffusion models.
1 code implementation • 21 May 2024 • Garrett Tanzer, Gustaf Ahdritz, Luke Melas-Kyriazi
Chatbots built upon language models have exploded in popularity, but they have largely been limited to synchronous, turn-by-turn dialogues.
1 code implementation • 15 Feb 2024 • Abdullah Hamdi, Luke Melas-Kyriazi, Jinjie Mai, Guocheng Qian, Ruoshi Liu, Carl Vondrick, Bernard Ghanem, Andrea Vedaldi
With the aid of a frequency-modulated loss, GES achieves competitive performance in novel-view synthesis benchmarks while requiring less than half the memory storage of Gaussian Splatting and increasing the rendering speed by up to 39%.
no code implementations • 13 Feb 2024 • Luke Melas-Kyriazi, Iro Laina, Christian Rupprecht, Natalia Neverova, Andrea Vedaldi, Oran Gafni, Filippos Kokkinos
A mitigation is to fine-tune the 2D generator to be multi-view aware, which can help distillation or can be combined with reconstruction networks to output 3D objects directly.
1 code implementation • CVPR 2024 • Xingjian Bai, Luke Melas-Kyriazi
We introduce the Fixed Point Diffusion Model (FPDM), a novel approach to image generation that integrates the concept of fixed point solving into the framework of diffusion-based generative modeling.
no code implementations • CVPR 2024 • Abdullah Hamdi, Luke Melas-Kyriazi, Jinjie Mai, Guocheng Qian, Ruoshi Liu, Carl Vondrick, Bernard Ghanem, Andrea Vedaldi
With the aid of a frequency-modulated loss GES achieves competitive performance in novel-view synthesis benchmarks while requiring less than half the memory storage of Gaussian Splatting and increasing the rendering speed by up to 39%.
no code implementations • 24 Nov 2023 • Paul Engstler, Luke Melas-Kyriazi, Christian Rupprecht, Iro Laina
Self-supervised learning (SSL) can be used to solve complex visual tasks without human labels.
Ranked #3 on
Unsupervised Instance Segmentation
on COCO val2017
1 code implementation • 28 Sep 2023 • Garrett Tanzer, Mirac Suzgun, Eline Visser, Dan Jurafsky, Luke Melas-Kyriazi
In this paper, we introduce MTOB (Machine Translation from One Book), a benchmark for learning to translate between English and Kalamang -- a language with less than 200 speakers and therefore virtually no presence on the web -- using several hundred pages of field linguistics reference materials.
no code implementations • 23 Aug 2023 • Luke W. Sagers, James A. Diao, Luke Melas-Kyriazi, Matthew Groh, Pranav Rajpurkar, Adewole S. Adamson, Veronica Rotemberg, Roxana Daneshjou, Arjun K. Manrai
While hundreds of artificial intelligence (AI) algorithms are now approved or cleared by the US Food and Drugs Administration (FDA), many studies have shown inconsistent generalization or latent bias, particularly for underrepresented populations.
2 code implementations • 21 Feb 2023 • Luke Melas-Kyriazi, Christian Rupprecht, Andrea Vedaldi
Reconstructing the 3D shape of an object from a single RGB image is a long-standing and highly challenging problem in computer vision.
4 code implementations • 21 Feb 2023 • Luke Melas-Kyriazi, Christian Rupprecht, Iro Laina, Andrea Vedaldi
We consider the problem of reconstructing a full 360{\deg} photographic model of an object from a single image of it.
no code implementations • CVPR 2023 • Luke Melas-Kyriazi, Iro Laina, Christian Rupprecht, Andrea Vedaldi
We consider the problem of reconstructing a full 360deg photographic model of an object from a single image of it.
no code implementations • CVPR 2023 • Luke Melas-Kyriazi, Christian Rupprecht, Andrea Vedaldi
Reconstructing the 3D shape of an object from a single RGB image is a long-standing problem in computer vision.
1 code implementation • 14 Nov 2022 • Mirac Suzgun, Luke Melas-Kyriazi, Dan Jurafsky
In open-ended natural-language generation, existing text decoding methods typically struggle to produce text which is both diverse and high-quality.
1 code implementation • NeurIPS 2023 • Mirac Suzgun, Luke Melas-Kyriazi, Suproteem K. Sarkar, Scott Duke Kominers, Stuart M. Shieber
Innovation is a major driver of economic and social development, and information about many kinds of innovation is embedded in semi-structured data from patents and patent applications.
1 code implementation • 23 May 2022 • Mirac Suzgun, Luke Melas-Kyriazi, Dan Jurafsky
We propose a method for arbitrary textual style transfer (TST)--the task of transforming a text into any given style--utilizing general-purpose pre-trained language models.
1 code implementation • CVPR 2022 • Luke Melas-Kyriazi, Christian Rupprecht, Iro Laina, Andrea Vedaldi
We find that these eigenvectors already decompose an image into meaningful segments, and can be readily used to localize objects in a scene.
no code implementations • 5 Dec 2021 • Luke Melas-Kyriazi, Franklyn Wang
Federated learning is a rapidly-growing area of research which enables a large number of clients to jointly train a machine learning model on privately-held data.
1 code implementation • CVPR 2021 • Luke Melas-Kyriazi, Arjun K. Manrai
In this work, we present a novel framework for unsupervised domain adaptation based on the notion of target-domain consistency training.
Ranked #29 on
Synthetic-to-Real Translation
on SYNTHIA-to-Cityscapes
1 code implementation • ICLR 2022 • Luke Melas-Kyriazi, Christian Rupprecht, Iro Laina, Andrea Vedaldi
Recent research has shown that numerous human-interpretable directions exist in the latent space of GANs.
2 code implementations • 6 May 2021 • Luke Melas-Kyriazi
These results indicate that aspects of vision transformers other than attention, such as the patch embedding, may be more responsible for their strong performance than previously thought.
Ranked #967 on
Image Classification
on ImageNet
no code implementations • 30 Oct 2020 • Luke Melas-Kyriazi
Manifold learning is a popular and quickly-growing subfield of machine learning based on the assumption that one's observed data lie on a low-dimensional manifold embedded in a higher-dimensional space.
1 code implementation • CVPR 2020 • Fawaz Sammani, Luke Melas-Kyriazi
Specifically, our caption-editing model consisting of two sub-modules: (1) EditNet, a language module with an adaptive copy mechanism (Copy-LSTM) and a Selective Copy Memory Attention mechanism (SCMA), and (2) DCNet, an LSTM-based denoising auto-encoder.
no code implementations • WS 2019 • Luke Melas-Kyriazi, George Han, Celine Liang
Recent research points to knowledge distillation as a potential solution, showing that when training data for a given task is abundant, it is possible to distill a large (teacher) LM into a small task-specific (student) network with minimal loss of performance.
1 code implementation • 19 Aug 2019 • Zachary M. Ziegler, Luke Melas-Kyriazi, Sebastian Gehrmann, Alexander M. Rush
Large pretrained language models have changed the way researchers approach discriminative natural language understanding tasks, leading to the dominance of approaches that adapt a pretrained model for arbitrary downstream tasks.
1 code implementation • EMNLP 2018 • Luke Melas-Kyriazi, Alex Rush, er, George Han
Image paragraph captioning models aim to produce detailed descriptions of a source image.