no code implementations • 14 Aug 2024 • Zongze Wu, Nicholas Kolkin, Jonathan Brandt, Richard Zhang, Eli Shechtman
We address the challenges of precise image inversion and disentangled image editing in the context of few-step diffusion models.
no code implementations • CVPR 2024 • Cusuh Ham, Matthew Fisher, James Hays, Nicholas Kolkin, Yuchen Liu, Richard Zhang, Tobias Hinz
We present personalized residuals and localized attention-guided sampling for efficient concept-driven generation using text-to-image diffusion models.
no code implementations • 28 Nov 2023 • Xiaodan Du, Nicholas Kolkin, Greg Shakhnarovich, Anand Bhattad
Generative models excel at mimicking real scenes, suggesting they might inherently encode important intrinsic scene properties.
no code implementations • 7 Nov 2023 • Xingzhe He, Zhiwen Cao, Nicholas Kolkin, Lantao Yu, Kun Wan, Helge Rhodin, Ratheesh Kalarot
This strategy enables the model to preserve fine details of the desired subjects, such as text and logos.
no code implementations • 9 Jul 2023 • Dan Ruta, Gemma Canet Tarrés, Andrew Gilbert, Eli Shechtman, Nicholas Kolkin, John Collomosse
Neural Style Transfer (NST) is the field of study applying neural techniques to modify the artistic appearance of a content image to match the style of a reference style image.
1 code implementation • 11 Apr 2023 • Dan Ruta, Andrew Gilbert, John Collomosse, Eli Shechtman, Nicholas Kolkin
As a component of curating this data, we present a novel model able to classify if an image is stylistic.
1 code implementation • 8 Sep 2022 • Xiaodan Du, Raymond A. Yeh, Nicholas Kolkin, Eli Shechtman, Greg Shakhnarovich
We propose Fast text2StyleGAN, a natural language interface that adapts pre-trained GANs for text-guided human face synthesis.
1 code implementation • 24 Mar 2022 • Nicholas Kolkin, Michal Kucera, Sylvain Paris, Daniel Sykora, Eli Shechtman, Greg Shakhnarovich
We propose Neural Neighbor Style Transfer (NNST), a pipeline that offers state-of-the-art quality, generalization, and competitive efficiency for artistic style transfer.
no code implementations • ICCV 2021 • Cooper Nederhood, Nicholas Kolkin, Deqing Fu, Jason Salavon
Multi-modal domain translation typically refers to synthesizing a novel image that inherits certain localized attributes from a 'content' image (e. g. layout, semantics, or geometry), and inherits everything else (e. g. texture, lighting, sometimes even semantics) from a 'style' image.
no code implementations • 29 Aug 2021 • Nicholas Kolkin
It seems easy to imagine a photograph of the Eiffel Tower painted in the style of Vincent van Gogh's 'The Starry Night', but upon introspection it is difficult to precisely define what this would entail.
1 code implementation • ECCV 2020 • Sunnie S. Y. Kim, Nicholas Kolkin, Jason Salavon, Gregory Shakhnarovich
Both geometry and texture are fundamental aspects of visual style.
6 code implementations • CVPR 2019 • Nicholas Kolkin, Jason Salavon, Greg Shakhnarovich
Style transfer algorithms strive to render the content of one image using the style of another.
no code implementations • ICCV 2017 • Nicholas Kolkin, Gregory Shakhnarovich, Eli Shechtman
In many computer vision tasks, for example saliency prediction or semantic segmentation, the desired output is a foreground map that predicts pixels where some criteria is satisfied.
no code implementations • 6 Dec 2016 • Mohammadreza Mostajabi, Nicholas Kolkin, Gregory Shakhnarovich
We propose an approach for learning category-level semantic segmentation purely from image-level classification tags indicating presence of categories.