1 code implementation • EMNLP 2021 • Sangwoo Cho, Franck Dernoncourt, Tim Ganter, Trung Bui, Nedim Lipka, Walter Chang, Hailin Jin, Jonathan Brandt, Hassan Foroosh, Fei Liu
With the explosive growth of livestream broadcasting, there is an urgent need for new summarization technology that enables us to create a preview of streamed content and tap into this wealth of knowledge.
2 code implementations • ICCV 2021 • Hareesh Ravi, Kushal Kafle, Scott Cohen, Jonathan Brandt, Mubbasir Kapadia
Visual storytelling and story comprehension are uniquely human skills that play a central role in how we learn about and experience the world.
no code implementations • 25 Sep 2019 • Biao Jia, Jonathan Brandt, Radomir Mech, Ning Xu, Byungmoon Kim, Dinesh Manocha
We present a novel approach to train a natural media painting using reinforcement learning.
no code implementations • 17 Jun 2019 • Biao Jia, Jonathan Brandt, Radomir Mech, Byungmoon Kim, Dinesh Manocha
We present a novel reinforcement learning-based natural media painting algorithm.
no code implementations • 3 Apr 2019 • Biao Jia, Chen Fang, Jonathan Brandt, Byungmoon Kim, Dinesh Manocha
Action selection is guided by a given reference image, which the agent attempts to replicate subject to the limitations of the action space and the agent's learned policy.
1 code implementation • 26 Jan 2019 • Abby Stylianou, Hong Xuan, Maya Shende, Jonathan Brandt, Richard Souvenir, Robert Pless
Recognizing a hotel from an image of a hotel room is important for human trafficking investigations.
no code implementations • 14 Oct 2018 • Tao Zhou, Chen Fang, Zhaowen Wang, Jimei Yang, Byungmoon Kim, Zhili Chen, Jonathan Brandt, Demetri Terzopoulos
Doodling is a useful and common intelligent skill that people can learn and master.
no code implementations • CVPR 2017 • Long Mai, Hailin Jin, Zhe Lin, Chen Fang, Jonathan Brandt, Feng Liu
We train a convolutional neural network to synthesize appropriate visual features that captures the spatial-semantic constraints from the user canvas query.
3 code implementations • 1 Aug 2016 • Jianming Zhang, Zhe Lin, Jonathan Brandt, Xiaohui Shen, Stan Sclaroff
We aim to model the top-down attention of a Convolutional Neural Network (CNN) classifier for generating task-specific attention maps.
no code implementations • CVPR 2016 • Jae-Pil Heo, Zhe Lin, Xiaohui Shen, Jonathan Brandt, Sung-Eui Yoon
We have tested the proposed method with the inverted index and multi-index on a diverse set of benchmarks including up to one billion data points with varying dimensions, and found that our method robustly improves the accuracy of shortlists (up to 127% relatively higher) over the state-of-the-art techniques with a comparable or even faster computational cost.
no code implementations • CVPR 2016 • Haoxiang Li, Jonathan Brandt, Zhe Lin, Xiaohui Shen, Gang Hua
Our new framework enables efficient use of these complementary multi-level contextual cues to improve overall recognition rates on the photo album person recognition task, as demonstrated through state-of-the-art results on a challenging public dataset.
1 code implementation • 12 Jul 2015 • Zhangyang Wang, Jianchao Yang, Hailin Jin, Eli Shechtman, Aseem Agarwala, Jonathan Brandt, Thomas S. Huang
As font is one of the core design concepts, automatic font identification and similar font suggestion from an image or photo has been on the wish list of many designers.
Ranked #1 on
Font Recognition
on VFR-Wild
no code implementations • CVPR 2015 • Haoxiang Li, Zhe Lin, Xiaohui Shen, Jonathan Brandt, Gang Hua
To improve localization effectiveness, and reduce the number of candidates at later stages, we introduce a CNN-based calibration stage after each of the detection stages in the cascade.
no code implementations • 31 Mar 2015 • Zhangyang Wang, Jianchao Yang, Hailin Jin, Eli Shechtman, Aseem Agarwala, Jonathan Brandt, Thomas S. Huang
We address a challenging fine-grain classification problem: recognizing a font style from an image of text.
no code implementations • 18 Dec 2014 • Zhangyang Wang, Jianchao Yang, Hailin Jin, Eli Shechtman, Aseem Agarwala, Jonathan Brandt, Thomas S. Huang
We present a domain adaption framework to address a domain mismatch between synthetic training and real-world testing data.
no code implementations • CVPR 2014 • Guang Chen, Jianchao Yang, Hailin Jin, Jonathan Brandt, Eli Shechtman, Aseem Agarwala, Tony X. Han
This paper addresses the large-scale visual font recognition (VFR) problem, which aims at automatic identification of the typeface, weight, and slope of the text in an image or photo without any knowledge of content.
Ranked #1 on
Font Recognition
on VFR-447
no code implementations • CVPR 2014 • Brandon M. Smith, Jonathan Brandt, Zhe Lin, Li Zhang
We propose a data-driven approach to facial landmark localization that models the correlations between each landmark and its surrounding appearance features.
no code implementations • CVPR 2014 • Haoxiang Li, Zhe Lin, Jonathan Brandt, Xiaohui Shen, Gang Hua
Despite the fact that face detection has been studied intensively over the past several decades, the problem is still not completely solved.
no code implementations • 24 Apr 2014 • Zhaowen Wang, Jianchao Yang, Zhe Lin, Jonathan Brandt, Shiyu Chang, Thomas Huang
In this paper, we present an image similarity learning method that can scale well in both the number of images and the dimensionality of image descriptors.
no code implementations • CVPR 2013 • Brandon M. Smith, Li Zhang, Jonathan Brandt, Zhe Lin, Jianchao Yang
Given a test image, our algorithm first selects a subset of exemplar images from the database, Our algorithm then computes a nonrigid warp for each exemplar image to align it with the test image.
no code implementations • CVPR 2013 • Xiaohui Shen, Zhe Lin, Jonathan Brandt, Ying Wu
In order to overcome these challenges, we present a novel and robust exemplarbased face detector that integrates image retrieval and discriminative learning.
no code implementations • CVPR 2013 • Haoxiang Li, Gang Hua, Zhe Lin, Jonathan Brandt, Jianchao Yang
By augmenting each feature with its location, a Gaussian mixture model (GMM) is trained to capture the spatialappearance distribution of all face images in the training corpus.