1 code implementation • 7 Nov 2023 • Xiangyong Lu, Masanori Suganuma, Takayuki Okatani
For the first time, it achieves an ImageNet-1K top-1 accuracy of around 80% at a speed of 1. 0 frame/sec on the SBC.
1 code implementation • 7 Oct 2023 • Korawat Charoenpitaks, Van-Quang Nguyen, Masanori Suganuma, Masahiro Takahashi, Ryoma Niihara, Takayuki Okatani
To enable research in this understudied area, a new dataset named the DHPR (Driving Hazard Prediction and Reasoning) dataset is created.
no code implementations • 6 Jul 2023 • Jie Zhang, Masanori Suganuma, Takayuki Okatani
They consider an unsupervised setting, specifically the one-class setting, in which we assume the availability of a set of normal (\textit{i. e.}, anomaly-free) images for training.
no code implementations • 6 Jul 2023 • Jie Zhang, Masanori Suganuma, Takayuki Okatani
The local student, which is used in previous studies mainly focuses on structural anomaly detection while the global student pays attention to logical anomalies.
Ranked #10 on
Anomaly Detection
on MVTec LOCO AD
no code implementations • 6 Jul 2023 • Han Zou, Masanori Suganuma, Takayuki Okatani
We can utilize an alternative shot of the identical scene, just like in video deblurring, or we can even employ a distinct image from another scene.
no code implementations • 6 Jul 2023 • Han Zou, Masanori Suganuma, Takayuki Okatani
Then, we propose an improved method, RefVSR++, which can aggregate two features in parallel in the temporal direction, one for aggregating the fused LR and Ref inputs and the other for Ref inputs over time.
Reference-based Video Super-Resolution
Video Super-Resolution
1 code implementation • 20 Jul 2022 • Van-Quang Nguyen, Masanori Suganuma, Takayuki Okatani
Current state-of-the-art methods for image captioning employ region-based features, as they provide object-level information that is essential to describe the content of images; they are usually extracted by an object detector such as Faster R-CNN.
Ranked #7 on
Image Captioning
on nocaps in-domain
no code implementations • 20 Jul 2022 • Yusuke Hosoya, Masanori Suganuma, Takayuki Okatani
In this paper, we first point out that the recent studies' formalization of OSOD, which generalizes open-set recognition (OSR) and thus considers an unlimited variety of unknown objects, has a fundamental issue.
no code implementations • 7 Jul 2022 • Qian Ye, Masanori Suganuma, Takayuki Okatani
Considering the spatial variant property of the defocus blur and the blur level indicated in the defocus map, we employ the defocus map as conditional guidance to adjust the features from the input blurring images instead of simple concatenation.
no code implementations • 6 Jul 2022 • Qian Ye, Masanori Suganuma, Jun Xiao, Takayuki Okatani
Reconstructing ghosting-free high dynamic range (HDR) images of dynamic scenes from a set of multi-exposure images is a challenging task, especially with large object motion and occlusions, leading to visible artifacts using existing methods.
2 code implementations • 30 Jun 2022 • Zhijie Wang, Masanori Suganuma, Takayuki Okatani
Due to its high annotation cost, researchers have developed many UDA methods for semantic segmentation, which assume no labeled sample is available in the target domain.
no code implementations • 14 Sep 2021 • Zhijie Wang, Masanori Suganuma, Takayuki Okatani
This study is concerned with few-shot segmentation, i. e., segmenting the region of an unseen object class in a query image, given support image(s) of its instances.
no code implementations • 14 Sep 2021 • Zhijie Wang, Xing Liu, Masanori Suganuma, Takayuki Okatani
To cope with this, we propose a method that applies adversarial training to align two feature distributions in the target domain.
Ranked #1 on
Domain Adaptation
on Synscapes-to-Cityscapes
no code implementations • ICCV 2021 • Wenzheng Song, Masanori Suganuma, Xing Liu, Noriyuki Shimobayashi, Daisuke Maruta, Takayuki Okatani
To consider if and how well we can utilize such information stored in RAW-format images for image matching, we have created a new dataset named MID (matching in the dark).
1 code implementation • 1 Jun 2021 • Van-Quang Nguyen, Masanori Suganuma, Takayuki Okatani
It then integrates the prediction with the visual information etc., yielding the final prediction of an action and an object.
no code implementations • 7 May 2020 • Rito Murase, Masanori Suganuma, Takayuki Okatani
We draw a mixed conclusion from the experimental results; the positional encoding certainly works in some cases, but the absolute image position may not be so important for segmentation tasks as we think.
1 code implementation • ECCV 2020 • Van-Quang Nguyen, Masanori Suganuma, Takayuki Okatani
It has been a primary concern in recent studies of vision and language tasks to design an effective attention mechanism dealing with interactions between the two modalities.
no code implementations • 21 Oct 2019 • Yusuke Hosoya, Masanori Suganuma, Takayuki Okatani
The employment of convolutional neural networks has led to significant performance improvement on the task of object detection.
1 code implementation • 10 Jul 2019 • Xing Liu, Masanori Suganuma, Xiyang Luo, Takayuki Okatani
The employment of convolutional neural networks has achieved unprecedented performance in the task of image restoration for a variety of degradation factors.
1 code implementation • 25 May 2019 • Engkarat Techapanurak, Masanori Suganuma, Takayuki Okatani
The ability to detect out-of-distribution (OOD) samples is vital to secure the reliability of deep neural networks in real-world applications.
1 code implementation • CVPR 2019 • Xing Liu, Masanori Suganuma, Zhun Sun, Takayuki Okatani
In this paper, we study design of deep neural networks for tasks of image restoration.
1 code implementation • CVPR 2019 • Masanori Suganuma, Xing Liu, Takayuki Okatani
There are many different types of distortion which affect image quality.
1 code implementation • ICML 2018 • Masanori Suganuma, Mete Ozay, Takayuki Okatani
Researchers have applied deep neural networks to image restoration tasks, in which they proposed various network architectures, loss functions, and training methods.
5 code implementations • 3 Apr 2017 • Masanori Suganuma, Shinichi Shirakawa, Tomoharu Nagao
To evaluate the proposed method, we constructed a CNN architecture for the image classification task with the CIFAR-10 dataset.