no code implementations • ECCV 2020 • Sho Inayoshi, Keita Otani, Antonio Tejero-de-Pablos, Tatsuya Harada
In this paper, we propose the bounding-box channels, a novel architecture capable of relating the semantic, spatial, and image features strongly.
no code implementations • CAI (COLING) 2022 • Kohtaro Tanaka, Hiroaki Yamane, Yusuke Mori, Yusuke Mukuta, Tatsuya Harada
Memes are a widely used means of communication on social media platforms, and are known for their ability to “go viral”.
no code implementations • In2Writing (ACL) 2022 • Yusuke Mori, Hiroaki Yamane, Ryohei Shimizu, Tatsuya Harada
Emotions are essential for storytelling and narrative generation, and as such, the relationship between stories and emotions has been extensively studied.
1 code implementation • COLING (LaTeCHCLfL, CLFL, LaTeCH) 2020 • Yusuke Mori, Hiroaki Yamane, Yusuke Mukuta, Tatsuya Harada
We first conduct an experiment focusing on MPP, and our analysis shows that highly accurate predictions can be obtained when the missing part of a story is the beginning or the end.
no code implementations • 10 Mar 2023 • Ziteng Cui, Lin Gu, Xiao Sun, Yu Qiao, Tatsuya Harada
Common capture low-light scenes are challenging for most computer vision techniques, including Neural Radiance Fields (NeRF).
no code implementations • 8 Mar 2023 • Yusuke Mukuta, Tatsuya Harada
To ensure that training is consistent with the equivariance, we propose two concepts for self-supervised tasks: equivariant pretext labels and invariant contrastive loss.
no code implementations • 7 Mar 2023 • Kazuma Kobayashi, Lin Gu, Ryuichiro Hataya, Takaaki Mizuno, Mototaka Miyake, Hirokazu Watanabe, Masamichi Takahashi, Yasuyuki Takamizawa, Yukihiro Yoshida, Satoshi Nakamura, Nobuji Kouno, Amina Bolatkan, Yusuke Kurose, Tatsuya Harada, Ryuji Hamamoto
As a result, our SBMIR system enabled users to overcome previous challenges, including image retrieval based on fine-grained image characteristics, image retrieval without example images, and image retrieval for isolated samples.
no code implementations • 19 Feb 2023 • Xinyue Hu, Lin Gu, Kazuma Kobayashi, Qiyuan An, Qingyu Chen, Zhiyong Lu, Chang Su, Tatsuya Harada, Yingying Zhu
Medical visual question answering (VQA) aims to answer clinically relevant questions regarding input medical images.
1 code implementation • WACV 2023 • Thomas Westfechtel, Hao-Wei Yeh, Qier Meng, Yusuke Mukuta, Tatsuya Harada
Firstly, it lets the domain classifier focus on features that are important for the classification, and, secondly, it couples the classification and adversarial branch more closely.
Ranked #5 on
Domain Adaptation
on Office-31
no code implementations • 7 Dec 2022 • Shenghan Su, Lin Gu, Yue Yang, Jingjing Shen, Hiroaki Yamane, Zenghui Zhang, Tatsuya Harada
Besides, our colour quantisation method also offers an efficient quantisation method that effectively compresses the image storage while maintaining high performance in high-level recognition tasks such as classification and detection.
no code implementations • 12 Oct 2022 • Kohei Uehara, Tatsuya Harada
Our pipeline consists of two components: the Object Classifier, which performs knowledge-based object recognition, and the Question Generator, which generates knowledge-aware questions to acquire novel knowledge.
no code implementations • 2 Oct 2022 • Bumjun Jung, Yusuke Mukuta, Tatsuya Harada
Time-series data analysis is important because numerous real-world tasks such as forecasting weather, electricity consumption, and stock market involve predicting data that vary over time.
no code implementations • 13 Aug 2022 • Xinyue Hu, Lin Gu, Liangchen Liu, Ruijiang Li, Chang Su, Tatsuya Harada, Yingying Zhu
Existing video domain adaption (DA) methods need to store all temporal combinations of video frames or pair the source and target videos, which are memory cost expensive and can't scale up to long videos.
1 code implementation • 5 Aug 2022 • Ziteng Cui, Yingying Zhu, Lin Gu, Guo-Jun Qi, Xiaoxiao Li, Renrui Zhang, Zenghui Zhang, Tatsuya Harada
Image restoration algorithms such as super resolution (SR) are indispensable pre-processing modules for object detection in low quality images.
no code implementations • 25 Jul 2022 • Tianhan Xu, Tatsuya Harada
Recent advances in radiance fields enable photorealistic rendering of static or dynamic 3D scenes, but still do not support explicit deformation that is used for scene manipulation or animation.
no code implementations • 13 Jul 2022 • Nabarun Goswami, Tatsuya Harada
The mapping of text to speech (TTS) is non-deterministic, letters may be pronounced differently based on context, or phonemes can vary depending on various physiological and stylistic factors like gender, age, accent, emotions, etc.
1 code implementation • 30 May 2022 • Ziteng Cui, Kunchang Li, Lin Gu, Shenghan Su, Peng Gao, Zhengkai Jiang, Yu Qiao, Tatsuya Harada
Challenging illumination conditions (low-light, under-exposure and over-exposure) in the real world not only cast an unpleasant visual appearance but also taint the computer vision tasks.
Ranked #2 on
Image Enhancement
on Exposure-Errors
1 code implementation • 25 May 2022 • Yang Li, Tatsuya Harada
Non-rigid point cloud registration is a key component in many computer vision and computer graphics applications.
no code implementations • 23 May 2022 • Yusuke Mori, Hiroaki Yamane, Yusuke Mukuta, Tatsuya Harada
Storytelling has always been vital for human nature.
2 code implementations • ICCV 2021 • Ziteng Cui, Guo-Jun Qi, Lin Gu, ShaoDi You, Zenghui Zhang, Tatsuya Harada
To enhance object detection in a dark environment, we propose a novel multitask auto encoding transformation (MAET) model which is able to explore the intrinsic pattern behind illumination translation.
Ranked #1 on
2D object detection
on ExDark
no code implementations • 19 Apr 2022 • Atsuhiro Noguchi, Xiao Sun, Stephen Lin, Tatsuya Harada
We propose an unsupervised method for 3D geometry-aware representation learning of articulated objects, in which no image-pose pairs or foreground masks are used for training.
1 code implementation • 24 Mar 2022 • Ryoma Kobayashi, Yusuke Mukuta, Tatsuya Harada
Learning from Label Proportions (LLP) is a weakly supervised learning method that aims to perform instance classification from training data consisting of pairs of bags containing multiple instances and the class label proportions within the bags.
1 code implementation • CVPR 2022 • Jiawei Zhang, Xiang Wang, Xiao Bai, Chen Wang, Lei Huang, Yimin Chen, Lin Gu, Jun Zhou, Tatsuya Harada, Edwin R. Hancock
The stereo contrastive feature loss function explicitly constrains the consistency between learned features of matching pixel pairs which are observations of the same 3D points.
no code implementations • 18 Mar 2022 • Takayuki Hara, Tatsuya Harada
We present a method for synthesizing novel views from a single 360-degree image based on the neural radiance field (NeRF) .
no code implementations • 15 Mar 2022 • Kohei Uehara, Tatsuya Harada
Visual Question Generation (VQG) is a task to generate questions from images.
no code implementations • 26 Feb 2022 • Yusuke Mori, Hiroaki Yamane, Ryohei Shimizu, Yusuke Mukuta, Tatsuya Harada
Furthermore, based on the novel task and methods, we developed a creative writing support system, COMPASS.
no code implementations • 15 Feb 2022 • Kohei Uehara, Yusuke Mori, Yusuke Mukuta, Tatsuya Harada
Image narrative generation is a task to create a story from an image with a subjective viewpoint.
no code implementations • 25 Jan 2022 • Antonio Tejero-de-Pablos, Hiroaki Yamane, Yusuke Kurose, Junichi Iho, Youji Tokunaga, Makoto Horie, Keisuke Nishizawa, Yusaku Hayashi, Yasushi Koyama, Tatsuya Harada
However, in the case of coronary artery wall, even state-of-the-art segmentation methods fail to produce an accurate boundary in the presence of plaques and bifurcations.
no code implementations • 7 Jan 2022 • Ziteng Cui, Yingying Zhu, Lin Gu, Guo-Jun Qi, Xiaoxiao Li, Peng Gao, Zenghui Zhang, Tatsuya Harada
Image restoration algorithms such as super resolution (SR) are indispensable pre-processing modules for object detection in degraded images.
1 code implementation • CVPR 2022 • Atsuhiro Noguchi, Umar Iqbal, Jonathan Tremblay, Tatsuya Harada, Orazio Gallo
Rendering articulated objects while controlling their poses is critical to applications such as virtual reality or animation for movies.
no code implementations • 2 Dec 2021 • Yifei HUANG, Xiaoxiao Li, Lijin Yang, Lin Gu, Yingying Zhu, Hirofumi Seo, Qiuming Meng, Tatsuya Harada, Yoichi Sato
Then we design a novel Auxiliary Attention Block (AAB) to allow information from SAN to be utilized by the backbone encoder to focus on selective areas.
1 code implementation • CVPR 2022 • Yang Li, Tatsuya Harada
We present Lepard, a Learning based approach for partial point cloud matching in rigid and deformable scenes.
Ranked #1 on
Partial Point Cloud Matching
on 4DMatch
no code implementations • 8 Oct 2021 • Yuki Kawana, Yusuke Mukuta, Tatsuya Harada
Articulated objects exist widely in the real world.
no code implementations • 29 Sep 2021 • Toru Makuuchi, Yusuke Mukuta, Tatsuya Harada
In this study, we analyze the generalization bound for the time-varying case by applying PAC-Bayes and experimentally show that the theoretical functional form for the batch size and learning rate approximates the generalization error well for both cases.
1 code implementation • 26 Sep 2021 • Hiromichi Kamata, Yusuke Mukuta, Tatsuya Harada
Spiking neural networks (SNNs) can be run on neuromorphic devices with ultra-high speed and ultra-low energy consumption because of their binary and event-driven nature.
no code implementations • 25 Jun 2021 • Sho Maeoki, Yusuke Mukuta, Tatsuya Harada
In this paper, we propose a novel training method that takes advantage of potentially relevant pairs, which are detected based on linguistic analysis about text annotation.
no code implementations • 7 Jun 2021 • Naoya Fushishita, Antonio Tejero-de-Pablos, Yusuke Mukuta, Tatsuya Harada
In this paper, we propose a novel method for generating future prediction videos with less memory usage than the conventional methods.
1 code implementation • ICCV 2021 • Atsuhiro Noguchi, Xiao Sun, Stephen Lin, Tatsuya Harada
We present Neural Articulated Radiance Field (NARF), a novel deformable 3D representation for articulated objects learned from images.
no code implementations • 23 Mar 2021 • Kazuma Kobayashi, Ryuichiro Hataya, Yusuke Kurose, Mototaka Miyake, Masamichi Takahashi, Akiko Nakagawa, Tatsuya Harada, Ryuji Hamamoto
To support comparative diagnostic reading, content-based image retrieval (CBIR), which can selectively utilize normal and abnormal features in medical images as two separable semantic components, will be useful.
1 code implementation • 7 Mar 2021 • Xiaoxiao Li, Ziteng Cui, Yifan Wu, Lin Gu, Tatsuya Harada
To tackle this issue, we propose an adversarial multi-task training strategy to simultaneously mitigate and detect bias in the deep learning-based medical image analysis system.
1 code implementation • CVPR 2021 • Yang Liu, Lei Zhou, Xiao Bai, Yifei HUANG, Lin Gu, Jun Zhou, Tatsuya Harada
Therefore, we introduce a novel goal-oriented gaze estimation module (GEM) to improve the discriminative attribute localization based on the class-level attributes for ZSL.
no code implementations • 1 Jan 2021 • Dexuan Zhang, Tatsuya Harada
In this paper, we argue that the joint error is essential for the domain adaptation problem, in particular if the samples from different classes in source/target are closely aligned when matching the marginal distributions.
no code implementations • 12 Nov 2020 • Kazuma Kobayashi, Ryuichiro Hataya, Yusuke Kurose, Tatsuya Harada, Ryuji Hamamoto
Medical images can be decomposed into normal and abnormal features, which is considered as the compositionality.
no code implementations • NeurIPS 2020 • Yuki Kawana, Yusuke Mukuta, Tatsuya Harada
We show that NSD is a universal approximator of the star domain and is not only parsimonious and semantic but also an implicit and explicit shape representation.
no code implementations • 16 Sep 2020 • Yang Liu, Lei Zhou, Xiao Bai, Lin Gu, Tatsuya Harada, Jun Zhou
Though many ZSL methods rely on a direct mapping between the visual and the semantic space, the calibration deviation and hubness problem limit the generalization capability to unseen classes.
no code implementations • 4 Aug 2020 • Adrien Bitton, Philippe Esling, Tatsuya Harada
In this setting the learned grain space is invertible, meaning that we can continuously synthesize sound when traversing its dimensions.
no code implementations • 3 Aug 2020 • Yujin Tang, Jie Tan, Tatsuya Harada
In contrast to prior works that used only one adversary, we find that training an ensemble of adversaries, each of which specializes in a different escaping strategy, is essential for the protagonist to master agility.
no code implementations • 27 Jul 2020 • Kenzo Lobos-Tsunekawa, Tatsuya Harada
Reinforcement Learning (RL), among other learning-based methods, represents powerful tools to solve complex robotic tasks (e. g., actuation, manipulation, navigation, etc.
1 code implementation • 13 Jul 2020 • Adrien Bitton, Philippe Esling, Tatsuya Harada
Although its definition is usually elusive, it can be seen from a signal processing viewpoint as all the spectral features that are perceived independently from pitch and loudness.
no code implementations • 4 Jul 2020 • Yang Li, Tianwei Zhang, Yoshihiko Nakamura, Tatsuya Harada
We present SplitFusion, a novel dense RGB-D SLAM framework that simultaneously performs tracking and dense reconstruction for both rigid and non-rigid components of the scene.
1 code implementation • ICLR 2021 • Ryohei Shimizu, Yusuke Mukuta, Tatsuya Harada
Hyperbolic spaces, which have the capacity to embed tree structures without distortion owing to their exponential volume growth, have recently been applied to machine learning to better capture the hierarchical nature of data.
1 code implementation • 26 May 2020 • Kazuma Kobayashi, Ryuichiro Hataya, Yusuke Kurose, Amina Bolatkan, Mototaka Miyake, Hirokazu Watanabe, Masamichi Takahashi, Jun Itami, Tatsuya Harada, Ryuji Hamamoto
In addition, we devise a metric to evaluate the anatomical fidelity of the reconstructed images and confirm that the overall detection performance is improved when the image reconstruction network achieves a higher score.
no code implementations • CVPR 2020 • Yang Li, Aljaž Božič, Tianwei Zhang, Yanli Ji, Tatsuya Harada, Matthias Nießner
One of the widespread solutions for non-rigid tracking has a nested-loop structure: with Gauss-Newton to minimize a tracking objective in the outer loop, and Preconditioned Conjugate Gradient (PCG) to solve a sparse linear system in the inner loop.
no code implementations • CVPR 2021 • Takuhiro Kaneko, Tatsuya Harada
However, in contrast to NR-GAN, to address irreversible characteristics, we introduce masking architectures adjusting degradation strength values in a data-driven manner using bypasses before and after degradation.
no code implementations • 6 Mar 2020 • Mikihiro Tanaka, Tatsuya Harada
In this study, we introduce a low cost method for generating descriptions from images containing novel objects.
no code implementations • 27 Feb 2020 • Kaikai Huang, Antonio Tejero-de-Pablos, Hiroaki Yamane, Yusuke Kurose, Junichi Iho, Youji Tokunaga, Makoto Horie, Keisuke Nishizawa, Yusaku Hayashi, Yasushi Koyama, Tatsuya Harada
In this paper, we propose a novel boundary detection method for coronary arteries that focuses on the continuity and connectivity of the boundaries.
no code implementations • 9 Jan 2020 • Takayuki Hara, Tatsuya Harada
We propose a method to generate spherical image from a single NFOV image, and control the degree of freedom of the generated regions using scene symmetry.
2 code implementations • CVPR 2020 • Takuhiro Kaneko, Tatsuya Harada
Therefore, we propose distribution and transformation constraints that encourage the noise generator to capture only the noise-specific components.
no code implementations • EMNLP (nlpbt) 2020 • Kohei Uehara, Tatsuya Harada
In the majority of the existing Visual Question Answering (VQA) research, the answers consist of short, often single words, as per instructions given to the annotators during dataset construction.
no code implementations • 20 Nov 2019 • Hiroharu Kato, Tatsuya Harada
We present a method to learn single-view reconstruction of the 3D shape, pose, and texture of objects from categorized natural images in a self-supervised manner.
2 code implementations • 18 Nov 2019 • Toshihiko Matsuura, Tatsuya Harada
Conventional methods assume that the domain to which each sample belongs is known in training.
Ranked #58 on
Domain Generalization
on PACS
no code implementations • ICCV 2019 • Shusuke Takahama, Yusuke Kurose, Yusuke Mukuta, Hiroyuki Abe, Masashi Fukayama, Akihiko Yoshizawa, Masanobu Kitagawa, Tatsuya Harada
If we consider the relationship of neighboring patches and global features, we can improve the classification performance.
no code implementations • 3 Oct 2019 • Dexuan Zhang, Tatsuya Harada
In this work, we present a novel upper bound of target error to address the problem for unsupervised domain adaptation.
no code implementations • WS 2019 • Yusuke Mori, Hiroaki Yamane, Yusuke Mukuta, Tatsuya Harada
In this study, we undertake the task of story ending generation.
no code implementations • 1 Oct 2019 • Akihiro Nakamura, Tatsuya Harada
With both tasks, we show that our method achieves higher accuracy than common few-shot learning algorithms.
2 code implementations • ICLR 2020 • Atsuhiro Noguchi, Tatsuya Harada
The loss is simple yet effective for any type of image generator such as DCGAN and StyleGAN to be conditioned on camera parameters.
no code implementations • 27 Sep 2019 • Kosuke Arase, Yusuke Mukuta, Tatsuya Harada
Certain existing studies have split input point clouds into small regions such as 1m x 1m; one reason for this is that models in the studies cannot consume a large number of points because of the large space complexity.
no code implementations • 25 Sep 2019 • Hiroaki Yamane, Chin-Yew Lin, Tatsuya Harada
To this end, we first used a crowdsourcing service to obtain sufficient data for a subjective agreement on numerical common sense.
no code implementations • 5 Jun 2019 • Yusuke Mukuta, Tatsuya Harada
Based on this result, a novel feature model that explicitly consider group action is proposed for principal component analysis and k-means clustering, which are commonly used in most feature coding methods, and global feature functions.
no code implementations • 5 Jun 2019 • Yusuke Mukuta, Tatsuaki Machida, Tatsuya Harada
Subsequently, we apply the proposed approximation to the polynomial corresponding to the matrix square root to obtain a compact approximation for the square root of the covariance feature.
no code implementations • ICLR 2020 • Wataru Kawai, Yusuke Mukuta, Tatsuya Harada
Graphs are ubiquitous real-world data structures, and generative models that approximate distributions over graphs and derive new samples from them have significant importance.
no code implementations • 21 May 2019 • Jen-Yen Chang, Antonio Tejero-de-Pablos, Tatsuya Harada
Gesture interaction is a natural way of communicating with a robot as an alternative to speech.
no code implementations • 7 May 2019 • Sho Maeoki, Kohei Uehara, Tatsuya Harada
We propose a system to retrieve videos by asking questions about the content of the videos and leveraging the user's responses to the questions.
no code implementations • 6 May 2019 • Takuhiro Kaneko, Tatsuya Harada
This problem is challenging in terms of scalability because it requires the learning of numerous mappings, the number of which increases proportional to the number of domains.
no code implementations • 16 Apr 2019 • Naoya Fushishita, Antonio Tejero-de-Pablos, Yusuke Mukuta, Tatsuya Harada
First, from an input human video, we generate sequences of future human poses (i. e., the image coordinates of their body-joints) via adversarial learning.
4 code implementations • ICCV 2019 • Atsuhiro Noguchi, Tatsuya Harada
To reduce the amount of data required, we propose a new method for transferring prior knowledge of the pre-trained generator, which is trained with a large dataset, to a small dataset in a different domain.
no code implementations • 25 Mar 2019 • Keisuke Hagiwara, Yusuke Mukuta, Tatsuya Harada
So far, research to generate captions from images has been carried out from the viewpoint that a caption holds sufficient information for an image.
no code implementations • 15 Mar 2019 • Yang Li, Yoshitaka Ushiku, Tatsuya Harada
In this paper, we propose to leverage graph optimization and loop closure detection to overcome limitations of unsupervised learning based monocular visual odometry.
no code implementations • 18 Dec 2018 • Toshihiko Matsuura, Kuniaki Saito, Tatsuya Harada
We utilize two classification networks to estimate the ratio of the target samples in each class with which a classification loss is weighted to adapt the classes present in the target domain.
2 code implementations • CVPR 2019 • Kuniaki Saito, Yoshitaka Ushiku, Tatsuya Harada, Kate Saenko
This motivates us to propose a novel method for detector adaptation based on strong local alignment and weak global alignment.
Ranked #2 on
Unsupervised Domain Adaptation
on SIM10K to BDD100K
1 code implementation • 11 Dec 2018 • Kohei Watanabe, Kuniaki Saito, Yoshitaka Ushiku, Tatsuya Harada
The other is a multitask learning approach that uses depth images as outputs.
no code implementations • CVPR 2019 • Atsushi Kanehira, Tatsuya Harada
This paper addresses the generation of explanations with visual examples.
no code implementations • 4 Dec 2018 • Shohei Yamamoto, Antonio Tejero-de-Pablos, Yoshitaka Ushiku, Tatsuya Harada
The results demonstrate that CFT-GAN is able to successfully generate videos containing the action and appearances indicated in the captions.
no code implementations • CVPR 2019 • Atsushi Kanehira, Kentaro Takemoto, Sho Inayoshi, Tatsuya Harada
This study addresses generating counterfactual explanations with multimodal information.
2 code implementations • ICCV 2019 • Mikihiro Tanaka, Takayuki Itamochi, Kenichi Narioka, Ikuro Sato, Yoshitaka Ushiku, Tatsuya Harada
Moreover, we regard that sentences that are easily understood are those that are comprehended correctly and quickly by humans.
3 code implementations • CVPR 2019 • Takuhiro Kaneko, Yoshitaka Ushiku, Tatsuya Harada
To remedy this, we propose a novel family of GANs called label-noise robust GANs (rGANs), which, by incorporating a noise transition model, can learn a clean label conditional generative distribution even when training labels are noisy.
2 code implementations • 27 Nov 2018 • Takuhiro Kaneko, Yoshitaka Ushiku, Tatsuya Harada
To overcome this limitation, we address a novel problem called class-distinct and class-mutual image generation, in which the goal is to construct a generator that can capture between-class relationships and generate an image selectively conditioned on the class specificity.
no code implementations • CVPR 2019 • Hiroharu Kato, Tatsuya Harada
The discriminator is trained to distinguish the reconstructed views of the observed viewpoints from those of the unobserved viewpoints.
1 code implementation • ECCV 2018 • Kohei Uehara, Antonio Tejero-de-Pablos, Yoshitaka Ushiku, Tatsuya Harada
In this paper, we propose a method for generating questions about unknown objects in an image, as means to get information about classes that have not been learned.
no code implementations • CVPR 2018 • Andrew Shin, Yoshitaka Ushiku, Tatsuya Harada
Image description task has been invariably examined in a static manner with qualitative presumptions held to be universally applicable, regardless of the scope or target of the description.
4 code implementations • ECCV 2018 • Kuniaki Saito, Shohei Yamamoto, Yoshitaka Ushiku, Tatsuya Harada
Almost all of them are proposed for a closed-set scenario, where the source and the target domain completely share the class of their samples.
no code implementations • CVPR 2018 • Atsushi Kanehira, Luc van Gool, Yoshitaka Ushiku, Tatsuya Harada
To satisfy these requirements (A)-(C) simultaneously, we proposed a novel video summarization method from multiple groups of videos.
no code implementations • IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2017 • Qishen Ha, Kohei Watanabe, Takumi Karasawa, Yoshitaka Ushiku, Tatsuya Harada
We benchmarked our method by creating an RGB-Thermal dataset in which thermal and RGB images are combined.
Ranked #4 on
Thermal Image Segmentation
on PST900
8 code implementations • CVPR 2018 • Kuniaki Saito, Kohei Watanabe, Yoshitaka Ushiku, Tatsuya Harada
To solve these problems, we introduce a new approach that attempts to align distributions of source and target by utilizing the task-specific decision boundaries.
Ranked #3 on
Domain Adaptation
on HMDBfull-to-UCF
Image Classification
Multi-Source Unsupervised Domain Adaptation
+2
5 code implementations • ICLR 2018 • Yuji Tokozume, Yoshitaka Ushiku, Tatsuya Harada
Deep learning methods have achieved high performance in sound recognition tasks.
3 code implementations • CVPR 2018 • Yuji Tokozume, Yoshitaka Ushiku, Tatsuya Harada
Second, we propose a mixing method that treats the images as waveforms, which leads to a further improvement in performance.
3 code implementations • 27 Nov 2017 • Katsunori Ohnishi, Shohei Yamamoto, Yoshitaka Ushiku, Tatsuya Harada
FlowGAN generates optical flow, which contains only the edge and motion of the videos to be begerated.
3 code implementations • CVPR 2018 • Hiroharu Kato, Yoshitaka Ushiku, Tatsuya Harada
Using this renderer, we perform single-image 3D mesh reconstruction with silhouette image supervision and our system outperforms the existing voxel-based approach.
Ranked #6 on
3D Object Reconstruction
on Data3D−R2N2
(Avg F1 metric)
no code implementations • ICLR 2018 • Kuniaki Saito, Yoshitaka Ushiku, Tatsuya Harada, Kate Saenko
However, a drawback of this approach is that the critic simply labels the generated features as in-domain or not, without considering the boundaries between classes.
Ranked #2 on
Synthetic-to-Real Translation
on Syn2Real-C
1 code implementation • 31 Oct 2017 • Andrew Shin, Leopold Crestel, Hiroharu Kato, Kuniaki Saito, Katsunori Ohnishi, Masataka Yamaguchi, Masahiro Nakawaki, Yoshitaka Ushiku, Tatsuya Harada
Automatic melody generation for pop music has been a long-time aspiration for both AI researchers and musicians.
Sound Multimedia Audio and Speech Processing
no code implementations • ICCV 2017 • Masataka Yamaguchi, Kuniaki Saito, Yoshitaka Ushiku, Tatsuya Harada
In this paper, we address the problem of spatio-temporal person retrieval from multiple videos using a natural language query, in which we output a tube (i. e., a sequence of bounding boxes) which encloses the person described by the query.
1 code implementation • ICML 2017 • Kuniaki Saito, Yoshitaka Ushiku, Tatsuya Harada
Deep-layered models trained on a large number of labeled samples boost the accuracy of many tasks.
Ranked #5 on
Sentiment Analysis
on Multi-Domain Sentiment Dataset
no code implementations • 7 Feb 2017 • Masatoshi Hidaka, Ken Miura, Tatsuya Harada
In the experiments, we demonstrate their practicality by training VGGNet in a distributed manner using web browsers as the client.
no code implementations • 23 Dec 2016 • Kuniaki Saito, Yusuke Mukuta, Yoshitaka Ushiku, Tatsuya Harada
To obtain the common representations under such a situation, we propose to make the distributions over different modalities similar in the learned representations, namely modality-invariant representations.
no code implementations • 21 Sep 2016 • Andrew Shin, Yoshitaka Ushiku, Tatsuya Harada
Visual Question Answering (VQA) task has showcased a new stage of interaction between language and vision, two of the most pivotal components of artificial intelligence.
no code implementations • 20 Jun 2016 • Kuniaki Saito, Andrew Shin, Yoshitaka Ushiku, Tatsuya Harada
Visual question answering (VQA) task not only bridges the gap between images and language, but also requires that specific contents within the image are understood as indicated by linguistic context of the question, in order to generate the accurate answers.
no code implementations • CVPR 2016 • Yusuke Mukuta, Tatsuya Harada
Our experiments show that the proposed method is better than the random features method and comparable with the Nystrom method in terms of the approximation error and classification accuracy.
no code implementations • CVPR 2016 • Atsushi Kanehira, Tatsuya Harada
Such a setting has been studied as a positive and unlabeled (PU) classification problem in a binary setting.
no code implementations • 18 May 2016 • Andrew Shin, Katsunori Ohnishi, Tatsuya Harada
Recent advances in image captioning task have led to increasing interests in video captioning task.
no code implementations • 29 Apr 2016 • Katsunori Ohnishi, Masatoshi Hidaka, Tatsuya Harada
This new descriptor is calculated by applying discriminative weights learned from one network to a convolutional layer of the other network.
Ranked #10 on
Action Classification
on Toyota Smarthome dataset
no code implementations • 30 Mar 2016 • Andrew Shin, Masataka Yamaguchi, Katsunori Ohnishi, Tatsuya Harada
The workflow of extracting features from images using convolutional neural networks (CNN) and generating captions with recurrent neural networks (RNN) has become a de-facto standard for image captioning task.
no code implementations • ICCV 2015 • Yoshitaka Ushiku, Masataka Yamaguchi, Yusuke Mukuta, Tatsuya Harada
In order to overcome the shortage of training samples, CoSMoS obtains a subspace in which (a) all feature vectors associated with the same phrase are mapped as mutually close, (b) classifiers for each phrase are learned, and (c) training samples are shared among co-occurring phrases.
no code implementations • CVPR 2016 • Katsunori Ohnishi, Atsushi Kanehira, Asako Kanezaki, Tatsuya Harada
We present a novel dataset and a novel algorithm for recognizing activities of daily living (ADL) from a first-person wearable camera.
no code implementations • 9 Nov 2015 • Hiroharu Kato, Tatsuya Harada
Additionally, a method to measure naturalness can be complementary to Convolutional Neural Network (CNN) based features, which are known to be insensitive to the naturalness of images.
no code implementations • CVPR 2014 • Hiroharu Kato, Tatsuya Harada
The objective of this work is to reconstruct an original image from Bag-of-Visual-Words (BoVW).
no code implementations • 19 Mar 2015 • Ken Miura, Tatsuya Harada
We have also developed a new JavaScript neural network framework called "Sukiyaki" that uses general purpose GPUs with web browsers.
no code implementations • 21 Feb 2015 • Ken Miura, Tetsuaki Mano, Atsushi Kanehira, Yuichiro Tsuchiya, Tatsuya Harada
Our core library offering a matrix calculation is called Sushi, which exhibits far better performance than any other leading machine learning libraries written in JavaScript.
no code implementations • CVPR 2014 • Yoshitaka Ushiku, Masatoshi Hidaka, Tatsuya Harada
In this paper, we would like to evaluate online learning algorithms for large-scale visual recognition using state-of-the-art features which are preselected and held fixed.
no code implementations • NeurIPS 2012 • Tatsuya Harada, Yasuo Kuniyoshi
This paper proposes a novel image representation called a Graphical Gaussian Vector, which is a counterpart of the codebook and local feature matching approaches.