no code implementations • 10 Apr 2023 • Qiao Jin, Andrew Shin, Zhiyong Lu
On all queries, LADER can improve the performance of a dense retriever by 24%-37% relative NDCG@10 while not requiring additional training, and further performance improvement is expected from more logs.
no code implementations • 6 Mar 2021 • Andrew Shin, Masato Ishii, Takuya Narihira
Transformer architectures have brought about fundamental changes to computational linguistic field, which had been dominated by recurrent neural networks for many years.
1 code implementation • 12 Feb 2021 • Takuya Narihira, Javier Alonsogarcia, Fabien Cardinaux, Akio Hayakawa, Masato Ishii, Kazunori Iwaki, Thomas Kemp, Yoshiyuki Kobayashi, Lukas Mauch, Akira Nakamura, Yukio Obuchi, Andrew Shin, Kenji Suzuki, Stephen Tiedmann, Stefan Uhlich, Takuya Yashima, Kazuki Yoshiyama
While there exist a plethora of deep learning tools and frameworks, the fast-growing complexity of the field brings new demands and challenges, such as more flexible network design, speedy computation on distributed setting, and compatibility between different tools.
no code implementations • 25 Nov 2020 • Naofumi Akimoto, Akio Hayakawa, Andrew Shin, Takuya Narihira
To address this issue, we warp colors only from the regions on the reference frame restricted by correspondence in time.
no code implementations • CVPR 2018 • Andrew Shin, Yoshitaka Ushiku, Tatsuya Harada
Image description task has been invariably examined in a static manner with qualitative presumptions held to be universally applicable, regardless of the scope or target of the description.
1 code implementation • 31 Oct 2017 • Andrew Shin, Leopold Crestel, Hiroharu Kato, Kuniaki Saito, Katsunori Ohnishi, Masataka Yamaguchi, Masahiro Nakawaki, Yoshitaka Ushiku, Tatsuya Harada
Automatic melody generation for pop music has been a long-time aspiration for both AI researchers and musicians.
Sound Multimedia Audio and Speech Processing
no code implementations • 21 Sep 2016 • Andrew Shin, Yoshitaka Ushiku, Tatsuya Harada
Visual Question Answering (VQA) task has showcased a new stage of interaction between language and vision, two of the most pivotal components of artificial intelligence.
no code implementations • 20 Jun 2016 • Kuniaki Saito, Andrew Shin, Yoshitaka Ushiku, Tatsuya Harada
Visual question answering (VQA) task not only bridges the gap between images and language, but also requires that specific contents within the image are understood as indicated by linguistic context of the question, in order to generate the accurate answers.
no code implementations • 18 May 2016 • Andrew Shin, Katsunori Ohnishi, Tatsuya Harada
Recent advances in image captioning task have led to increasing interests in video captioning task.
no code implementations • 30 Mar 2016 • Andrew Shin, Masataka Yamaguchi, Katsunori Ohnishi, Tatsuya Harada
The workflow of extracting features from images using convolutional neural networks (CNN) and generating captions with recurrent neural networks (RNN) has become a de-facto standard for image captioning task.