Search Results for author: Yutaro Yamada

Found 12 papers, 6 papers with code

L3GO: Language Agents with Chain-of-3D-Thoughts for Generating Unconventional Objects

no code implementations14 Feb 2024 Yutaro Yamada, Khyathi Chandu, YuChen Lin, Jack Hessel, Ilker Yildirim, Yejin Choi

In this paper, we propose a language agent with chain-of-3D-thoughts (L3GO), an inference-time approach that can reason about part-based 3D mesh generation of unconventional objects that current data-driven diffusion models struggle with.

Image Generation Text to 3D

Evaluating Spatial Understanding of Large Language Models

1 code implementation23 Oct 2023 Yutaro Yamada, Yihan Bao, Andrew K. Lampinen, Jungo Kasai, Ilker Yildirim

Large language models (LLMs) show remarkable capabilities across a variety of tasks.

Evaluating GPT-4 and ChatGPT on Japanese Medical Licensing Examinations

1 code implementation31 Mar 2023 Jungo Kasai, Yuhei Kasai, Keisuke Sakaguchi, Yutaro Yamada, Dragomir Radev

In this work, we evaluate LLM APIs (ChatGPT, GPT-3, and GPT-4) on the Japanese national medical licensing examinations from the past five years, including the current year.

When are Lemons Purple? The Concept Association Bias of Vision-Language Models

no code implementations22 Dec 2022 Yutaro Yamada, Yingtian Tang, Yoyo Zhang, Ilker Yildirim

Large-scale vision-language models such as CLIP have shown impressive performance on zero-shot image classification and image-to-text retrieval.

Attribute Image Classification +7

Does Robustness on ImageNet Transfer to Downstream Tasks?

no code implementations CVPR 2022 Yutaro Yamada, Mayu Otani

For object detection and semantic segmentation, we find that a vanilla Swin Transformer, a variant of Vision Transformer tailored for dense prediction tasks, transfers robustness better than Convolutional Neural Networks that are trained to be robust to the corrupted version of ImageNet.

Classification Image Classification +5

Can Wikipedia Help Offline Reinforcement Learning?

1 code implementation28 Jan 2022 Machel Reid, Yutaro Yamada, Shixiang Shane Gu

In this paper, we look to take advantage of this formulation of reinforcement learning as sequence modeling and investigate the transferability of pre-trained sequence models on other domains (vision, language) when finetuned on offline RL tasks (control, games).

Offline RL reinforcement-learning +1

Support Recovery with Stochastic Gates: Theory and Application for Linear Models

1 code implementation29 Oct 2021 Soham Jana, Henry Li, Yutaro Yamada, Ofir Lindenbaum

Consider the problem of simultaneous estimation and support recovery of the coefficient vector in a linear data model with additive Gaussian noise.

Exploiting 3D Shape Bias towards Robust Vision

no code implementations NeurIPS Workshop SVRHM 2021 Yutaro Yamada, Yuval Kluger, Sahand Negahban, Ilker Yildirim

To tackle the problem from a new perspective, we encourage closer collaboration between the robustness and 3D vision communities.

3D Reconstruction

Geon3D: Exploiting 3D Shape Bias towards Building Robust Machine Vision

no code implementations29 Sep 2021 Yutaro Yamada, Yuval Kluger, Sahand Negahban, Ilker Yildirim

To tackle the problem from a new perspective, we encourage closer collaboration between the robustness and 3D vision communities.

3D Reconstruction

Feature Selection using Stochastic Gates

1 code implementation ICML 2020 Yutaro Yamada, Ofir Lindenbaum, Sahand Negahban, Yuval Kluger

Feature selection problems have been extensively studied for linear estimation, for instance, Lasso, but less emphasis has been placed on feature selection for non-linear functions.

feature selection

Defending against Adversarial Images using Basis Functions Transformations

1 code implementation28 Mar 2018 Uri Shaham, James Garritano, Yutaro Yamada, Ethan Weinberger, Alex Cloninger, Xiuyuan Cheng, Kelly Stanton, Yuval Kluger

We study the effectiveness of various approaches that defend against adversarial attacks on deep networks via manipulations based on basis function representations of images.

Understanding Adversarial Training: Increasing Local Stability of Neural Nets through Robust Optimization

no code implementations17 Nov 2015 Uri Shaham, Yutaro Yamada, Sahand Negahban

We propose a general framework for increasing local stability of Artificial Neural Nets (ANNs) using Robust Optimization (RO).

Cannot find the paper you are looking for? You can Submit a new open access paper.