Search Results for author: Michael S Ryoo

Found 6 papers, 2 papers with code

LAM SIMULATOR: Advancing Data Generation for Large Action Model Training via Online Exploration and Trajectory Feedback

no code implementations2 Jun 2025 Thai Hoang, Kung-Hsiang Huang, Shirley Kokane, JianGuo Zhang, Zuxin Liu, Ming Zhu, Jake Grigsby, Tian Lan, Michael S Ryoo, Chien-Sheng Wu, Shelby Heinecke, Huan Wang, Silvio Savarese, Caiming Xiong, Juan Carlos Niebles

Large Action Models (LAMs) for AI Agents offer incredible potential but face challenges due to the need for high-quality training data, especially for multi-steps tasks that involve planning, executing tool calls, and responding to feedback.

Large Language Model

Pixel Motion as Universal Representation for Robot Control

no code implementations12 May 2025 Kanchana Ranasinghe, Xiang Li, Cristina Mata, Jongwoo Park, Michael S Ryoo

We present LangToMo, a vision-language-action framework structured as a dual-system architecture that uses pixel motion forecasts as intermediate representations.

Vision-Language-Action

LatentCRF: Continuous CRF for Efficient Latent Diffusion

no code implementations24 Dec 2024 Kanchana Ranasinghe, Sadeep Jayasumana, Andreas Veit, Ayan Chakrabarti, Daniel Glasner, Michael S Ryoo, Srikumar Ramalingam, Sanjiv Kumar

Latent Diffusion Models (LDMs) produce high-quality, photo-realistic images, however, the latency incurred by multiple costly inference iterations can restrict their applicability.

Diversity

Hybrid Random Features

1 code implementation ICLR 2022 Krzysztof Choromanski, Haoxian Chen, Han Lin, Yuanzhe Ma, Arijit Sehanobish, Deepali Jain, Michael S Ryoo, Jake Varley, Andy Zeng, Valerii Likhosherstov, Dmitry Kalashnikov, Vikas Sindhwani, Adrian Weller

We propose a new class of random feature methods for linearizing softmax and Gaussian kernels called hybrid random features (HRFs) that automatically adapt the quality of kernel estimation to provide most accurate approximation in the defined regions of interest.

Benchmarking

Vi-MIX FOR SELF-SUPERVISED VIDEO REPRESENTATION

no code implementations29 Sep 2021 Srijan Das, Michael S Ryoo

We find that our video mixing strategy: Vi-Mix, i. e. preliminary mixing of videos followed by CMMC across different modalities in a video, improves the qual- ity of learned video representations.

Action Recognition Representation Learning +3

Cannot find the paper you are looking for? You can Submit a new open access paper.