no code implementations • 4 Jan 2025 • Abbaas Alif Mohamed Nishar, Shrinivas Kudekar, Bernard Kintzing, Ashwin Ashok
We present `Revelio', a real-world screen-camera communication system leveraging temporal flicker fusion in the OKLAB color space.
no code implementations • 15 Nov 2024 • Faith Johnson, Bryan Bo Cao, Ashwin Ashok, Shubham Jain, Kristin Dana
Key to our approach is a memory proxy map (MPM), an intermediate representation of the environment learned in a self-supervised manner by the high-level manager agent that serves as a simplified memory, approximating what the agent has seen.
no code implementations • 22 Feb 2024 • Faith Johnson, Bryan Bo Cao, Kristin Dana, Shubham Jain, Ashwin Ashok
However, recent advancements in the visual navigation field face challenges due to the lack of human datasets in the real world for efficient supervised representation learning of the environments.
no code implementations • 19 Feb 2024 • Faith Johnson, Bryan Bo Cao, Ashwin Ashok, Shubham Jain, Kristin Dana
We introduce a new approach to visual navigation using feudal learning, which employs a hierarchical structure consisting of a worker agent, a mid-level manager, and a high-level manager.
1 code implementation • MobiCom ISACom 2023 • Bryan Bo Cao, Abrar Alali, Hansi Liu, Nicholas Meegan, Marco Gruteser, Kristin Dana, Ashwin Ashok, Shubham Jain
Tracking subjects in videos is one of the most widely used functions in camera-based IoT applications such as security surveillance, smart city traffic safety enhancement, vehicle to pedestrian communication and so on.
no code implementations • 11 Oct 2022 • Nicholas Meegan, Hansi Liu, Bryan Cao, Abrar Alali, Kristin Dana, Marco Gruteser, Shubham Jain, Ashwin Ashok
We introduce ViFiCon, a self-supervised contrastive learning scheme which uses synchronized information across vision and wireless modalities to perform cross-modal association.
1 code implementation • IEEE International Conference on Sensing, Communication, and Networking 2022 • Bryan Bo Cao, Abrar Alali, Hansi Liu, Nicholas Meegan, Marco Gruteser, Kristin Dana, Ashwin Ashok, Shubham Jain
ViTag associates a sequence of vision tracker generated bounding boxes with Inertial Measurement Unit (IMU) data and Wi-Fi Fine Time Measurements (FTM) from smartphones.
1 code implementation • ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN) 2022 • Hansi Liu, Abrar Alali, Mohamed Ibrahim, Bryan Bo Cao, Nicholas Meegan, Hongyu Li, Marco Gruteser, Shubham Jain, Kristin Dana, Ashwin Ashok, Bin Cheng, HongSheng Lu
In this paper, we present Vi-Fi, a multi-modal system that leverages a user’s smartphone WiFi Fine Timing Measurements (FTM) and inertial measurement unit (IMU) sensor data to associate the user detected on a camera footage with their corresponding smartphone identifier (e. g. WiFi MAC address).
no code implementations • 11 May 2021 • Vu Tran, Gihan Jayatilaka, Ashwin Ashok, Archan Misra
We show that a fully functional DeepLight system is able to robustly achieve high decoding accuracy (frame error rate < 0. 2) and moderately-high data goodput (>=0. 95Kbps) using a human-held smartphone camera, even over larger screen-camera distances (approx =2m).
1 code implementation • 22 Nov 2019 • Amey Parundekar, Susan Elias, Ashwin Ashok
Furthermore, we intend to create this data set not only for classification of the news but also to find patterns that reason the intent behind misinformation.
no code implementations • 8 Jan 2015 • Wenjia Yuan, Eric Wengrowski, Kristin J. Dana, Ashwin Ashok, Marco Gruteser, Narayan Mandayam
We present a novel method for communicating between a camera and display by embedding and recovering hidden and dynamic information within a displayed image.