Search Results for author: Joo Hwee Lim

Found 17 papers, 6 papers with code

An Overview of Challenges in Egocentric Text-Video Retrieval

no code implementations7 Jun 2023 Burak Satar, Hongyuan Zhu, Hanwang Zhang, Joo Hwee Lim

Text-video retrieval contains various challenges, including biases coming from diverse sources.

Retrieval Video Retrieval

Is Bio-Inspired Learning Better than Backprop? Benchmarking Bio Learning vs. Backprop

no code implementations9 Dec 2022 Manas Gupta, Sarthak Ketanbhai Modi, Hang Zhang, Joon Hei Lee, Joo Hwee Lim

Four of the five Bio-algorithms tested outperform BP by upto 5% accuracy when only 20% of the training dataset is available.

Benchmarking

Portmanteauing Features for Scene Text Recognition

no code implementations9 Nov 2022 Yew Lee Tan, Ernest Yu Kai Chew, Adams Wai-Kin Kong, Jung-jae Kim, Joo Hwee Lim

To generate the portmanteau feature, a non-linear input pipeline with a block matrix initialization is presented.

Scene Text Recognition

Semantic Role Aware Correlation Transformer for Text to Video Retrieval

1 code implementation26 Jun 2022 Burak Satar, Hongyuan Zhu, Xavier Bresson, Joo Hwee Lim

With the emergence of social media, voluminous video clips are uploaded every day, and retrieving the most relevant visual content with a language query becomes critical.

Retrieval Text to Video Retrieval +1

RoME: Role-aware Mixture-of-Expert Transformer for Text-to-Video Retrieval

1 code implementation26 Jun 2022 Burak Satar, Hongyuan Zhu, Hanwang Zhang, Joo Hwee Lim

Most methods consider only one joint embedding space between global visual and textual features without considering the local structures of each modality.

Retrieval Text to Video Retrieval +1

FashionSearchNet-v2: Learning Attribute Representations with Localization for Image Retrieval with Attribute Manipulation

no code implementations28 Nov 2021 Kenan E. Ak, Joo Hwee Lim, Ying Sun, Jo Yew Tham, Ashraf A. Kassim

A key challenge in e-commerce is that images have multiple attributes where users would like to manipulate and it is important to estimate discriminative feature representations for each of these attributes.

Attribute Image Retrieval +1

Joint Learning On The Hierarchy Representation for Fine-Grained Human Action Recognition

no code implementations12 Oct 2021 Mei Chee Leong, Hui Li Tan, Haosong Zhang, Liyuan Li, Feng Lin, Joo Hwee Lim

Inspired by the recently proposed hierarchy representation of fine-grained actions in FineGym and SlowFast network for action recognition, we propose a novel multi-task network which exploits the FineGym hierarchy representation to achieve effective joint learning and prediction for fine-grained human action recognition.

Action Recognition Multi-Task Learning +1

Prototype Recalls for Continual Learning

no code implementations25 Sep 2019 Mengmi Zhang, Tao Wang, Joo Hwee Lim, Jiashi Feng

Without tampering with the performance on initial tasks, our method learns novel concepts given a few training examples of each class in new tasks.

Continual Learning Metric Learning +1

Variational Prototype Replays for Continual Learning

1 code implementation23 May 2019 Mengmi Zhang, Tao Wang, Joo Hwee Lim, Gabriel Kreiman, Jiashi Feng

In each classification task, our method learns a set of variational prototypes with their means and variances, where embedding of the samples from the same class can be represented in a prototypical distribution and class-representative prototypes are separated apart.

Continual Learning General Classification +2

Egocentric Spatial Memory

1 code implementation31 Jul 2018 Mengmi Zhang, Keng Teck Ma, Shih-Cheng Yen, Joo Hwee Lim, Qi Zhao, Jiashi Feng

Egocentric spatial memory (ESM) defines a memory system with encoding, storing, recognizing and recalling the spatial information about the environment from an egocentric perspective.

Feature Engineering

Egocentric Spatial Memory Network

no code implementations ICLR 2018 Mengmi Zhang, Keng Teck Ma, Joo Hwee Lim, Shih-Cheng Yen, Qi Zhao, Jiashi Feng

During the exploration, our proposed ESM network model updates belief of the global map based on local observations using a recurrent neural network.

Navigate Simultaneous Localization and Mapping

Deep Future Gaze: Gaze Anticipation on Egocentric Videos Using Adversarial Networks

1 code implementation CVPR 2017 Mengmi Zhang, Keng Teck Ma, Joo Hwee Lim, Qi Zhao, Jiashi Feng

Through competition with discriminator, the generator progressively improves quality of the future frames and thus anticipates future gaze better.

Gaze Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.