1 code implementation • 17 Jul 2024 • Hyun Seok Seong, WonJun Moon, SuBeen Lee, Jae-Pil Heo
Then, considering the distribution of positive samples, we relocate the proxy anchor towards areas with a higher concentration of positives and adjust the positiveness boundary based on the propagation degree of the proxy anchor.
1 code implementation • 16 Jul 2024 • Gilhan Park, WonJun Moon, SuBeen Lee, Tae-Young Kim, Jae-Pil Heo
Additionally, in the case of the second approach, initializing the new class classifier with background knowledge triggers a similar background shift issue, but towards the new classes.
Ranked #1 on
Overlapped 5-3
on PASCAL VOC 2012
1 code implementation • 27 Dec 2023 • Seunggu Kang, WonJun Moon, Euiyeon Kim, Jae-Pil Heo
Zero-Shot Object Counting (ZSOC) aims to count referred instances of arbitrary classes in a query image without human-annotated exemplars.
Ranked #5 on
Object Counting
on CARPK
2 code implementations • 15 Nov 2023 • WonJun Moon, Sangeek Hyun, SuBeen Lee, Jae-Pil Heo
Dummy tokens conditioned by text query take portions of the attention weights, preventing irrelevant video clips from being represented by the text query.
Ranked #3 on
Highlight Detection
on TvSum
1 code implementation • 28 Jul 2023 • SuBeen Lee, WonJun Moon, Hyun Seok Seong, Jae-Pil Heo
While TDM influences high-level feature maps by task-adaptive calibration of channel-wise importance, we further introduce Instance Attention Module (IAM) operating in intermediate layers of feature extractors to instance-wisely highlight object-relevant channels, by extending QAM.
1 code implementation • CVPR 2023 • Hyun Seok Seong, WonJun Moon, SuBeen Lee, Jae-Pil Heo
Specifically, we add the loss propagating to local hidden positives, semantically similar nearby patches, in proportion to the predefined similarity scores.
Ranked #3 on
Unsupervised Semantic Segmentation
on Potsdam-3
1 code implementation • CVPR 2023 • WonJun Moon, Sangeek Hyun, Sanguk Park, Dongchan Park, Jae-Pil Heo
As we observe the insignificant role of a given query in transformer architectures, our encoding module starts with cross-attention layers to explicitly inject the context of text query into video representation.
Ranked #3 on
Video Grounding
on QVHighlights
1 code implementation • 24 Nov 2022 • WonJun Moon, Hyun Seok Seong, Jae-Pil Heo
A dramatic increase in real-world video volume with extremely diverse and emerging topics naturally forms a long-tailed video distribution in terms of their categories, and it spotlights the need for Video Long-Tailed Recognition (VLTR).
1 code implementation • 20 Jul 2022 • WonJun Moon, Junho Park, Hyun Seok Seong, Cheol-Ho Cho, Jae-Pil Heo
Furthermore, moderate- and easy-difficulty samples are also yielded by our modified GAN and Copycat, respectively.
1 code implementation • 20 Jul 2022 • WonJun Moon, Ji-Hwan Kim, Jae-Pil Heo
Our exhaustive experiments validate the merits of LoRot as a pretext task tailored for supervised learning in terms of robustness and generalization capability.
Ranked #9 on
Data Augmentation
on ImageNet
1 code implementation • CVPR 2022 • SuBeen Lee, WonJun Moon, Jae-Pil Heo
Specifically, TDM learns task-specific channel weights based on two novel components: Support Attention Module (SAM) and Query Attention Module (QAM).
Ranked #11 on
Few-Shot Image Classification
on CUB 200 5-way 5-shot
(using extra training data)