no code implementations • 1 May 2023 • Wenbin He, Suphanut Jamonnak, Liang Gou, Liu Ren
To further improve the pixel embeddings and enable language-driven semantic segmentation, we design two types of consistency guided by vision-language models: 1) embedding consistency, aligning our pixel embeddings to the joint feature space of a pre-trained vision-language model, CLIP; and 2) semantic consistency, forcing our model to make the same predictions as CLIP over a set of carefully designed target classes with both known and unknown prototypes.
no code implementations • CVPR 2023 • Wenbin He, Suphanut Jamonnak, Liang Gou, Liu Ren
To further improve the pixel embeddings and enable language-driven semantic segmentation, we design two types of consistency guided by vision-language models: 1) embedding consistency, aligning our pixel embeddings to the joint feature space of a pre-trained vision-language model, CLIP; and 2) semantic consistency, forcing our model to make the same predictions as CLIP over a set of carefully designed target classes with both known and unknown prototypes.
no code implementations • 20 Aug 2021 • Suphanut Jamonnak, Ye Zhao, Xinyi Huang, Md Amiruzzaman
The visual study is seamlessly integrated with the geographical environment by combining DL model performance with geospatial visualization techniques.
no code implementations • 3 Sep 2020 • Xinyi Huang, Suphanut Jamonnak, Ye Zhao, Boyu Wang, Minh Hoai, Kevin Yager, Wei Xu
Existing interactive visualization tools for deep learning are mostly applied to the training, debugging, and refinement of neural network models working on natural images.
no code implementations • 10 Oct 2019 • Xinyi Huang, Suphanut Jamonnak, Ye Zhao, Boyu Wang, Minh Hoai, Kevin Yager, Wei Xu
This extended abstract presents a visualization system, which is designed for domain scientists to visually understand their deep learning model of extracting multiple attributes in x-ray scattering images.