no code implementations • Findings (EMNLP) 2021 • Zhan Shi, Hui Liu, Martin Renqiang Min, Christopher Malon, Li Erran Li, Xiaodan Zhu
Image captioning systems are expected to have the ability to combine individual concepts when describing scenes with concept combinations that are not observed during training.
no code implementations • 3 Jan 2025 • Zhan Shi, Shanglin Yang
Fashion, deeply rooted in sociocultural dynamics, evolves as individuals emulate styles popularized by influencers and iconic figures.
no code implementations • 30 Dec 2024 • Subramaniam Vincent, Phoebe Wang, Zhan Shi, Sahas Koka, Yi Fang
Since the launch of ChatGPT in late 2022, the capacities of Large Language Models and their evaluation have been in constant discussion and evaluation both in academic research and in the industry.
no code implementations • 10 Sep 2024 • Tao Ni, Xin Zhan, Tao Luo, wenbin liu, Zhan Shi, Junbo Chen
The integration of multiple data sources in road segmentation presents both opportunities and challenges.
no code implementations • 13 Aug 2024 • ZiChao Dong, Yilin Zhang, Xufeng Huang, Hang Ji, Zhan Shi, Xin Zhan, Junbo Chen
Unfortunately, single RGBD dataset with thousands of data is not enough for training an discriminating filter for visual texture feature extraction.
Ranked #4 on
3D Object Detection
on ScanNetV2
1 code implementation • 25 Jul 2024 • Ruijie Tao, Zhan Shi, Yidi Jiang, Duc-Tuan Truong, Eng-Siong Chng, Massimo Alioto, Haizhou Li
In this paper, we propose a ``Multi-stage Face-voice Association Learning with Keynote Speaker Diarization''~(MFV-KSD) framework.
no code implementations • 20 Jun 2024 • Zhongshen Zeng, Yinhong Liu, Yingjia Wan, Jingyao Li, Pengguang Chen, Jianbo Dai, Yuxuan Yao, Rongwu Xu, Zehan Qi, Wanru Zhao, Linling Shen, Jianqiao Lu, Haochen Tan, Yukang Chen, Hao Zhang, Zhan Shi, Bailin Wang, Zhijiang Guo, Jiaya Jia
Large language models (LLMs) have shown increasing capability in problem-solving and decision-making, largely based on the step-by-step chain-of-thought reasoning processes.
no code implementations • 7 May 2024 • Chunlin Tian, Zhan Shi, Xinpeng Qin, Li Li, Chengzhong Xu
Federated Learning (FL) enables multiple devices to collaboratively train a shared model while ensuring data privacy.
1 code implementation • 30 Apr 2024 • Chunlin Tian, Zhan Shi, Zhijiang Guo, Li Li, Chengzhong Xu
Through a series of experiments, we have uncovered two critical insights that shed light on the training and parameter inefficiency of LoRA.
no code implementations • 1 Apr 2024 • Ruijie Tao, Zhan Shi, Yidi Jiang, Tianchi Liu, Haizhou Li
Our experimental results on three created datasets demonstrated that VCA-NN effectively mitigates these dataset problems, which provides a new direction for handling the speaker recognition problems from the data aspect.
no code implementations • 26 Mar 2024 • Zhan Shi, Jingwei Zhang, Jun Kong, Fusheng Wang
In digital pathology, the multiple instance learning (MIL) strategy is widely used in the weakly supervised histopathology whole slide image (WSI) classification task where giga-pixel WSIs are only labeled at the slide level.
1 code implementation • 26 Jan 2024 • Haochen Tan, Zhijiang Guo, Zhan Shi, Lu Xu, Zhili Liu, Yunlong Feng, Xiaoguang Li, Yasheng Wang, Lifeng Shang, Qun Liu, Linqi Song
Large Language Models (LLMs) have succeeded remarkably in understanding long-form contents.
1 code implementation • 37th Conference on Neural Information Processing Systems (NeurIPS 2023) 2023 • Fangchen Yu, Runze Zhao, Zhan Shi, Yiwen Lu, Jicong Fan, Yicheng Zeng, Jianfeng Mao, Wenye Li
Secondly, we develop a series of affinity learning methods that equip the selfexpressive framework with ℓp-norm to construct an intrinsic affinity matrix with an adaptive extension.
1 code implementation • ICCV 2023 • Tao Lv, Hao Ye, Quan Yuan, Zhan Shi, Yibo Wang, Shuming Wang, Xun Cao
We demonstrate a compact, cost-effective snapshot spectral imaging system named Aperture Diffraction Imaging Spectrometer (ADIS), which consists only of an imaging lens with an ultra-thin orthogonal aperture mask and a mosaic filter sensor, requiring no additional physical footprint compared to common RGB cameras.
1 code implementation • 13 Jul 2023 • Zhan Shi, Xin Ding, Peng Ding, Chun Yang, Ru Huang, Xiaoxuan Song
Four tiny SOAP models are also created by replacing the convolutional blocks in Mobile-SOAP with four small-scale networks, respectively.
1 code implementation • 28 Oct 2022 • Ruijie Tao, Kong Aik Lee, Zhan Shi, Haizhou Li
However, noisy samples (i. e., with wrong labels) in the training set induce confusion and cause the network to learn the incorrect representation.
no code implementations • 26 Oct 2022 • Mohamed Ashraf Abdelsalam, Zhan Shi, Federico Fancellu, Kalliopi Basioti, Dhaivat J. Bhatt, Vladimir Pavlovic, Afsaneh Fazly
The success of scene graphs for visual scene understanding has brought attention to the benefits of abstracting a visual input (e. g., image) into a structured representation, where entities (people and objects) are nodes connected by edges specifying their relations.
1 code implementation • 31 Mar 2022 • Kevin Miller, John Mauro, Jason Setiadi, Xoaquin Baca, Zhan Shi, Jeff Calder, Andrea L. Bertozzi
We use a Convolutional Neural Network Variational Autoencoder (CNNVAE) to embed SAR data into a feature space, and then construct a similarity graph from the embedded data and apply graph-based semi-supervised learning techniques.
no code implementations • 8 Sep 2021 • Xiaoyu Yang, Xiaodan Zhu, Zhan Shi, Tianda Li
There have been two lines of approaches that can be used to further address the limitation: (1) unsupervised pretraining can leverage knowledge in much larger unstructured text data; (2) structured (often human-curated) knowledge has started to be considered in neural-network-based models for NLI.
1 code implementation • EMNLP 2021 • Hui Liu, Zhan Shi, Xiaodan Zhu
For the message-pair classifier, we enrich its training data by retrieving message pairs with high confidence from the disentangled sessions predicted by the session classifier.
1 code implementation • ACL 2021 • Zhan Shi, Hui Liu, Xiaodan Zhu
In this paper we propose a novel approach to encourage captioning models to produce more detailed captions using natural language inference, based on the motivation that, among different captions of an image, descriptive captions are more likely to entail less descriptive captions.
no code implementations • 3 Dec 2020 • Elie Aïdékon, Yueyun Hu, Zhan Shi
The classical Ray-Knight theorems for Brownian motion determine the law of its local time process either at the first hitting time of a given value a by the local time at the origin, or at the first hitting time of a given position b by Brownian motion.
Probability
1 code implementation • NeurIPS 2020 • Hongwei Jin, Zhan Shi, Venkata Jaya Shankar Ashish Peruri, Xinhua Zhang
Graph convolution networks (GCNs) have become effective models for graph classification.
no code implementations • 5 Oct 2020 • Zhan Shi, Chirag Sakhuja, Milad Hashemi, Kevin Swersky, Calvin Lin
The use of deep learning has grown at an exponential rate, giving rise to numerous specialized hardware and software systems for deep learning.
no code implementations • ACL 2020 • Zhan Shi, Xu Zhou, Xipeng Qiu, Xiaodan Zhu
Image captioning is a multimodal problem that has drawn extensive attention in both the natural language processing and computer vision community.
1 code implementation • 21 Jun 2020 • Zhan Shi, Xu Zhou, Xipeng Qiu, Xiaodan Zhu
Image captioning is a multimodal problem that has drawn extensive attention in both the natural language processing and computer vision community.
no code implementations • 11 Feb 2020 • Zac Cranko, Zhan Shi, Xinhua Zhang, Richard Nock, Simon Kornblith
The problem of adversarial examples has highlighted the need for a theory of regularisation that is general enough to apply to exotic function classes, such as universal approximators.
no code implementations • 25 Sep 2019 • Zac Cranko, Zhan Shi, Xinhua Zhang, Simon Kornblith, Richard Nock
Distributional robust risk (DRR) minimisation has arisen as a flexible and effective framework for machine learning.
no code implementations • ICLR 2020 • Zhan Shi, Kevin Swersky, Daniel Tarlow, Parthasarathy Ranganathan, Milad Hashemi
In this work, we propose a new approach to use GNNs to learn fused representations of general source code and its execution.
no code implementations • 22 Apr 2019 • Yu-Ping Ruan, Xiaodan Zhu, Zhen-Hua Ling, Zhan Shi, Quan Liu, Si Wei
Winograd Schema Challenge (WSC) was proposed as an AI-hard problem in testing computers' intelligence on common sense representation and reasoning.
no code implementations • 4 Sep 2018 • Zac Cranko, Simon Kornblith, Zhan Shi, Richard Nock
Robust risk minimisation has several advantages: it has been studied with regards to improving the generalisation properties of models and robustness to adversarial perturbation.
no code implementations • ICML 2018 • Vignesh Ganapathiraman, Zhan Shi, Xinhua Zhang, Yao-Liang Yu
Latent prediction models, exemplified by multi-layer networks, employ hidden variables that automate abstract feature discovery.
no code implementations • 8 Jun 2018 • Zac Cranko, Aditya Krishna Menon, Richard Nock, Cheng Soon Ong, Zhan Shi, Christian Walder
A key feature of our result is that it holds for all proper losses, and for a popular subset of these, the optimisation of this central measure appears to be independent of the loss.
3 code implementations • 30 Apr 2018 • Zhan Shi, Xinchi Chen, Xipeng Qiu, Xuanjing Huang
Similar to the adversarial models, the reward and policy function in IRL are optimized alternately.
no code implementations • NeurIPS 2017 • Zhan Shi, Xinhua Zhang, Yao-Liang Yu
Adversarial machines, where a learner competes against an adversary, have regained much recent interest in machine learning.
no code implementations • 2 Jul 2017 • Xinchi Chen, Zhan Shi, Xipeng Qiu, Xuanjing Huang
In this paper, we propose a new neural model to incorporate the word-level information for Chinese word segmentation.
no code implementations • ACL 2017 • Xinchi Chen, Zhan Shi, Xipeng Qiu, Xuanjing Huang
Different linguistic perspectives causes many diverse segmentation criteria for Chinese word segmentation (CWS).