no code implementations • NLP4ConvAI (ACL) 2022 • Tong Zhang, Yong liu, Boyang Li, Peixiang Zhong, Chen Zhang, Hao Wang, Chunyan Miao
Conversational Recommendation Systems recommend items through language based interactions with users. In order to generate naturalistic conversations and effectively utilize knowledge graphs (KGs) containing background information, we propose a novel Bag-of-Entities loss, which encourages the generated utterances to mention concepts related to the item being recommended, such as the genre or director of a movie.
no code implementations • CCL 2021 • Hao Wang, Junhui Li, ZhengXian Gong
“在汉语等其他有省略代词习惯的语言中, 通常会删掉可从上下文信息推断出的代词。尽管以Transformer为代表的的神经机器翻译模型取得了巨大的成功, 但这种省略现象依旧对神经机器翻译模型造成了很大的挑战。本文在Transformer基础上提出了一个融合零指代识别的翻译模型, 并引入篇章上下文来丰富指代信息。具体地, 该模型采用联合学习的框架, 在翻译模型基础上, 联合了一个分类任务, 即判别句子中省略代词在句子所表示的成分, 使得模型能够融合零指代信息辅助翻译。通过在中英对话数据集上的实验, 验证了本文提出方法的有效性, 与基准模型相比, 翻译性能提升了1. 48个BLEU值。”
1 code implementation • ECCV 2020 • Xiangyu Zhu, Fan Yang, Di Huang, Chang Yu, Hao Wang, Jianzhu Guo, Zhen Lei, Stan Z. Li
However, most of their training data is constructed by 3D Morphable Model, whose space spanned is only a small part of the shape space.
no code implementations • 10 Sep 2024 • Yang Wen, Anyu Lai, Bo Qian, Hao Wang, Wuzhen Shi, Wenming Cao
In this paper, we propose a Task Sequence Generator module that, in conjunction with the Task Intra-patch Block, effectively extracts task-specific features embedded in degraded images.
no code implementations • 5 Sep 2024 • Yang Wen, Anyu Lai, Bo Qian, Hao Wang, Wuzhen Shi, Wenming Cao
In this paper, we introduce a novel multi-task severe weather removal model that can effectively handle complex weather conditions in an adaptive manner.
no code implementations • 24 Aug 2024 • Hao Wang, Jindong Han, Wei Fan, Hao liu
Moreover, the linear mapping of continuous time series to a compressed subset vocabulary in reprogramming constrains the spatio-temporal semantic expressivity of PLMs and may lead to potential information bottleneck.
no code implementations • 21 Aug 2024 • Hao Wang, Yongqiang Han, Kefan Wang, Kai Cheng, Zhen Wang, Wei Guo, Yong liu, Defu Lian, Enhong Chen
Its objective is to extract knowledge from extensive pre-training data and fine-tune the model for downstream tasks.
no code implementations • 19 Aug 2024 • Haoyu Zhao, Chen Yang, Hao Wang, Xingyue Zhao, Wei Shen
To address this issue, we propose SG-GS, which uses semantics-embedded 3D Gaussians, skeleton-driven rigid deformation, and non-rigid cloth dynamics deformation to create photo-realistic animatable human avatars from monocular videos.
no code implementations • 19 Aug 2024 • Haoyu Zhao, Hao Wang, Chen Yang, Wei Shen
To address this challenge, we propose CHASE, which introduces supervision from intrinsic 3D consistency across poses and 3D geometry contrastive learning, achieving performance comparable with sparse inputs to that with full inputs.
1 code implementation • 10 Aug 2024 • Zeyu Gao, Hao Wang, Yuanda Wang, Chao Zhang
Assembly code search is vital for reducing the burden on reverse engineers, allowing them to quickly identify specific functions using natural language within vast binary programs.
no code implementations • 6 Aug 2024 • Ammar Mansoor Kamoona, Hui Song, Mahdi Jalili, Hao Wang, Reza Razzaghi, Xinghuo Yu
It dynamically leverages coarse-scale historical information using an M-TR encoder from an extended global temporal window, in conjunction with an M-TR decoder that concentrates on a limited time frame, local window, aiming to capture the fine-scale characteristics of the smart meter data.
no code implementations • 6 Aug 2024 • Yuxin Wang, Duanyu Feng, Yongfu Dai, Zhengyu Chen, Jimin Huang, Sophia Ananiadou, Qianqian Xie, Hao Wang
In this paper, we take a step forward to explore LLMs for tabular data synthesis and privacy protection, by introducing a new framework HARMONIC for tabular data generation and evaluation.
no code implementations • 29 Jul 2024 • Shohei Tanaka, Hao Wang, Yoshitaka Ushiku
All of the posters and papers in our dataset are under the CC-BY license and are publicly available.
no code implementations • 29 Jul 2024 • Hao Wang, Lingfeng Zhang, Erjia Xiao, Xin Wang, Zhongrui Wang, Renjing Xu
Non-invasive mobile electroencephalography (EEG) acquisition systems have been utilized for long-term monitoring of seizures, yet they suffer from limited battery life.
no code implementations • 29 Jul 2024 • Zixuan Chen, Xuandong Liu, Minglin Li, Yinfan Hu, Hao Mei, Huifeng Xing, Hao Wang, Wanxin Shi, Sen Liu, Yang Xu
The emerging In-network Aggregation (INA) has been proposed to integrate with PS to mitigate its incast issue.
1 code implementation • 24 Jul 2024 • Hao Wang, Xiangyu Yang, Yichen Zhu
This paper explores a specific type of nonconvex sparsity-promoting regularization problems, namely those involving $\ell_p$-norm regularization, in conjunction with a twice continuously differentiable loss function.
no code implementations • 22 Jul 2024 • Hanxi Guo, Hao Wang, Tao Song, Tianhang Zheng, Yang Hua, Haibing Guan, Xiangyu Zhang
Without direct access to the client's data, federated learning (FL) is well-known for its unique strength in data privacy protection among existing distributed machine learning techniques.
no code implementations • 20 Jul 2024 • Jiayu Lin, Guanrong Chen, Bojun Jin, Chenyang Li, Shutong Jia, Wancong Lin, Yang Sun, Yuhang He, Caihua Yang, Jianzhu Bao, Jipeng Wu, Wen Su, Jinglu Chen, Xinyi Li, Tianyu Chen, Mingjie Han, Shuaiwen Du, Zijian Wang, Jiyin Li, Fuzhong Suo, Hao Wang, Nuanchen Lin, Xuanjing Huang, Changjian Jiang, Ruifeng Xu, Long Zhang, Jiuxin Cao, Ting Jin, Zhongyu Wei
In this paper we present the results of the AI-Debater 2023 Challenge held by the Chinese Conference on Affect Computing (CCAC 2023), and introduce the related datasets.
1 code implementation • 17 Jul 2024 • Hao Wang, Wenhui Zhu, Jiayou Qin, Xin Li, Oana Dumitrascu, Xiwen Chen, Peijie Qiu, Abolfazl Razi
Detecting retinal image analysis, particularly the geometrical features of branching points, plays an essential role in diagnosing eye diseases.
1 code implementation • 16 Jul 2024 • Wenxiang Jiang, Hanwei Zhang, Shuo Zhao, Zhongwen Guo, Hao Wang
In this study, we contribute to this area by introducing the Illusory Poisoning Attack against Neural Radiance Fields (IPA-NeRF).
no code implementations • 14 Jul 2024 • Marawan Elbatel, Hualiang Wang, Jixiang Chen, Hao Wang, Xiaomeng Li
Existing FedSemi methods typically fail to aggregate models from unlabeled clients due to their inherent unreliability, thus overlooking unique information from their heterogeneous data distribution, leading to sub-optimal results.
1 code implementation • 10 Jul 2024 • Hao Wang, Pengzhen Ren, Zequn Jie, Xiao Dong, Chengjian Feng, Yinlong Qian, Lin Ma, Dongmei Jiang, YaoWei Wang, Xiangyuan Lan, Xiaodan Liang
To address these challenges, we propose a novel unified open-vocabulary detection method called OV-DINO, which is pre-trained on diverse large-scale datasets with language-aware selective fusion in a unified framework.
Ranked #4 on Zero-Shot Object Detection on MSCOCO (using extra training data)
1 code implementation • 9 Jul 2024 • Mingjia Yin, Chuhan Wu, YuFei Wang, Hao Wang, Wei Guo, Yasheng Wang, Yong liu, Ruiming Tang, Defu Lian, Enhong Chen
Inspired by the information compression nature of LLMs, we uncover an ``entropy law'' that connects LLM performance with data compression ratio and first-epoch training loss, which reflect the information redundancy of a dataset and the mastery of inherent knowledge encoded in this dataset, respectively.
no code implementations • 5 Jul 2024 • Ye Bai, Jingping Chen, Jitong Chen, Wei Chen, Zhuo Chen, Chuang Ding, Linhao Dong, Qianqian Dong, Yujiao Du, Kepan Gao, Lu Gao, Yi Guo, Minglun Han, Ting Han, Wenchao Hu, Xinying Hu, Yuxiang Hu, Deyu Hua, Lu Huang, Mingkun Huang, Youjia Huang, Jishuo Jin, Fanliu Kong, Zongwei Lan, Tianyu Li, Xiaoyang Li, Zeyang Li, Zehua Lin, Rui Liu, Shouda Liu, Lu Lu, Yizhou Lu, Jingting Ma, Shengtao Ma, Yulin Pei, Chen Shen, Tian Tan, Xiaogang Tian, Ming Tu, Bo wang, Hao Wang, Yuping Wang, Yuxuan Wang, Hanzhang Xia, Rui Xia, Shuangyi Xie, Hongmin Xu, Meng Yang, Bihong Zhang, Jun Zhang, Wanyi Zhang, Yang Zhang, Yawei Zhang, Yijie Zheng, Ming Zou
Modern automatic speech recognition (ASR) model is required to accurately transcribe diverse speech signals (from different domains, languages, accents, etc) given the specific contextual information in various application scenarios.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
2 code implementations • 4 Jul 2024 • Keyu An, Qian Chen, Chong Deng, Zhihao Du, Changfeng Gao, Zhifu Gao, Yue Gu, Ting He, Hangrui Hu, Kai Hu, Shengpeng Ji, Yabin Li, Zerui Li, Heng Lu, Haoneng Luo, Xiang Lv, Bin Ma, Ziyang Ma, Chongjia Ni, Changhe Song, Jiaqi Shi, Xian Shi, Hao Wang, Wen Wang, Yuxuan Wang, Zhangyu Xiao, Zhijie Yan, Yexin Yang, Bin Zhang, Qinglin Zhang, Shiliang Zhang, Nan Zhao, Siqi Zheng
This report introduces FunAudioLLM, a model family designed to enhance natural voice interactions between humans and large language models (LLMs).
no code implementations • 3 Jul 2024 • Yu Huang, Min Zhou, Menglin Yang, Zhen Wang, Muhan Zhang, Jie Wang, Hong Xie, Hao Wang, Defu Lian, Enhong Chen
Recent advancements in graph learning have revolutionized the way to understand and analyze data with complex structures.
no code implementations • 1 Jul 2024 • Hao Wang, Zhichao Chen, Yuan Shen, Jiajun Fan, Zhaoran Liu, Degui Yang, Xinggao Liu, Haoxuan Li
Heterogeneous treatment effect (HTE) estimation from observational data poses significant challenges due to treatment selection bias.
1 code implementation • 1 Jul 2024 • Zelong Li, Shuyuan Xu, Kai Mei, Wenyue Hua, Balaji Rama, Om Raheja, Hao Wang, He Zhu, Yongfeng Zhang
We believe that the automatic generation and interpretation of workflows in natural language represent a promising paradigm for solving complex tasks, particularly with the rapid development of LLMs.
1 code implementation • 27 Jun 2024 • Hao Yu, Xin Yang, Xin Gao, Yan Kang, Hao Wang, Junbo Zhang, Tianrui Li
In addition, we design a selective prompt fusion mechanism for aggregating knowledge of global prompts distilled from different clients.
no code implementations • 24 Jun 2024 • Haoxuan Li, Chunyuan Zheng, Wenjie Wang, Hao Wang, Fuli Feng, Xiao-Hua Zhou
Ratings of a user to most items in recommender systems are usually missing not at random (MNAR), largely because users are free to choose which items to rate.
1 code implementation • 22 Jun 2024 • Hao Wang, Ye Wang, Xiangyu Yang
We prove the global convergence of the proposed algorithm, guaranteeing that every limit point of the iterates is a critical point.
no code implementations • 19 Jun 2024 • Hao Wang, Euijoon Ahn, Jinman Kim
Further, due to their customization of the transformer architecture, they cannot use the advancements made in general video transformers (GVT).
no code implementations • 18 Jun 2024 • Egor Ershov, Artyom Panshin, Oleg Karasev, Sergey Korchagin, Shepelev Lev, Alexandr Startsev, Daniil Vladimirov, Ekaterina Zaychenkova, Nikola Banić, Dmitrii Iarchuk, Maria Efimova, Radu Timofte, Arseniy Terekhin, Shuwei Yue, Yuyang Liu, Minchen Wei, Lu Xu, Chao Zhang, Yasi Wang, Furkan Kınlı, Doğa Yılmaz, Barış Özcan, Furkan Kıraç, Shuai Liu, Jingyuan Xiao, Chaoyu Feng, Hao Wang, Guangqi Shao, Yuqian Zhang, Yibin Huang, Wei Luo, Liming Wang, Xiaotao Wang, Lei Lei, Simone Zini, Claudio Rota, Marco Buzzelli, Simone Bianco, Raimondo Schettini, Jin Guo, Tianli Liu, Mohao Wu, Ben Shao, Qirui Yang, Xianghui Li, Qihua Cheng, Fangpu Zhang, Zhiqiang Xu, Jingyu Yang, Huanjing Yue
The top ranking participants' solutions effectively represent the state-of-the-art in nighttime photography rendering.
no code implementations • 18 Jun 2024 • Hengyi Wang, Shiwei Tan, Hao Wang
We introduce a variational Bayesian explanation framework, dubbed ProbAbilistic Concept Explainers (PACE), which models the distributions of patch embeddings to provide trustworthy post-hoc conceptual explanations.
no code implementations • 17 Jun 2024 • Yibin Wang, Haizhou Shi, Ligong Han, Dimitris Metaxas, Hao Wang
Large Language Models (LLMs) often suffer from overconfidence during inference, particularly when adapted to downstream domain-specific tasks with limited data.
1 code implementation • 17 Jun 2024 • Hengyi Wang, Haizhou Shi, Shiwei Tan, Weiyi Qin, Wenyuan Wang, Tunyu Zhang, Akshay Nambi, Tanuja Ganu, Hao Wang
However, a comprehensive evaluation of their long-context capabilities remains underexplored.
Ranked #1000000000 on Hallucination on MMNeedle
1 code implementation • 17 Jun 2024 • Nvidia, :, Bo Adler, Niket Agarwal, Ashwath Aithal, Dong H. Anh, Pallab Bhattacharya, Annika Brundyn, Jared Casper, Bryan Catanzaro, Sharon Clay, Jonathan Cohen, Sirshak Das, Ayush Dattagupta, Olivier Delalleau, Leon Derczynski, Yi Dong, Daniel Egert, Ellie Evans, Aleksander Ficek, Denys Fridman, Shaona Ghosh, Boris Ginsburg, Igor Gitman, Tomasz Grzegorzek, Robert Hero, Jining Huang, Vibhu Jawa, Joseph Jennings, Aastha Jhunjhunwala, John Kamalu, Sadaf Khan, Oleksii Kuchaiev, Patrick Legresley, Hui Li, Jiwei Liu, Zihan Liu, Eileen Long, Ameya Sunil Mahabaleshwarkar, Somshubra Majumdar, James Maki, Miguel Martinez, Maer Rodrigues de Melo, Ivan Moshkov, Deepak Narayanan, Sean Narenthiran, Jesus Navarro, Phong Nguyen, Osvald Nitski, Vahid Noroozi, Guruprasad Nutheti, Christopher Parisien, Jupinder Parmar, Mostofa Patwary, Krzysztof Pawelec, Wei Ping, Shrimai Prabhumoye, Rajarshi Roy, Trisha Saar, Vasanth Rao Naik Sabavat, Sanjeev Satheesh, Jane Polak Scowcroft, Jason Sewall, Pavel Shamis, Gerald Shen, Mohammad Shoeybi, Dave Sizer, Misha Smelyanskiy, Felipe Soares, Makesh Narsimhan Sreedhar, Dan Su, Sandeep Subramanian, Shengyang Sun, Shubham Toshniwal, Hao Wang, Zhilin Wang, Jiaxuan You, Jiaqi Zeng, Jimmy Zhang, Jing Zhang, Vivienne Zhang, Yian Zhang, Chen Zhu
We release the Nemotron-4 340B model family, including Nemotron-4-340B-Base, Nemotron-4-340B-Instruct, and Nemotron-4-340B-Reward.
no code implementations • 16 Jun 2024 • Shuyang Lin, Tong Jia, Hao Wang, Bowen Ma, Mingyuan Li, Dongyue Chen
To address aforementioned challenges, in this paper, we introduce distillation-based open-vocabulary object detection (OVOD) task into X-ray security inspection domain by extending CLIP to learn visual representations in our specific X-ray domain, aiming to detect novel prohibited item categories beyond base categories on which the detector is trained.
no code implementations • 15 Jun 2024 • Wentian Wang, Paul Kantor, Jacob Feldman, Lazaros Gallos, Hao Wang
We propose MMLU-SR, a novel dataset designed to measure the true comprehension abilities of Large Language Models (LLMs) by challenging their performance in question-answering tasks with modified terms.
no code implementations • 14 Jun 2024 • Jingtao Cao, Zheng Zhang, Hongru Wang, Bin Liang, Hao Wang, Kam-Fai Wong
Utilizing the BLIP model for image captioning, PP-OCR and TrOCR for text recognition across multiple languages, and the Qwen LLM for nuanced language understanding, our system is capable of identifying harmful content in memes created in English, Chinese, Malay, and Tamil.
no code implementations • 13 Jun 2024 • Ziyan Wang, Hao Wang
Our approach learns a bandit model for the target domain by collecting feedback from the source domain.
no code implementations • 8 Jun 2024 • Hengguan Huang, Xing Shen, Songtao Wang, Dianbo Liu, Hao Wang
Faced with complex problems, the human brain demonstrates a remarkable capacity to transcend sensory input and form latent understandings of perceived world patterns.
1 code implementation • 5 Jun 2024 • Tingjia Shen, Hao Wang, Jiaqing Zhang, Sirui Zhao, Liangyue Li, Zulong Chen, Defu Lian, Enhong Chen
To this end, we propose a novel framework named URLLM, which aims to improve the CDSR performance by exploring the User Retrieval approach and domain grounding on LLM simultaneously.
1 code implementation • 5 Jun 2024 • Mingyuan Li, Tong Jia, Hui Lu, Bowen Ma, Hao Wang, Dongyue Chen
Prohibited Item detection in X-ray images is one of the most effective security inspection methods. However, differing from natural light images, the unique overlapping phenomena in X-ray images lead to the coupling of foreground and background features, thereby lowering the accuracy of general object detectors. Therefore, we propose a Multi-Class Min-Margin Contrastive Learning (MMCL) method that, by clarifying the category semantic information of content queries under the deformable DETR architecture, aids the model in extracting specific category foreground information from coupled features. Specifically, after grouping content queries by the number of categories, we employ the Multi-Class Inter-Class Exclusion (MIE) loss to push apart content queries from different groups.
no code implementations • 4 Jun 2024 • Kun Zhou, Shengkui Zhao, Yukun Ma, Chong Zhang, Hao Wang, Dianwen Ng, Chongjia Ni, Nguyen Trung Hieu, Jia Qi Yip, Bin Ma
Recent language model-based text-to-speech (TTS) frameworks demonstrate scalability and in-context learning capabilities.
no code implementations • 31 May 2024 • Xinxi Zhang, Song Wen, Ligong Han, Felix Juefei-Xu, Akash Srivastava, Junzhou Huang, Hao Wang, Molei Tao, Dimitris N. Metaxas
We introduce Spectral Orthogonal Decomposition Adaptation (SODA), which balances computational efficiency and representation capacity.
no code implementations • 30 May 2024 • Xiaoyu Wu, Jiaru Zhang, Yang Hua, Bohan Lyu, Hao Wang, Tao Song, Haibing Guan
Through this modeling, we identify the primary cause of this corruption stage: a narrowed learning distribution inherent in the nature of few-shot fine-tuning.
no code implementations • 30 May 2024 • Songning Lai, Ninghui Feng, Jiechao Gao, Hao Wang, Haochen Sui, Xin Zou, Jiayu Yang, Wenshuo Chen, Hang Zhao, Xuming Hu, Yutao Yue
The field of time series forecasting has garnered significant attention in recent years, prompting the development of advanced models like TimeSieve, which demonstrates impressive performance.
no code implementations • 29 May 2024 • Kaveh Alimohammadi, Hao Wang, Ojas Gulati, Akash Srivastava, Navid Azizan
Existing differentially private (DP) synthetic data generation mechanisms typically assume a single-source table.
no code implementations • 28 May 2024 • Youlong Ding, Xueyang Wu, Yining Meng, Yonggang Luo, Hao Wang, Weike Pan
Deep learning with differential privacy (DP) has garnered significant attention over the past years, leading to the development of numerous methods aimed at enhancing model accuracy and training efficiency.
1 code implementation • 28 May 2024 • Mingjia Yin, Hao Wang, Wei Guo, Yong liu, Suojuan Zhang, Sirui Zhao, Defu Lian, Enhong Chen
The sequential recommender (SR) system is a crucial component of modern recommender systems, as it aims to capture the evolving preferences of users.
1 code implementation • 27 May 2024 • Tianshu Wang, Xiaoyang Chen, Hongyu Lin, Xuanang Chen, Xianpei Han, Hao Wang, Zhenyu Zeng, Le Sun
Based on our findings, we further design a compound entity matching framework (ComEM) that leverages the composition of multiple strategies and LLMs.
1 code implementation • 26 May 2024 • Xijie Huang, Xinyuan Wang, Hantao Zhang, Yinghao Zhu, Jiawen Xi, Jingkun An, Hao Wang, Hao Liang, Chengwei Pan
Security concerns related to Large Language Models (LLMs) have been extensively explored, yet the safety implications for Multimodal Large Language Models (MLLMs), particularly in medical contexts (MedMLLMs), remain insufficiently studied.
no code implementations • 26 May 2024 • Hao Wang, Jianwei Li, Zhengyu Li
In conclusion, the AI-generated text detection model based on the BERT algorithm proposed in this study shows high accuracy and stability in experiments, providing an effective solution for related fields.
1 code implementation • 23 May 2024 • Zhuowei Li, Zihao Xu, Ligong Han, Yunhe Gao, Song Wen, Di Liu, Hao Wang, Dimitris N. Metaxas
In-context Learning (ICL) empowers large language models (LLMs) to adapt to unseen tasks during inference by prefixing a few demonstration examples prior to test queries.
no code implementations • 21 May 2024 • Mingjia Yin, Hao Wang, Wei Guo, Yong liu, Zhi Li, Sirui Zhao, Zhen Wang, Defu Lian, Enhong Chen
Cross-domain sequential recommendation (CDSR) aims to uncover and transfer users' sequential preferences across multiple recommendation domains.
no code implementations • 14 May 2024 • Hao Wang, Nao Li
To verify that the FiiNet model can dynamically learn the importance of feature interaction combinations in a fine-grained manner and improve the model's recommendation performance and interpretability, this paper compares it with many click-through rate prediction models on two real datasets, proving that the FiiNet model incorporating the selective kernel network can effectively improve the recommendation effect and provide better interpretability.
no code implementations • 9 May 2024 • Hao Wang, Angel E. Rodriguez-Fernandez, Lourdes Uribe, André Deutz, Oziel Cortés-Piña, Oliver Schütze
In this work, we propose a set-based Newton method for Hausdorff approximations of the Pareto front to be used within multi-objective evolutionary algorithms.
no code implementations • 7 May 2024 • Hao Wu, Ruochong LI, Hao Wang, Hui Xiong
To address this issue, we propose COM3D, making the first attempt to exploit the cross-view correspondence and cross-modal mining to enhance the retrieval performance.
2 code implementations • 6 May 2024 • Xiwen Chen, Peijie Qiu, Wenhui Zhu, Huayu Li, Hao Wang, Aristeidis Sotiras, Yalin Wang, Abolfazl Razi
Deep neural networks, including transformers and convolutional neural networks, have significantly improved multivariate time series classification (MTSC).
no code implementations • 2 May 2024 • Hao Wang, Tetsuro Morimura, Ukyo Honda, Daisuke Kawahara
Non-autoregressive (NAR) language models are known for their low latency in neural machine translation (NMT).
1 code implementation • 1 May 2024 • Yucheng Shi, Alexandros Agapitos, David Lynch, Giorgio Cruciata, Cengis Hasan, Hao Wang, Yayu Yao, Aleksandar Milenovic
In Multi-objective Reinforcement Learning (MORL) agents are tasked with optimising decision-making behaviours that trade-off between multiple, possibly conflicting, objectives.
no code implementations • 1 May 2024 • Huai-an Su, Jiaxiang Geng, Liang Li, Xiaoqi Qin, Yanzhao Hou, Hao Wang, Xin Fu, Miao Pan
Although such fixed size subnetwork assignment enables FL training over heterogeneous mobile devices, it is unaware of (i) the dynamic changes of devices' communication and computing conditions and (ii) FL training progress and its dynamic requirements of local training contributions, both of which may cause very long FL training delay.
no code implementations • 30 Apr 2024 • Cengis Hasan, Alexandros Agapitos, David Lynch, Alberto Castagna, Giorgio Cruciata, Hao Wang, Aleksandar Milenovic
We present a method that addresses the pain point of long lead-time required to deploy cell-level parameter optimisation policies to new wireless network sites.
1 code implementation • 29 Apr 2024 • Chuni Liu, Boyuan Ma, Xiaojuan Ban, Yujie Xie, Hao Wang, Weihua Xue, Jingchao Ma, Ke Xu
Topological consistency plays a crucial role in the task of boundary segmentation for reticular images, such as cell membrane segmentation in neuron electron microscopic images, grain boundary segmentation in material microscopic images and road segmentation in aerial images.
no code implementations • 27 Apr 2024 • Chenghao Huang, Xiaolu Chen, Yanru Zhang, Hao Wang
FedCRL introduces contrastive representation learning (CRL) on shared representations to facilitate knowledge acquisition of clients.
2 code implementations • 25 Apr 2024 • Haizhou Shi, Zihao Xu, Hengyi Wang, Weiyi Qin, Wenyuan Wang, Yibin Wang, Zifeng Wang, Sayna Ebrahimi, Hao Wang
In this survey, we provide a comprehensive overview of the current research progress on LLMs within the context of CL.
1 code implementation • 25 Apr 2024 • Hao Wang, Jiayou Qin, Xiwen Chen, Ashish Bastola, John Suchanek, Zihao Gong, Abolfazl Razi
Furthermore, in the experiments part, we show the qualitative analysis of motor focus estimation between the conventional dense optical flow-based method and the proposed method.
1 code implementation • 24 Apr 2024 • Zhuoqun Li, Hongyu Lin, Tianshu Wang, Boxi Cao, Yaojie Lu, Weixiang Zhou, Hao Wang, Zhenyu Zeng, Le Sun, Xianpei Han
Linking a claim to grounded references is a critical ability to fulfill human demands for authentic and reliable information.
no code implementations • 22 Apr 2024 • Hao Wang, Qingshan Xu, Hongyuan Chen, Rui Ma
In this work, we introduce PGAHum, a prior-guided geometry and appearance learning framework for high-fidelity animatable human reconstruction.
no code implementations • 13 Apr 2024 • Shan Gao, Amit K. Chakraborty, Russell Greiner, Mark A. Lewis, Hao Wang
In summary, we showed that there are statistical features that distinguish outbreak and non-outbreak sequences long before outbreaks occur.
no code implementations • 7 Apr 2024 • Hao Wang, Yanping Chen, Weizhe Yang, Yongbin Qin, Ruizhang Huang
The results indicate that two-dimensional feature engineering can take advantage of a two-dimensional sentence representation and make full use of prior knowledge in traditional feature engineering.
no code implementations • 6 Apr 2024 • Siyuan Tian, Hao Wang, Yiren Rong, Junhao Wang, Renjie Dai, Zhengxiao He
Modern displays nowadays possess the capability to render video content with a high dynamic range (HDR) and an extensive color gamut . However, the majority of available resources are still in standard dynamic range (SDR).
no code implementations • 30 Mar 2024 • Luankang Zhang, Hao Wang, Suojuan Zhang, Mingjia Yin, Yongqiang Han, Jiaqing Zhang, Defu Lian, Enhong Chen
To this end, we propose a Unified Framework for Adaptive Representation Enhancement and Inversed Learning in Cross-Domain Recommendation (AREIL).
1 code implementation • 26 Mar 2024 • Jinyi Li, Yihuai Lan, Lei Wang, Hao Wang
Prompt compression is an innovative method for efficiently condensing input prompts while preserving essential information.
no code implementations • 26 Mar 2024 • Yongqiang Han, Hao Wang, Kefan Wang, Likang Wu, Zhi Li, Wei Guo, Yong liu, Defu Lian, Enhong Chen
In recommendation systems, users frequently engage in multiple types of behaviors, such as clicking, adding to a cart, and purchasing.
1 code implementation • 25 Mar 2024 • Xiaoxuan Yu, Hao Wang, Weiming Li, Qiang Wang, SoonYong Cho, Younghun Sung
In this work, we propose a novel Disentangled Object-Centric TRansformer (DOCTR) that explores object-centric representation to facilitate learning with multiple objects for the multiple sub-tasks in a unified manner.
1 code implementation • 24 Mar 2024 • Amit K. Chakraborty, Shan Gao, Reza Miry, Pouria Ramazi, Russell Greiner, Mark A. Lewis, Hao Wang
The timely detection of disease outbreaks through reliable early warning signals (EWSs) is indispensable for effective public health mitigation strategies.
no code implementations • 23 Mar 2024 • Hao Wang, Tang Li, Chenhui Chu, Nengjun Zhu, Rui Wang, Pinpin Zhu
This approach aims to generate relation representations that are more aware of the spatial context and unseen relation in a manner similar to human perception.
no code implementations • 20 Mar 2024 • Fucai Ke, Hao Wang
To address this research gap, inspired by the concept of non-intrusive load monitoring (NILM), we develop a home charging prediction method using historical smart meter data.
no code implementations • 20 Mar 2024 • Canchen Jiang, Hao Wang
Community battery systems have been widely deployed to provide services to the grid.
no code implementations • 20 Mar 2024 • Jiarong Fan, Ariel Liebman, Hao Wang
The increasing integration of electric vehicles (EVs) into the grid can pose a significant risk to the distribution system operation in the absence of coordination.
1 code implementation • 19 Mar 2024 • Hao Wang, Jiayou Qin, Ashish Bastola, Xiwen Chen, John Suchanek, Zihao Gong, Abolfazl Razi
This paper explores the potential of Large Language Models(LLMs) in zero-shot anomaly detection for safe visual navigation.
no code implementations • CVPR 2024 • Xiaoyu Wu, Yang Hua, Chumeng Liang, Jiaru Zhang, Hao Wang, Tao Song, Haibing Guan
In response, we present Contrasting Gradient Inversion for Diffusion Models (CGI-DM), a novel method featuring vivid visual representations for digital copyright authentication.
4 code implementations • 12 Mar 2024 • Zhicheng Guo, Sijie Cheng, Hao Wang, Shihao Liang, Yujia Qin, Peng Li, Zhiyuan Liu, Maosong Sun, Yang Liu
The virtual API server contains a caching system and API simulators which are complementary to alleviate the change in API status.
1 code implementation • 12 Mar 2024 • Abdul Fatir Ansari, Lorenzo Stella, Caner Turkmen, Xiyuan Zhang, Pedro Mercado, Huibin Shen, Oleksandr Shchur, Syama Sundar Rangapuram, Sebastian Pineda Arango, Shubham Kapoor, Jasper Zschiegner, Danielle C. Maddix, Hao Wang, Michael W. Mahoney, Kari Torkkola, Andrew Gordon Wilson, Michael Bohlke-Schneider, Yuyang Wang
We introduce Chronos, a simple yet effective framework for pretrained probabilistic time series models.
no code implementations • 8 Mar 2024 • Jiajie Fan, Amal Trigui, Thomas Bäck, Hao Wang
As such, FID might not be suitable to assess the performance of DGMs for a generative design task.
1 code implementation • 8 Mar 2024 • Yunhao Li, Qin Li, Hao Wang, Xue Ma, Jiali Yao, Shaohua Dong, Heng Fan, Libo Zhang
Current multi-object tracking (MOT) aims to predict trajectories of targets (i. e., ''where'') in videos.
1 code implementation • 7 Mar 2024 • Mingyuan Li, Tong Jia, Hao Wang, Bowen Ma, Shuyang Lin, Da Cai, Dongyue Chen
Considering the significant overlapping phenomenon in X-ray prohibited item images, we propose an Anti-Overlapping DETR (AO-DETR) based on one of the state-of-the-art general object detectors, DINO.
1 code implementation • 6 Mar 2024 • Wenfeng Feng, Chuzhan Hao, Yuewei Zhang, Yu Han, Hao Wang
These LoRA modules can be aligned with the expert design principles observed in Mixture-of-Experts (MoE).
1 code implementation • 6 Mar 2024 • Hao Wang, Sayed Pedram Haeri Boroujeni, Xiwen Chen, Ashish Bastola, Huayu Li, Abolfazl Razi
Thus, our proposed framework can generate a massive dataset of that images are high-quality and ground truth-paired, which well addresses the needs of the annotated datasets in specific tasks.
no code implementations • 6 Mar 2024 • Cheng-Yen Yang, Hsiang-Wei Huang, Zhongyu Jiang, Hao Wang, Farron Wallace, Jenq-Neng Hwang
Dense object counting or crowd counting has come a long way thanks to the recent development in the vision community.
no code implementations • 29 Feb 2024 • Ji Ma, Hongming Dai, Yao Mu, Pengying Wu, Hao Wang, Xiaowei Chi, Yang Fei, Shanghang Zhang, Chang Liu
Zero-Shot Object Navigation (ZSON) requires agents to autonomously locate and approach unseen objects in unfamiliar environments and has emerged as a particularly challenging task within the domain of Embodied AI.
no code implementations • 29 Feb 2024 • Jinhao Li, Changlong Wang, Yanru Zhang, Hao Wang
To bridge this gap, we develop a novel BESS joint bidding strategy that utilizes deep reinforcement learning (DRL) to bid in the spot and contingency frequency control ancillary services (FCAS) markets.
1 code implementation • 28 Feb 2024 • Lei Wang, Wanyu Xu, Zhiqiang Hu, Yihuai Lan, Shan Dong, Hao Wang, Roy Ka-Wei Lee, Ee-Peng Lim
This paper introduces a new in-context learning (ICL) mechanism called In-Image Learning (I$^2$L) that combines demonstration examples, visual cues, and chain-of-thought reasoning into an aggregated image to enhance the capabilities of Large Multimodal Models (e. g., GPT-4V) in multimodal reasoning tasks.
1 code implementation • 26 Feb 2024 • Hao Wang, Shengda Luo, Guosheng Hu, JianGuo Zhang
In aid of this indicator, we present a novel Gradient-guided Modality Decoupling (GMD) method to decouple the dependency on dominating modalities.
1 code implementation • 26 Feb 2024 • Hao Wang, Zeyu Gao, Chao Zhang, Zihan Sha, Mingyang Sun, Yuchen Zhou, Wenyu Zhu, Wenju Sun, Han Qiu, Xi Xiao
At the core, our approach boosts superior transfer learning capabilities by effectively aligning binary code with their semantics explanations (in natural language), resulting a model able to generate better embeddings for binary code.
no code implementations • 25 Feb 2024 • Tianyu Chen, Haoyi Zhou, Ying Li, Hao Wang, Chonghan Gao, Shanghang Zhang, JianXin Li
Foundation models have revolutionized knowledge acquisition across domains, and our study introduces OmniArch, a paradigm-shifting approach designed for building foundation models in multi-physics scientific computing.
no code implementations • 25 Feb 2024 • Hao Wang, Hao Li, Minlie Huang, Lei Sha
In addition, our approach can be generalized into a broader method for generating transferable adversarial suffixes that can successfully attack multiple LLMs, even black-box LLMs, such as ChatGPT and Gemini.
no code implementations • 22 Feb 2024 • Ziqi Yin, Hao Wang, Kaito Horio, Daisuke Kawahara, Satoshi Sekine
We investigate the impact of politeness levels in prompts on the performance of large language models (LLMs).
2 code implementations • 20 Feb 2024 • Qianqian Xie, Weiguang Han, Zhengyu Chen, Ruoyu Xiang, Xiao Zhang, Yueru He, Mengxi Xiao, Dong Li, Yongfu Dai, Duanyu Feng, Yijing Xu, Haoqiang Kang, Ziyan Kuang, Chenhan Yuan, Kailai Yang, Zheheng Luo, Tianlin Zhang, Zhiwei Liu, Guojun Xiong, Zhiyang Deng, Yuechen Jiang, Zhiyuan Yao, Haohang Li, Yangyang Yu, Gang Hu, Jiajia Huang, Xiao-Yang Liu, Alejandro Lopez-Lira, Benyou Wang, Yanzhao Lai, Hao Wang, Min Peng, Sophia Ananiadou, Jimin Huang
Our evaluation of 15 representative LLMs, including GPT-4, ChatGPT, and the latest Gemini, reveals several key findings: While LLMs excel in IE and textual analysis, they struggle with advanced reasoning and complex tasks like text generation and forecasting.
no code implementations • 15 Feb 2024 • Diederick Vermetten, Carola Doerr, Hao Wang, Anna V. Kononova, Thomas Bäck
The number of proposed iterative optimization heuristics is growing steadily, and with this growth, there have been many points of discussion within the wider community.
no code implementations • 10 Feb 2024 • Behzad Akbari, Mingfeng Yuan, Hao Wang, Haibin Zhu, Jinjun Shan
In the field of Multi-Agent Systems (MAS), known for their openness, dynamism, and cooperative nature, the ability to trust the resources and services of other agents is crucial.
1 code implementation • 8 Feb 2024 • Hengguan Huang, Songtao Wang, Hongfu Liu, Hao Wang, Ye Wang
Traditional applications of natural language processing (NLP) in healthcare have predominantly focused on patient-centered services, enhancing patient interactions and care delivery, such as through medical dialogue systems.
no code implementations • 7 Feb 2024 • Mengqi Chen, Bin Guo, Hao Wang, Haoyu Li, Qian Zhao, Jingqi Liu, Yasan Ding, Yan Pan, Zhiwen Yu
To depict the research trends of CogAgent, in this paper, we first present several fundamental cognitive psychology theories and give the formalized definition of three typical cognitive strategies, including the persuasion strategy, the topic path planning strategy, and the argument structure prediction strategy.
no code implementations • 6 Feb 2024 • Hao Wang, Xin Zhang, JinZhe Jiang, YaQian Zhao, Chen Li
However, it has been shown that multimodal NLP are vulnerable to adversarial attacks, where the outputs of a model can be dramatically changed by a perturbation to the input.
no code implementations • 6 Feb 2024 • Hao Wang, Lei Sha
The proposed approach aims to enhance the fluency of generated text by guiding the generation process with PPCs.
no code implementations • 5 Feb 2024 • Xu Huang, Weiwen Liu, Xiaolong Chen, Xingmei Wang, Hao Wang, Defu Lian, Yasheng Wang, Ruiming Tang, Enhong Chen
As Large Language Models (LLMs) have shown significant intelligence, the progress to leverage LLMs as planning modules of autonomous agents has attracted more attention.
1 code implementation • 4 Feb 2024 • Hao Wang, Licheng Pan, Zhichao Chen, Degui Yang, Sen Zhang, Yifei Yang, Xinggao Liu, Haoxuan Li, DaCheng Tao
Time series modeling is uniquely challenged by the presence of autocorrelation in both historical and label sequences.
1 code implementation • 3 Feb 2024 • Guang-Yuan Hao, Hengguan Huang, Haotian Wang, Jie Gao, Hao Wang
In this paper, we propose the first general method, dubbed composite active learning (CAL), for multi-domain AL. Our approach explicitly considers the domain-level and instance-level information in the problem; CAL first assigns domain-level budgets according to domain-level importance, which is estimated by optimizing an upper error bound that we develop; with the domain-level budgets, CAL then leverages a certain instance-level query strategy to select samples to label from each domain.
no code implementations • 2 Feb 2024 • Guang-Yuan Hao, Jiji Zhang, Biwei Huang, Hao Wang, Kun Zhang
Counterfactual reasoning is pivotal in human cognition and especially important for providing explanations and making decisions.
1 code implementation • 1 Feb 2024 • Zelong Li, Wenyue Hua, Hao Wang, He Zhu, Yongfeng Zhang
A stack-based LLM plan generation process is then conducted under the supervision of the automaton to ensure that the generated plan satisfies the constraints, making the planning process controllable.
1 code implementation • 29 Jan 2024 • Hao Wang, Tao Xiang, Shangwei Guo, Jialing He, Hangcheng Liu, Tianwei Zhang
Adopting untrusted PTMs may suffer from backdoor attacks, where the adversary can compromise the downstream models by injecting backdoors into the PTM.
no code implementations • 29 Jan 2024 • Jinhao Li, Changlong Wang, Hao Wang
This paper studies the synergy of solar-battery energy storage system (BESS) and develops a viable strategy for the BESS to unlock its economic potential by serving as a backup to reduce solar curtailments while also participating in the electricity market.
no code implementations • 29 Jan 2024 • Xiangzhao Qin, Sha Hu, Jiankun Zhang, Jing Qian, Hao Wang
Deep learning (DL) based channel estimation (CE) and multiple input and multiple output detection (MIMODet), as two separate research topics, have provided convinced evidence to demonstrate the effectiveness and robustness of artificial intelligence (AI) for receiver design.
no code implementations • 29 Jan 2024 • Jinhao Li, Ruichang Zhang, Hao Wang, Zhi Liu, Hongyang Lai, Yanru Zhang
Renewable energy resources (RERs) have been increasingly integrated into distribution networks (DNs) for decarbonization.
no code implementations • 27 Jan 2024 • Yuxin Liang, Zhuoyang Song, Hao Wang, Jiaxing Zhang
We evaluate the ability of Large Language Models (LLMs) to discern and express their internal knowledge state, a key factor in countering factual hallucination and ensuring reliable application of LLMs.
no code implementations • 26 Jan 2024 • Ashish Bastola, Julian Brinkley, Hao Wang, Abolfazl Razi
This paper presents a comprehensive literature review of the current state of in-vehicle human-computer interaction (HCI) in the context of self-driving vehicles, with a specific focus on inclusion and accessibility.
no code implementations • 25 Jan 2024 • Chaofan Pan, Xin Yang, Hao Wang, Wei Wei, Tianrui Li
Despite the progress in continual reinforcement learning (CRL), existing methods often suffer from insufficient knowledge transfer, particularly when the tasks are diverse.
1 code implementation • 25 Jan 2024 • Xinyue Xu, Yi Qin, Lu Mi, Hao Wang, Xiaomeng Li
Our ECBMs address both limitations of existing CBMs, providing higher accuracy and richer concept interpretations.
no code implementations • 24 Jan 2024 • Yunfan Zhang, Hong Huang, Zhiwei Xiong, Zhiqi Shen, Guosheng Lin, Hao Wang, Nicholas Vun
The core strength of our pipeline lies in its ability to generate 3D scenes that are not only visually impressive but also exhibit features like photorealism, multi-view consistency, and diversity.
no code implementations • 24 Jan 2024 • Yangsen Chen, Hao Wang
The accurate 3D reconstruction of deformable soft body tissues from endoscopic videos is a pivotal challenge in medical applications such as VR surgery and medical image analysis.
1 code implementation • 23 Jan 2024 • Chengzhi Mao, Carl Vondrick, Hao Wang, Junfeng Yang
We find that large language models (LLMs) are more likely to modify human-written text than AI-generated text when tasked with rewriting.
no code implementations • 18 Jan 2024 • Hao Wang
Fairness is a popular research topic in recent years.
no code implementations • 18 Jan 2024 • Hao Wang, Shuhei Kurita, Shuichiro Shimizu, Daisuke Kawahara
Audio-visual speech recognition (AVSR) is a multimodal extension of automatic speech recognition (ASR), using video as a complement to audio.
Audio-Visual Speech Recognition Automatic Speech Recognition +4
no code implementations • 17 Jan 2024 • Luwei Bai, Yaohua Hu, Hao Wang, Xiaoqi Yang
For DIRL$_1$, we show the reweighted $\ell_1$ subproblem has support identification property so that DIRL$_1$ locally reverts to a gradient descent algorithm around a stationary point.
no code implementations • 16 Jan 2024 • Gengyue Han, Xiaohan Liu, Xianyue Peng, Hao Wang, Yu Han
This study introduces CycLight, a novel cycle-level deep reinforcement learning (RL) approach for network-level adaptive traffic signal control (NATSC) systems.
no code implementations • 11 Jan 2024 • Xiwen Chen, Hao Wang, Zhao Zhang, Zhenmin Li, Huayu Li, Tong Ye, Abolfazl Razi
Untrained Physics-based Deep Learning (DL) methods for digital holography have gained significant attention due to their benefits, such as not requiring an annotated training dataset, and providing interpretability since utilizing the governing laws of hologram formation.
no code implementations • 29 Dec 2023 • Hao Wang, Bo Tang, Chi Harold Liu, Shangqin Mao, Jiahong Zhou, Zipeng Dai, Yaqi Sun, Qianlong Xie, Xingxing Wang, Dong Wang
Online display advertising platforms service numerous advertisers by providing real-time bidding (RTB) for the scale of billions of ad requests every day.
no code implementations • 27 Dec 2023 • Xin Yang, Hao Yu, Xin Gao, Hao Wang, Junbo Zhang, Tianrui Li
The key objective of FCL is to fuse heterogeneous knowledge from different clients and retain knowledge of previous tasks while learning on new ones.
no code implementations • 25 Dec 2023 • Hao Wang, Huabing Zhou, Yanduo Zhang, Tao Lu, Jiayi Ma
Scene text spotting is essential in various computer vision applications, enabling extracting and interpreting textual information from images.
no code implementations • 24 Dec 2023 • Rui Zhou, Haiyang Zhang, Hao Wang, Jin He, Qijun Huang, Sheng Chang
By integrating the local voltage-controlled magnetic anisotropy (VCMA) effect, Dzyaloshinskii-Moriya interaction (DMI) effect, and spin-orbit torque (SOT) effect, we propose a novel device structure for field-free magnetic tunnel junction (MTJ).
no code implementations • 24 Dec 2023 • Ming Yan, Ruihao Li, Hao Zhang, Hao Wang, Zhilan Yang, Ji Yan
Language agents have shown impressive problem-solving skills within defined settings and brief timelines.
1 code implementation • 22 Dec 2023 • Honghao Fu, Zhiqi Shen, Jing Jih Chin, Hao Wang
This leads to substantial limitations in existing works of visual stimuli reconstruction from EEG, such as difficulties in aligning EEG embeddings with the fine-grained semantic information and a heavy reliance on additional large self-collected dataset for training.
no code implementations • 22 Dec 2023 • Yujie Li, Xin Yang, Hao Wang, Xiangkun Wang, Tianrui Li
This paper studies the problem of continual learning in an open-world scenario, referred to as Open-world Continual Learning (OwCL).
2 code implementations • 20 Dec 2023 • Weibo Gao, Qi Liu, Hao Wang, Linan Yue, Haoyang Bi, Yin Gu, Fangzhou Yao, Zheng Zhang, Xin Li, Yuanjing He
Consequently, we refine the cognitive states of cold-start students as diagnostic outcomes via virtual data, aligning with the diagnosis-oriented goal.
1 code implementation • 19 Dec 2023 • Peishen Yan, Hao Wang, Tao Song, Yang Hua, Ruhui Ma, Ningxin Hu, Mohammad R. Haghighat, Haibing Guan
Federated Learning (FL) is becoming a popular paradigm for leveraging distributed data and preserving data privacy.
no code implementations • 16 Dec 2023 • Hao Wang
In this paper, we rely on the theory developed by Wang from 2021 to 2023 to demonstrate that online cultural rating platform rating data often evolve into Poisson/Pareto behavior, and individualistic voting preferences are predictable without any data input, so Borda Count Method (or, Range Voting Method) has intrinsic fallacy and should not be used as a voting theory method.