1 code implementation • EMNLP 2021 • Yingya Li, Jun Wang, Bei Yu
We also conducted a case study that applied this prediction model to retrieve specific health advice on COVID-19 treatments from LitCovid, a large COVID research literature portal, demonstrating the usefulness of retrieving health advice sentences as an advanced research literature navigation function for health researchers and the general public.
no code implementations • ACL 2022 • Jun Wang, Benjamin Rubinstein, Trevor Cohn
In this paper we describe a new source of bias prevalent in NMT systems, relating to translations of sentences containing person names.
no code implementations • Findings (EMNLP) 2021 • Chang Xu, Jun Wang, Francisco Guzmán, Benjamin Rubinstein, Trevor Cohn
NLP models are vulnerable to data poisoning attacks.
no code implementations • COLING 2022 • Alex X. Zhang, Xun Liang, Bo Wu, Xiangping Zheng, Sensen Zhang, Yuhui Guo, Jun Wang, Xinyao Liu
The human recognition system has presented the remarkable ability to effortlessly learn novel knowledge from only a few trigger events based on prior knowledge, which is called insight learning.
no code implementations • ACL 2022 • Ling.Yu Zhu, Zhengkun Zhang, Jun Wang, Hongbin Wang, Haiying Wu, Zhenglu Yang
Empathetic dialogue assembles emotion understanding, feeling projection, and appropriate response generation.
no code implementations • SemEval (NAACL) 2022 • Qizhi Lin, Changyu Hou, Xiaopeng Wang, Jun Wang, Yixuan Qiao, Peng Jiang, Xiandi Jiang, Benqi Wang, Qifeng Xiao
From pretrained contextual embedding to document-level embedding, the selection and construction of embedding have drawn more and more attention in the NER domain in recent research.
no code implementations • CCL 2022 • Zekun Deng, Hao Yang, Jun Wang
"《史记》和《汉书》具有经久不衰的研究价值。尽管两书异同的研究已经较为丰富, 但研究的全面性、完备性、科学性、客观性均仍显不足。在数字人文的视角下, 本文利用计算语言学方法, 通过对字、词、命名实体、段落等的多粒度、多角度分析, 开展对于《史》《汉》的比较研究。首先, 本文对于《史》《汉》中的字、词、命名实体的分布和特点进行对比, 以遍历穷举的考察方式提炼出两书在主要内容上的相同点与不同点, 揭示了汉武帝之前和汉武帝到西汉灭亡两段历史时期在政治、文化、思想上的重要变革与承袭。其次, 本文使用一种融入命名实体作为外部特征的文本相似度算法对于《史记》《汉书》的异文进行自动发现, 成功识别出过去研究者通过人工手段没有发现的袭用段落, 使得我们对于《史》《汉》的承袭关系形成更加完整和立体的认识。再次, 本文通过计算异文段落之间的最长公共子序列来自动得出两段异文之间存在的差异, 从宏观统计上证明了《汉书》文字风格《史记》的差别, 并从微观上进一步对二者语言特点进行了阐释, 为理解《史》《汉》异文特点提供了新的角度和启发。本研究站在数字人文的视域下, 利用先进的计算方法对于传世千年的中国古代经典进行了再审视、再发现, 其方法对于今人研究古籍有一定的借鉴价值。”
no code implementations • 10 Oct 2024 • Xue Yan, Yan Song, Xidong Feng, Mengyue Yang, Haifeng Zhang, Haitham Bou Ammar, Jun Wang
In sequential decision-making (SDM) tasks, methods like reinforcement learning (RL) and heuristic search have made notable advances in specific cases.
1 code implementation • 8 Oct 2024 • Wei Liu, Zhiying Deng, Zhongyu Niu, Jun Wang, Haozhao Wang, Yuankai Zhang, Ruixuan Li
In the optimization objectives of these methods, spurious features are still distinguished from plain noise, which hinders the discovery of causal rationales.
no code implementations • 7 Oct 2024 • Fenia Christopoulou, Ronald Cardenas, Gerasimos Lampouras, Haitham Bou-Ammar, Jun Wang
Preference Optimization (PO) has proven an effective step for aligning language models to human-desired behaviors.
no code implementations • 7 Oct 2024 • Ao Hu, Dongkai Wang, Yong Dai, shiyi qi, Liangjian Wen, Jun Wang, Zhi Chen, Xun Zhou, Zenglin Xu, Jiang Duan
Transformer-based models demonstrate significant potential in modeling cross-time and cross-variable interaction.
1 code implementation • 6 Oct 2024 • Qiqiang Lin, Muning Wen, Qiuying Peng, Guanyu Nie, Junwei Liao, Xiaoyun Mo, Jiamu Zhou, Cheng Cheng, Yin Zhao, Jun Wang, Weinan Zhang
Large language models have demonstrated impressive value in performing as autonomous agents when equipped with external tools and API calls.
1 code implementation • 6 Oct 2024 • Jiakai Tang, Heyang Gao, Xuchen Pan, Lei Wang, Haoran Tan, Dawei Gao, Yushuo Chen, Xu Chen, Yankai Lin, Yaliang Li, Bolin Ding, Jingren Zhou, Jun Wang, Ji-Rong Wen
With the rapid advancement of large language models (LLMs), recent years have witnessed many promising studies on leveraging LLM-based agents to simulate human social behavior.
no code implementations • 4 Oct 2024 • Matthieu Zimmer, Milan Gritta, Gerasimos Lampouras, Haitham Bou Ammar, Jun Wang
The growth in the number of parameters of Large Language Models (LLMs) has led to a significant surge in computational requirements, making them challenging and costly to deploy.
no code implementations • 2 Oct 2024 • Jun Wang, Yu Mao, Nan Guan, Chun Jason Xue
For each dimension of the bag-level representation, attribution values are calculated to indicate how changes in the specific dimensions of the input affect the model output.
no code implementations • 21 Sep 2024 • ZhiHao Lin, Wei Ma, Mingyi Zhou, Yanjie Zhao, Haoyu Wang, Yang Liu, Jun Wang, Li Li
During our manual attempts to perform jailbreak attacks, we found that the vocabulary of the response of the target model gradually became richer and eventually produced harmful responses.
no code implementations • 19 Sep 2024 • Wenbo Wei, Jun Wang, Abhir Bhalerao
To help address the occlusion problem in panoptic segmentation and image understanding, this paper proposes a new large-scale dataset, COCO-Occ, which is derived from the COCO dataset by manually labelling the COCO images into three perceived occlusion levels.
no code implementations • 10 Sep 2024 • Zihan Liao, Hang Yu, Lingxiao Wei, Jianguo Li, Jun Wang, Wei zhang
In the realm of Large Language Models (LLMs), the ability to process long contexts is increasingly crucial for tasks such as multi-round dialogues, code generation, and document summarization.
no code implementations • 9 Sep 2024 • Zhipeng Li, Xiaofen Xing, Jun Wang, Shuaiqi Chen, Guoqiao Yu, Guanglu Wan, Xiangmin Xu
In recent years, there has been significant progress in Text-to-Speech (TTS) synthesis technology, enabling the high-quality synthesis of voices in common scenarios.
1 code implementation • 3 Sep 2024 • Wenlin Li, Yucheng Xu, Xiaoqing Zheng, Suoya Han, Jun Wang, Xiaobo Sun
Sparse and noisy images (SNIs), like those in spatial gene expression data, pose significant challenges for effective representation learning and clustering, which are essential for thorough data analysis and interpretation.
no code implementations • 1 Sep 2024 • Xuechen Liang, Meiling Tao, Yinghui Xia, Tianyu Shi, Jun Wang, Jingsong Yang
As large language models (LLMs) improve their capabilities in handling complex tasks, the issues of computational cost and efficiency due to long prompts are becoming increasingly prominent.
no code implementations • 1 Sep 2024 • Xuechen Liang, Meiling Tao, Yinghui Xia, Tianyu Shi, Jun Wang, Jingsong Yang
Large language models (LLMs) have made significant advances in the field of natural language processing, but they still face challenges such as continuous decision-making.
no code implementations • 27 Aug 2024 • Yuheng Feng, Yangfan He, Yinghui Xia, Tianyu Shi, Jun Wang, Jinsong Yang
In this research, we aim to enhance the user-friendliness of our image generation system.
1 code implementation • 26 Aug 2024 • Yang Qiu, Wei Liu, Jun Wang, Ruixuan Li
Due to the dimensionality reduction of features in the latent space of the auto-encoder, it becomes easier to extract causal features leading to the model's output, which can be easily employed to generate explanations.
1 code implementation • 26 Aug 2024 • Cong Xu, Zhangchi Zhu, Mo Yu, Jun Wang, Jianyong Wang, Wei zhang
Some studies have observed that LLMs, when fine-tuned by the cross-entropy (CE) loss with a full softmax, could achieve `state-of-the-art' performance in sequential recommendation.
no code implementations • 24 Aug 2024 • Tianxiang Huang, Jing Shi, Ge Jin, Juncheng Li, Jun Wang, Jun Du, Jun Shi
In this work, we propose a novel hip landmark detection model by integrating the Topological GCN (TGCN) with an Improved Conformer (TGCN-ICF) into a unified frame-work to improve detection performance.
no code implementations • 22 Aug 2024 • Can Qin, Congying Xia, Krithika Ramakrishnan, Michael Ryoo, Lifu Tu, Yihao Feng, Manli Shu, Honglu Zhou, Anas Awadalla, Jun Wang, Senthil Purushwalkam, Le Xue, Yingbo Zhou, Huan Wang, Silvio Savarese, Juan Carlos Niebles, Zeyuan Chen, ran Xu, Caiming Xiong
We present xGen-VideoSyn-1, a text-to-video (T2V) generation model capable of producing realistic scenes from textual descriptions.
no code implementations • 22 Aug 2024 • Kaihui Cheng, Ce Liu, Qingkun Su, Jun Wang, Liwei Zhang, Yining Tang, Yao Yao, Siyu Zhu, Yuan Qi
Protein structure prediction is pivotal for understanding the structure-function relationship of proteins, advancing biological research, and facilitating pharmaceutical development and experimental design.
no code implementations • 22 Aug 2024 • Ce Liu, Jun Wang, Zhiqiang Cai, Yingxu Wang, Huizhen Kuang, Kaihui Cheng, Liwei Zhang, Qingkun Su, Yining Tang, Fenglei Cao, Limei Han, Siyu Zhu, Yuan Qi
Despite significant progress in static protein structure collection and prediction, the dynamic behavior of proteins, one of their most vital characteristics, has been largely overlooked in prior research.
no code implementations • 20 Aug 2024 • Yuankai Zhang, Lingxiao Kong, Haozhao Wang, Ruixuan Li, Jun Wang, Yuhua Li, Wei Liu
Based on this, we make a series of recommendations for improving rationalization models in terms of explanation.
no code implementations • 20 Aug 2024 • Huafeng Qin, Yuming Fu, Huiyan Zhang, Mounim A. El-Yacoubi, Xinbo Gao, Qun Song, Jun Wang
At the testing stage, given an adversarial sample, the MsMemoryGAN retrieves its most relevant normal patterns in memory for the reconstruction.
no code implementations • 18 Aug 2024 • Jun Wang, Yu Mao, Nan Guan, Chun Jason Xue
Whole slide images (WSIs) are gigapixel-scale digital images of H\&E-stained tissue samples widely used in pathology.
Interpretable Machine Learning Multiple Instance Learning +2
1 code implementation • 16 Aug 2024 • Le Xue, Manli Shu, Anas Awadalla, Jun Wang, An Yan, Senthil Purushwalkam, Honglu Zhou, Viraj Prabhu, Yutong Dai, Michael S Ryoo, Shrikant Kendre, Jieyu Zhang, Can Qin, Shu Zhang, Chia-Chih Chen, Ning Yu, Juntao Tan, Tulika Manoj Awalgaonkar, Shelby Heinecke, Huan Wang, Yejin Choi, Ludwig Schmidt, Zeyuan Chen, Silvio Savarese, Juan Carlos Niebles, Caiming Xiong, ran Xu
The framework comprises meticulously curated datasets, a training recipe, model architectures, and a resulting suite of LMMs.
no code implementations • 15 Aug 2024 • Bohao Wang, Feng Liu, Jiawei Chen, Yudi Wu, Xingyu Lou, Jun Wang, Yan Feng, Chun Chen, Can Wang
To tackle these challenges, we propose LLM4DSR, a tailored approach for denoising sequential recommendation using LLMs.
1 code implementation • 15 Aug 2024 • Jun Wang, Linyan Li, Qi Liu, Yu Yang
In summary, this paper presents our well-structured investigations and new findings when applying offline reinforcement learning to building HVAC systems.
no code implementations • 15 Aug 2024 • Jun Wang, Likang Wu, Qi Liu, Yu Yang
However, previous studies mainly focus on discrete action and policy spaces, which might have difficulties in handling dramatically growing items efficiently.
1 code implementation • 9 Aug 2024 • Zhibo Zhang, Wuxia Bai, Yuxi Li, Mark Huasong Meng, Kailong Wang, Ling Shi, Li Li, Jun Wang, Haoyu Wang
In this work, we aim to enhance the understanding of glitch tokens and propose techniques for their detection and mitigation.
no code implementations • 2 Aug 2024 • Xingyu Lou, Yu Yang, Kuiyao Dong, Heyuan Huang, Wenyi Yu, Ping Wang, Xiu Li, Jun Wang
As the recommendation service needs to address increasingly diverse distributions, such as multi-population, multi-scenario, multitarget, and multi-interest, more and more recent works have focused on multi-distribution modeling and achieved great progress.
no code implementations • 26 Jul 2024 • Jun Wang, Ying Yuan, Haichuan Che, Haozhi Qi, Yi Ma, Jitendra Malik, Xiaolong Wang
This serves two purposes: 1) pre-training a sensorimotor policy in simulation; 2) conducting open-loop trajectory replay in the real world.
no code implementations • 19 Jul 2024 • Tingkai Zhang, Chaoyu Chen, Cong Liao, Jun Wang, Xudong Zhao, Hang Yu, Jianchao Wang, Jianguo Li, Wenhui Shi
Text-to-SQL conversion is a critical innovation, simplifying the transition from complex SQL to intuitive natural language queries, especially significant given SQL's prevalence in the job market across various roles.
no code implementations • 18 Jul 2024 • Yuchen Weng, Zhengwen Shen, Ruofan Chen, Qi Wang, Jun Wang
3D deblurring reconstruction techniques have recently seen significant advancements with the development of Neural Radiance Fields (NeRF) and 3D Gaussian Splatting (3DGS).
no code implementations • 15 Jul 2024 • Jun Wang, Eleftheria Briakou, Hamid Dadkhahi, Rishabh Agarwal, Colin Cherry, Trevor Cohn
A critical component in knowledge distillation is the means of coupling the teacher and student.
no code implementations • 12 Jul 2024 • Jie Zheng, Ru Wen, Haiqin Hu, Lina Wei, Kui Su, Wei Chen, Chen Liu, Jun Wang
Existing Masked Image Modeling (MIM) depends on a spatial patch-based masking-reconstruction strategy to perceive objects'features from unlabeled images, which may face two limitations when applied to chest CT: 1) inefficient feature learning due to complex anatomical details presented in CT images, and 2) suboptimal knowledge transfer owing to input disparity between upstream and downstream models.
no code implementations • 12 Jul 2024 • Zafeirios Fountas, Martin A Benfeghoul, Adnan Oomerjee, Fenia Christopoulou, Gerasimos Lampouras, Haitham Bou-Ammar, Jun Wang
Large language models (LLMs) have shown remarkable capabilities, but still struggle with processing extensive contexts, limiting their ability to maintain coherence and accuracy over long sequences.
no code implementations • 11 Jul 2024 • Linglong Qian, Tao Wang, Jun Wang, Hugh Logan Ellis, Robin Mitra, Richard Dobson, Zina Ibrahim
By identifying conceptual gaps in the literature and existing reviews, we devise a taxonomy grounded on the inductive bias of neural imputation frameworks, resulting in a classification of existing deep imputation strategies based on their suitability for specific imputation scenarios and data-specific properties.
no code implementations • 1 Jul 2024 • Dan Peng, Zhihui Fu, Jun Wang
To tackle this, we propose employing derivative-free optimization techniques to enable on-device fine-tuning of LLM, even on memory-limited mobile devices.
1 code implementation • 28 Jun 2024 • Christopher E. Mower, Yuhui Wan, Hongzhan Yu, Antoine Grosnit, Jonas Gonzalez-Billandon, Matthieu Zimmer, Jinlong Wang, Xinyu Zhang, Yao Zhao, Anbang Zhai, Puze Liu, Daniel Palenicek, Davide Tateo, Cesar Cadena, Marco Hutter, Jan Peters, Guangjian Tian, Yuzheng Zhuang, Kun Shao, Xingyue Quan, Jianye Hao, Jun Wang, Haitham Bou-Ammar
Key features of the framework include: integration of ROS with an AI agent connected to a plethora of open-source and commercial LLMs, automatic extraction of a behavior from the LLM output and execution of ROS actions/services, support for three behavior modes (sequence, behavior tree, state machine), imitation learning for adding new robot actions to the library of possible actions, and LLM reflection via human and environment feedback.
1 code implementation • 25 Jun 2024 • Zihan Liao, Hang Yu, Jianguo Li, Jun Wang, Wei zhang
In this paper, we present D2LLMs-Decomposed and Distilled LLMs for semantic search-that combines the best of both worlds.
1 code implementation • 21 Jun 2024 • Chengzhe Piao, Taiyu Zhu, Yu Wang, Stephanie E Baldeweg, Paul Taylor, Pantelis Georgiou, Jiahao Sun, Jun Wang, Kezhi Li
Newly diagnosed Type 1 Diabetes (T1D) patients often struggle to obtain effective Blood Glucose (BG) prediction models due to the lack of sufficient BG data from Continuous Glucose Monitoring (CGM), presenting a significant "cold start" problem in patient care.
4 code implementations • 18 Jun 2024 • Wenjie Du, Jun Wang, Linglong Qian, Yiyuan Yang, Fanxing Liu, Zepu Wang, Zina Ibrahim, Haoxin Liu, Zhiyuan Zhao, Yingjie Zhou, Wenjia Wang, Kaize Ding, Yuxuan Liang, B. Aditya Prakash, Qingsong Wen
Despite the development of numerous deep learning algorithms for time series imputation, the community lacks standardized and comprehensive benchmark platforms to effectively evaluate imputation performance across different settings.
1 code implementation • 17 Jun 2024 • Xingjian Hu, Baole Wei, Liangcai Gao, Jun Wang
In this paper, we propose a general framework for historical document text detection (SegHist), enabling existing segmentation-based text detection methods to effectively address the challenges, especially text lines with high aspect ratios.
no code implementations • 13 Jun 2024 • Zhiyuan Wang, Wei zhang, Jun Wang
As the demand for programming skills grows across industries and academia, students often turn to Programming Online Judge (POJ) platforms for coding practice and competition.
no code implementations • 11 Jun 2024 • Yuxuan Mu, Shihao Zou, Kangning Yin, Zheng Tian, Li Cheng, Weinan Zhang, Jun Wang
The retriever searches motion experts from a user-specified database in a task-oriented fashion, which boosts the responsiveness to the user's control.
no code implementations • 7 Jun 2024 • Xihan Li, Xing Li, Lei Chen, Xing Zhang, Mingxuan Yuan, Jun Wang
While deep learning has achieved significant success in various domains, its application to logic circuit design has been limited due to complex constraints and strict feasibility requirement.
no code implementations • 7 Jun 2024 • Yinghui Xia, Yuyan Chen, Tianyu Shi, Jun Wang, Jinsong Yang
Therefore, we construct AICoderEval, a dataset focused on real-world tasks in various domains based on HuggingFace, PyTorch, and TensorFlow, along with comprehensive metrics for evaluation and enhancing LLMs' task-specific code generation capability.
no code implementations • 5 Jun 2024 • Yu Zhang, Rui Yu, Zhipeng Yao, Wenyuan Zhang, Jun Wang, Liming Zhang
However, we find that its principle can lead to overestimation phenomenon for the value function.
no code implementations • 30 May 2024 • Xuanfa Jin, Ziyan Wang, Yali Du, Meng Fang, Haifeng Zhang, Jun Wang
Communication is a fundamental aspect of human society, facilitating the exchange of information and beliefs among people.
1 code implementation • 28 May 2024 • Hongze Sun, Rui Liu, Wuque Cai, Jun Wang, Yue Wang, Huajin Tang, Yan Cui, Dezhong Yao, Daqing Guo
In this study, we propose a novel multimodal hybrid tracker (MMHT) that utilizes frame-event-based data for reliable single object tracking.
no code implementations • 26 May 2024 • Anjie Liu, Jianhong Wang, Haoxuan Li, Xu Chen, Jun Wang, Samuel Kaski, Mengyue Yang
In human-AI interaction, a prominent goal is to attain human`s desirable outcome with the assistance of AI agents, which can be ideally delineated as a problem of seeking the optimal Nash Equilibrium that matches the human`s desirable outcome.
no code implementations • 19 May 2024 • Xuanli He, Qiongkai Xu, Jun Wang, Benjamin I. P. Rubinstein, Trevor Cohn
Modern NLP models are often trained on public datasets drawn from diverse sources, rendering them vulnerable to data poisoning attacks.
no code implementations • 19 May 2024 • Jun Wang, Benedetta Tondi, Mauro Barni
Extensive efforts have been made to explore unique representations of generative models and use them to attribute a synthetic image to the model that produced it.
no code implementations • 17 May 2024 • Yi Yao, Jun Wang, Yabai Hu, LiFeng Wang, Yi Zhou, Jack Chen, Xuming Gai, Zhenming Wang, Wenjun Liu
The evolution of software testing from manual to automated methods has significantly influenced quality assurance (QA) practices.
no code implementations • 13 May 2024 • Jun Wang, Yu Mao, Yufei Cui, Nan Guan, Chun Jason Xue
Immunohistochemistry (IHC) plays a crucial role in pathology as it detects the over-expression of protein in tissue samples.
no code implementations • 9 May 2024 • Tianfu Qi, Jun Wang, Zexue Zhao
For the MSK demodulation based on the Viterbi algorithm, we derive a lower and upper bound of BER.
no code implementations • 5 May 2024 • Jinmin Li, Tao Dai, Jingyun Zhang, Kang Liu, Jun Wang, Shaoming Wang, Shu-Tao Xia, rizen guo
Recently developed generative methods, including invertible rescaling network (IRN) based and generative adversarial network (GAN) based methods, have demonstrated exceptional performance in image rescaling.
1 code implementation • 1 May 2024 • Yu Cui, Feng Liu, Pengbo Wang, Bohao Wang, Heng Tang, Yi Wan, Jun Wang, Jiawei Chen
Owing to their powerful semantic reasoning capabilities, Large Language Models (LLMs) have been effectively utilized as recommenders, achieving impressive performance.
no code implementations • 30 Apr 2024 • Xuanli He, Jun Wang, Qiongkai Xu, Pasquale Minervini, Pontus Stenetorp, Benjamin I. P. Rubinstein, Trevor Cohn
The implications of backdoor attacks on English-centric large language models (LLMs) have been widely examined - such attacks can be achieved by embedding malicious behaviors during training and activated under specific conditions that trigger malicious outputs.
no code implementations • 30 Apr 2024 • Jiabao Wang, Yang Wu, Jun Wang, Ni Chen
The multi-plane phase retrieval method provides a budget-friendly and effective way to perform phase imaging, yet it often encounters alignment challenges due to shifts along the optical axis in experiments.
no code implementations • 30 Apr 2024 • Wentao Lei, Li Liu, Jun Wang
Therefore, we propose a novel Gloss-prompted Diffusion-based CS Gesture generation framework (called GlossDiff).
1 code implementation • 26 Apr 2024 • Kaichen Xu, Yueyang Ding, Suyang Hou, Weiqiang Zhan, Nisang Chen, Jun Wang, Xiaobo Sun
In response, we propose ACSleuth, a novel, reconstruction deviation-guided generative framework that integrates the detection, domain adaptation, and fine-grained annotating of anomalous cells into a methodologically cohesive workflow.
1 code implementation • 24 Apr 2024 • Junfeng Tian, Rui Wang, Cong Li, Yudong Zhou, Jun Liu, Jun Wang
This report details the development and key achievements of our latest language model designed for custom large language models.
1 code implementation • ICCV 2023 • Renrong Shao, Wei zhang, Jianhua Yin, Jun Wang
Our approach utilizes an adversarial distillation framework with attention generator, mixed high-order attention distillation, and semantic feature contrast learning.
Data-free Knowledge Distillation Fine-Grained Visual Categorization +1
no code implementations • 18 Apr 2024 • WenHao Zhang, Jun Wang, Yong Luo, Lei Yu, Wei Yu, Zheng He
Then we design a spatio-temporal fusion module based on temporal granularity alignment, where the global spatial features extracted from event frames, together with the local relative spatial and temporal features contained in voxel graph list are effectively aligned and integrated.
1 code implementation • 18 Apr 2024 • Yongcheng Zeng, Guoqing Liu, Weiyu Ma, Ning Yang, Haifeng Zhang, Jun Wang
Fine-tuning pre-trained Large Language Models (LLMs) is essential to align them with human values and intentions.
1 code implementation • 17 Apr 2024 • Qiyu Hou, Jun Wang, Meixuan Qiao, Lujun Tian
By leveraging the actual structure and content of tables from Chinese financial announcements, we have developed the first extensive table annotation dataset in this domain.
no code implementations • 17 Apr 2024 • Jun Wang, Yufei Cui, Yu Mao, Nan Guan, Chun Jason Xue
Pre-processing whole slide images (WSIs) can impact classification performance.
no code implementations • 9 Apr 2024 • ZhiHao Lin, Wei Ma, Tao Lin, Yaowen Zheng, Jingquan Ge, Jun Wang, Jacques Klein, Tegawende Bissyande, Yang Liu, Li Li
We introduce a governance framework centered on federated learning (FL), designed to foster the joint development and maintenance of open-source AI code models while safeguarding data privacy and security.
no code implementations • 7 Apr 2024 • Haifeng Wang, Hao Xu, Jun Wang, Jian Zhou, Ke Deng
Recognizing various surgical tools, actions and phases from surgery videos is an important problem in computer vision with exciting clinical applications.
1 code implementation • CVPR 2024 • Sichen Chen, Yingyi Zhang, Siming Huang, Ran Yi, Ke Fan, Ruixin Zhang, Peixian Chen, Jun Wang, Shouhong Ding, Lizhuang Ma
To mitigate the problem of under-fitting, we design a transformer module named Multi-Cycled Transformer(MCT) based on multiple-cycled forwards to more fully exploit the potential of small model parameters.
no code implementations • 3 Apr 2024 • Jun Wang, Qiongkai Xu, Xuanli He, Benjamin I. P. Rubinstein, Trevor Cohn
Our aim is to bring attention to these vulnerabilities within MNMT systems with the hope of encouraging the community to address security concerns in machine translation, especially in the context of low-resource languages.
1 code implementation • 2 Apr 2024 • Xuechen Liang, Meiling Tao, Yinghui Xia, Tianyu Shi, Jun Wang, Jingsong Yang
Open large language models (LLMs) have significantly advanced the field of natural language processing, showcasing impressive performance across various tasks. Despite the significant advancements in LLMs, their effective operation still relies heavily on human input to accurately guide the dialogue flow, with agent tuning being a crucial optimization technique that involves human adjustments to the model for better response to such guidance. Addressing this dependency, our work introduces the TinyAgent model, trained on a meticulously curated high-quality dataset.
no code implementations • 1 Apr 2024 • Zuyu Xu, Kang Shen, Pengnian Cai, Tao Yang, Yuanming Hu, Shixian Chen, Yunlai Zhu, Zuheng Wu, Yuehua Dai, Jun Wang, Fei Yang
The recent emergence of the hybrid quantum-classical neural network (HQCNN) architecture has garnered considerable attention due to the potential advantages associated with integrating quantum principles to enhance various facets of machine learning algorithms and computations.
no code implementations • 26 Mar 2024 • Youpeng Zhao, Di wu, Jun Wang
In a single GPU-CPU system, we demonstrate that under varying workloads, ALISA improves the throughput of baseline systems such as FlexGen and vLLM by up to 3X and 1. 9X, respectively.
1 code implementation • 22 Mar 2024 • Xuemei Tang, Zekun Deng, Qi Su, Hao Yang, Jun Wang
Additionally, we have evaluated the capabilities of Large Language Models (LLMs) in the context of tasks related to ancient Chinese history.
2 code implementations • CVPR 2024 • Yuxi Mi, Zhizhou Zhong, Yuge Huang, Jiazhen Ji, Jianqing Xu, Jun Wang, Shaoming Wang, Shouhong Ding, Shuigeng Zhou
Recognizable identity features within the image are encouraged by co-training a recognition model on its high-dimensional feature representation.
no code implementations • 14 Mar 2024 • Qirui Mi, Zhiyu Zhao, Siyu Xia, Yan Song, Jun Wang, Haifeng Zhang
This paper proposes a novel general framework named Dynamic Stackelberg Mean Field Games (Dynamic SMFG) to model such policymaking within sequential decision-making processes, with the government as the leader and households as dynamic followers.
no code implementations • 14 Mar 2024 • Xihan Li, Xing Li, Lei Chen, Xing Zhang, Mingxuan Yuan, Jun Wang
Then, can circuits also be mastered by a a sufficiently large "circuit model", which can conquer electronic design tasks by simply predicting the next logic gate?
no code implementations • 13 Mar 2024 • Ben Athiwaratkun, Sujan Kumar Gonugondla, Sanjay Krishna Gouda, Haifeng Qian, Hantian Ding, Qing Sun, Jun Wang, Jiacheng Guo, Liangfu Chen, Parminder Bhatia, Ramesh Nallapati, Sudipta Sengupta, Bing Xiang
This study introduces bifurcated attention, a method designed to enhance language model inference in shared-context batch decoding scenarios.
1 code implementation • 11 Mar 2024 • Siyu Duan, Jun Wang, Qi Su
Cultural heritage serves as the enduring record of human thought and history.
no code implementations • 11 Mar 2024 • Chaochao Chen, Yizhao Zhang, Yuyuan Li, Dan Meng, Jun Wang, Xiaoli Zheng, Jianwei Yin
The first component is distinguishability loss, where we design a distribution-based measurement to make attribute labels indistinguishable from attackers.
no code implementations • 9 Mar 2024 • Jingyun Xue, Tao Wang, Jun Wang, Kaihao Zhang, Wenhan Luo, Wenqi Ren, Zikun Liu, Hyunhee Park, Xiaochun Cao
Specifically, we utilize sparse self-attention to filter out redundant information and noise, directing the model's attention to focus on the features more relevant to the degraded regions in need of reconstruction.
no code implementations • 8 Mar 2024 • Jun Wang, Lixing Zhu, Abhir Bhalerao, Yulan He
Radiology report generation (RRG) methods often lack sufficient medical knowledge to produce clinically accurate reports.
no code implementations • 8 Mar 2024 • Jingxiao Chen, Ziqin Gong, Minghuan Liu, Jun Wang, Yong Yu, Weinan Zhang
To overcome this problem and to have an effective solution against hard constraints, we proposed a novel learning-based method that uses looking-ahead information as the feature to improve the legality of TSP with Time Windows (TSPTW) solutions.
no code implementations • 5 Mar 2024 • Hanlei Jin, Yang Zhang, Dan Meng, Jun Wang, Jinghua Tan
Automatic Text Summarization (ATS), utilizing Natural Language Processing (NLP) algorithms, aims to create concise and accurate summaries, thereby significantly reducing the human effort required in processing large volumes of text.
1 code implementation • 2 Mar 2024 • Chenchen Tao, Xiaohao Peng, Chong Wang, Jiafei Wu, Puning Zhao, Jun Wang, Jiangbo Qian
Most models for weakly supervised video anomaly detection (WS-VAD) rely on multiple instance learning, aiming to distinguish normal and abnormal snippets without specifying the type of anomaly.
no code implementations • 28 Feb 2024 • Yang Cao, Shuo Shang, Jun Wang, Wei zhang
This paper explores providing explainability for session-based recommendation (SR) by path reasoning.
no code implementations • 28 Feb 2024 • Youpeng Zhao, Ming Lin, Huadong Tang, Qiang Wu, Jun Wang
Generative Large Language Models (LLMs) stand as a revolutionary advancement in the modern era of artificial intelligence (AI).
1 code implementation • 27 Feb 2024 • Siyuan Guo, Cheng Deng, Ying Wen, Hechang Chen, Yi Chang, Jun Wang
In this work, we investigate the potential of large language models (LLMs) based agents to automate data science tasks, with the goal of comprehending task requirements, then building and training the best-fit machine learning models.
no code implementations • 26 Feb 2024 • Chengzhe Piao, Taiyu Zhu, Stephanie E Baldeweg, Paul Taylor, Pantelis Georgiou, Jiahao Sun, Jun Wang, Kezhi Li
Accurate prediction of future blood glucose (BG) levels can effectively improve BG management for people living with diabetes, thereby reducing complications and improving quality of life.
no code implementations • 23 Feb 2024 • Jun Wang, Guocheng He, Yiannis Kantaros
To address this challenge, we introduce a new distributed LLM-based planner, called S-ATLAS for Safe plAnning for Teams of Language-instructed AgentS, that is capable of achieving user-defined mission success rates.
no code implementations • 22 Feb 2024 • Xuemei Tang, Jun Wang, Qi Su
Recently, large language models (LLMs) have been successful in relational extraction (RE) tasks, especially in the few-shot learning.
no code implementations • CVPR 2024 • Jun Wang, Yuzhe Qin, Kaiming Kuang, Yigit Korkmaz, Akhilan Gurumoorthy, Hao Su, Xiaolong Wang
We introduce CyberDemo, a novel approach to robotic imitation learning that leverages simulated human demonstrations for real-world tasks.
no code implementations • 20 Feb 2024 • Adam X. Yang, Maxime Robeyns, Thomas Coste, Zhengyan Shi, Jun Wang, Haitham Bou-Ammar, Laurence Aitchison
To ensure that large language model (LLM) responses are helpful and non-toxic, a reward model trained on human preference data is usually used.
no code implementations • 15 Feb 2024 • Min Zhang, Sato Takumi, Jack Zhang, Jun Wang
Large Language Models (LLMs) excel in generating personalized content and facilitating interactive dialogues, showcasing their remarkable aptitude for a myriad of applications.
no code implementations • 13 Feb 2024 • Zhaoan Wang, Shaoping Xiao, Jun Wang, Ashwin Parab, Shivam Patel
This study examines how artificial intelligence (AI), especially Reinforcement Learning (RL), can be used in farming to boost crop yields, fine-tune nitrogen use and watering, and reduce nitrate runoff and greenhouse gases, focusing on Nitrous Oxide (N$_2$O) emissions from soil.
no code implementations • 11 Feb 2024 • Xidong Feng, Ziyu Wan, Mengyue Yang, Ziyan Wang, Girish A. Koushik, Yali Du, Ying Wen, Jun Wang
Reinforcement Learning (RL) has shown remarkable abilities in learning policies for decision-making tasks.
no code implementations • 9 Feb 2024 • Cong Xu, Zhangchi Zhu, Jun Wang, Jianyong Wang, Wei zhang
Large language models (LLMs) have gained much attention in the recommendation community; some studies have observed that LLMs, fine-tuned by the cross-entropy loss with a full softmax, could achieve state-of-the-art performance already.
no code implementations • 9 Feb 2024 • Tianfu Qi, Jun Wang
In the first part, we propose a closed-form heavy-tailed multivariate probability density function (PDF) that to model the bursty mixed noise.
1 code implementation • 9 Feb 2024 • Muning Wen, Junwei Liao, Cheng Deng, Jun Wang, Weinan Zhang, Ying Wen
We assess the effectiveness of ETPO within a simulated environment that models data science code generation as a series of multi-step interactive tasks; results underline ETPO's potential as a robust method for refining the interactive decision-making capabilities of language agents.
no code implementations • 8 Feb 2024 • Jun Wang, Haoxuan Li, Chi Zhang, Dongxu Liang, Enyun Yu, Wenwu Ou, Wenjia Wang
Recommender systems are designed to learn user preferences from observed feedback and comprise many fundamental tasks, such as rating prediction and post-click conversion rate (pCVR) prediction.
5 code implementations • 6 Feb 2024 • Jun Wang, Wenjie Du, Wei Cao, Keli Zhang, Wenjia Wang, Yuxuan Liang, Qingsong Wen
In this paper, we conduct a comprehensive survey on the recently proposed deep learning imputation methods.
no code implementations • 23 Jan 2024 • Jiarui Jin, Zexue He, Mengyue Yang, Weinan Zhang, Yong Yu, Jun Wang, Julian McAuley
Subsequently, we minimize the mutual information between the observation estimation and the relevance estimation conditioned on the input features.
no code implementations • 22 Jan 2024 • Chao Song, Zhihao Ye, Qiqiang Lin, Qiuying Peng, Jun Wang
In practice, there are two prevailing ways, in which the adaptation can be achieved: (i) Multiple Independent Models: Pre-trained LLMs are fine-tuned a few times independently using the corresponding training samples from each task.
no code implementations • 18 Jan 2024 • Jun Wang, Chengfeng Zhou, Zhaoyan Ming, Lina Wei, Xudong Jiang, Dahong Qian
One of the fundamental challenges in microscopy (MS) image analysis is instance segmentation (IS), particularly when segmenting cluster regions where multiple objects of varying sizes and shapes may be connected or even overlapped in arbitrary orientations.
1 code implementation • 17 Jan 2024 • Meng Fang, Shilong Deng, Yudi Zhang, Zijing Shi, Ling Chen, Mykola Pechenizkiy, Jun Wang
A wide range of real-world applications is characterized by their symbolic nature, necessitating a strong capability for symbolic reasoning.
no code implementations • 16 Jan 2024 • Huafeng Qin, Yiquan Wu, Mounim A. El-Yacoubi, Jun Wang, Guangxiang Yang
To overcome this problem, in this paper, we propose an adversarial masking contrastive learning (AMCL) approach, that generates challenging samples to train a more robust contrastive learning model for the downstream palm-vein recognition task, by alternatively optimizing the encoder in the contrastive learning model and a set of latent variables.
no code implementations • 10 Jan 2024 • Zekun Deng, Hao Yang, Jun Wang
Some argue that the essence of humanity, such as creativity and sentiment, can never be mimicked by machines.
no code implementations • 4 Jan 2024 • Wenqi Zhang, Yongliang Shen, Linjuan Wu, Qiuying Peng, Jun Wang, Yueting Zhuang, Weiming Lu
Experiments conducted on a series of reasoning and translation tasks with different LLMs serve to underscore the effectiveness and generality of our strategy.
no code implementations • 2 Jan 2024 • Zhaoan Wang, Shaoping Xiao, Junchao Li, Jun Wang
However, our study illuminates the need for agent retraining to acquire new optimal policies under extreme weather events.
no code implementations • CVPR 2024 • Jun Wang
Starting from the initial boundary Mudslide executes a pixel-by-pixel collapse along various force directions.
no code implementations • 30 Dec 2023 • Jun Wang, Hao Ruan, Mingjie Wang, Chuanghui Zhang, Huachun Li, Jun Zhou
Over the past decade, visual gaze estimation has garnered increasing attention within the research community, owing to its wide-ranging application scenarios.
no code implementations • 27 Dec 2023 • Wei Huang, Jun Wang, Xiaoping Li, Qihang Peng
Orthogonal frequency division multiplexing (OFDM) is a widely adopted wireless communication technique but is sensitive to the carrier frequency offset (CFO).
1 code implementation • 22 Dec 2023 • Long Shi, Lei Cao, Jun Wang, Badong Chen
Specifically, we stack the data matrices from various views into the block-diagonal locations of the augmented matrix to exploit the complementary information.
no code implementations • 22 Dec 2023 • Filippos Christianos, Georgios Papoudakis, Matthieu Zimmer, Thomas Coste, Zhihao Wu, Jingxuan Chen, Khyati Khandelwal, James Doran, Xidong Feng, Jiacheng Liu, Zheng Xiong, Yicheng Luo, Jianye Hao, Kun Shao, Haitham Bou-Ammar, Jun Wang
This paper presents a general framework model for integrating and learning structured reasoning into AI agents' policies.
no code implementations • 21 Dec 2023 • Jie Han, Yixiong Zou, Haozhao Wang, Jun Wang, Wei Liu, Yao Wu, Tao Zhang, Ruixuan Li
Therefore, current works first train a model on source domains with sufficiently labeled data, and then transfer the model to target domains where only rarely labeled data is available.
no code implementations • 20 Dec 2023 • Bo Yang, Hong Peng, Chenggang Guo, Xiaohui Luo, Jun Wang, Xianzhong Long
Prompt treatment for melanoma is crucial.
no code implementations • 20 Dec 2023 • Bo Yang, Hong Peng, Xiaohui Luo, Jun Wang
Downsampling in deep networks may lead to loss of information, so for compensating the detail and edge information and allowing convolutional neural networks to pay more attention to seek the lesion region, we propose a multi-stages attention architecture based on NSNP neurons with autapses.
1 code implementation • 19 Dec 2023 • Weiyu Ma, Qirui Mi, Yongcheng Zeng, Xue Yan, Yuqiao Wu, Runji Lin, Haifeng Zhang, Jun Wang
StarCraft II is a challenging benchmark for AI agents due to the necessity of both precise micro level operations and strategic macro awareness.
no code implementations • 19 Dec 2023 • Yuang Liu, Jing Wang, Qiang Zhou, Fan Wang, Jun Wang, Wei zhang
Numerous self-supervised learning paradigms, such as contrastive learning and masked image modeling, have been proposed to acquire powerful and general representations from unlabeled data.
no code implementations • 18 Dec 2023 • Hanyu Li, Wenhan Huang, Zhijian Duan, David Henry Mguni, Kun Shao, Jun Wang, Xiaotie Deng
This paper reviews various algorithms computing the Nash equilibrium and its approximation solutions in finite normal-form games from both theoretical and empirical perspectives.