no code implementations • 29 Apr 2025 • Zechuan Zhang, Ji Xie, Yu Lu, Zongxin Yang, Yi Yang
Instruction-based image editing enables robust image modification via natural language prompts, yet current methods face a precision-efficiency tradeoff.
1 code implementation • 23 Mar 2025 • Yang Luo, Shiru Wang, Jun Liu, Jiaxuan Xiao, Rundong Xue, Zeyu Zhang, Hao Zhang, Yu Lu, Yang Zhao, Yutong Xie
Breast cancer survival prediction in computational pathology presents a remarkable challenge due to tumor heterogeneity.
no code implementations • 3 Mar 2025 • Zitang Zhou, Ke Mei, Yu Lu, Tianyi Wang, Fengyun Rao
This paper introduces HarmonySet, a comprehensive dataset designed to advance video-music understanding.
no code implementations • 26 Feb 2025 • Che Liu, Yingji Zhang, Dong Zhang, Weijie Zhang, Chenggong Gong, Haohan Li, Yu Lu, Shilin Zhou, Yue Lu, Ziliang Gan, Ziao Wang, Junwei Liao, Haipang Wu, Ji Liu, André Freitas, Qifan Wang, Zenglin Xu, Rongjuncheng Zhang, Yong Dai
For the third question, we design the speech data synthesis pipeline to obtain high-quality speech training datasets, covering various real-world scenarios.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+3
1 code implementation • 30 Dec 2024 • Zijie Chen, Zhanchao Zhou, Yu Lu, Renjun Xu, Lili Pan, Zhenzhong Lan
Solving NP-hard problems traditionally relies on heuristics, yet manually designing effective heuristics for complex problems remains a significant challenge.
no code implementations • 23 Dec 2024 • Yu Lu, Jiayi Zhang, Enyu Shi, Peng Zhang, Derrick Wing Kwan Ng, Dusit Niyato, Bo Ai
Reconfigurable intelligent surfaces (RIS)-assisted cell-free massive multiple-input multiple-output (CF mMIMO) systems have emerged as a promising technology for sixth-generation communication systems.
no code implementations • 26 Nov 2024 • Liyun Zhang, Dian Ding, Yu Lu, Yi-Chao Chen, Guangtao Xue
In this paper, we present a framework, Lantern, that can improve the performance of a certain vanilla model by prompting large language models with receptive-field-aware attention weighting.
no code implementations • 5 Nov 2024 • Ziliang Gan, Yu Lu, Dong Zhang, Haohan Li, Che Liu, Jian Liu, Ji Liu, Haipang Wu, Chaoyou Fu, Zenglin Xu, Rongjunchen Zhang, Yong Dai
In recent years, multimodal benchmarks for general domains have guided the rapid development of multimodal models on general tasks.
no code implementations • 18 Oct 2024 • Xiang Hu, Hongyu Fu, Jinge Wang, Yifeng Wang, Zhikun Li, Renjun Xu, Yu Lu, Yaochu Jin, Lili Pan, Zhenzhong Lan
Scientific innovation is pivotal for humanity, and harnessing large language models (LLMs) to generate research ideas could transform discovery.
no code implementations • 18 Sep 2024 • Jiongnan Liu, Yutao Zhu, Shuting Wang, Xiaochi Wei, Erxue Min, Yu Lu, Shuaiqiang Wang, Dawei Yin, Zhicheng Dou
By attaching this embedding to the task input, LLMs can better understand and capture user habits and preferences, thereby producing more personalized outputs without tuning their own parameters.
no code implementations • 29 Aug 2024 • Yu Lu, Roummel F. Marcia
The negative binomial model, which generalizes the Poisson distribution model, can be found in applications involving low-photon signal recovery, including medical imaging.
no code implementations • 28 Aug 2024 • Yu Lu, Kevin Bui, Roummel F. Marcia
In our experiments, we demonstrate that the NB model outperforms Poisson matrix completion in various noise and missing data settings on real data.
no code implementations • 28 Aug 2024 • Yu Lu, Kevin Bui, Roummel F. Marcia
However, this model assumes that the mean and variance of the signal's noise distribution are equal.
no code implementations • 20 Aug 2024 • Yu Lu, Yuyu Chen, Gang Zhou, Zhenghua Lan
To the best of our knowledge, it is the first publicly available eight-categories 3D point cloud dataset in the field of textile pilling assessment.
no code implementations • 15 Aug 2024 • Xihong Yang, Heming Jing, Zixing Zhang, Jindong Wang, Huakang Niu, Shuaiqiang Wang, Yu Lu, Junfeng Wang, Dawei Yin, Xinwang Liu, En Zhu, Defu Lian, Erxue Min
In this work, we prove that directly aligning the representations of LLMs and collaborative models is sub-optimal for enhancing downstream recommendation tasks performance, based on the information theorem.
no code implementations • 29 Jul 2024 • Yu Lu, Yuanzhi Liang, Linchao Zhu, Yi Yang
FreeLong blends the low-frequency components of global video features, which encapsulate the entire video sequence, with the high-frequency components of local video features that focus on shorter subsequences of frames.
no code implementations • 25 Jun 2024 • Xin Yang, Heng Chang, Zhijian Lai, Jinze Yang, Xingrun Li, Yu Lu, Shuaiqiang Wang, Dawei Yin, Erxue Min
Cross-Domain Recommendation (CDR) seeks to utilize knowledge from different domains to alleviate the problem of data sparsity in the target recommendation domain, and it has been gaining more attention in recent years.
1 code implementation • 4 Jun 2024 • Zhiwei Cao, Qian Cao, Yu Lu, Ningxin Peng, Luyang Huang, Shanbo Cheng, Jinsong Su
This decline can be attributed to the loss of key information during the compression process.
1 code implementation • 21 May 2024 • Xingyuan Pan, Luyang Huang, Liyan Kang, Zhicheng Liu, Yu Lu, Shanbo Cheng
Moreover, to enhance the diversity of the training data we maximize the variety of influences they have on the model by clustering on their gradients and resampling.
1 code implementation • 18 May 2024 • Mengxi Zhang, Wenhao Wu, Yu Lu, Yuxin Song, Kang Rong, Huanjin Yao, Jianbo Zhao, Fanglong Liu, Yifan Sun, Haocheng Feng, Jingdong Wang
To verify our viewpoint, we present the Automated Multi-level Preference (AMP) framework for MLLMs.
no code implementations • 19 Feb 2024 • Anqi Li, Yu Lu, Nirui Song, Shuai Zhang, Lizhi Ma, Zhenzhong Lan
Through further LLM-based evaluations on additional conversations, our findings underscore the challenges counselors face in cultivating strong online relationships with clients.
no code implementations • 18 Feb 2024 • Shuai Zhang, Yu Lu, Junwen Liu, JIA YU, Huachuan Qiu, Yuming Yan, Zhenzhong Lan
With the growing humanlike nature of dialog agents, people are now engaging in extended conversations that can stretch from brief moments to substantial periods of time.
no code implementations • 27 Nov 2023 • Yu Lu, Linchao Zhu, Hehe Fan, Yi Yang
Text-to-video (T2V) generation is a rapidly growing research area that aims to translate the scenes, objects, and actions within complex video text into a sequence of coherent visual frames.
no code implementations • CVPR 2024 • Wenshuai Xu, Zhenghui Hu, Yu Lu, Jinzhou Meng, Qingjie Liu, Yunhong Wang
Firstly, we select samples for annotation by optimizing the distribution similarity between the subset to be selected and the entire unlabeled pool in continuous space.
no code implementations • 24 Sep 2023 • Jonghyuk Baek, Yanran Wang, Xiaolong He, Yu Lu, John S. McCartney, J. S. Chen
In deep geological repositories for high level nuclear waste with close canister spacings, bentonite buffers can experience temperatures higher than 100 {\deg}C. In this range of extreme temperatures, phenomenological constitutive laws face limitations in capturing the thermo-hydro-mechanical (THM) behavior of the bentonite, since the pre-defined functional constitutive laws often lack generality and flexibility to capture a wide range of complex coupling phenomena as well as the effects of stress state and path dependency.
no code implementations • 8 Aug 2023 • Richard Jiarui Tong, Cassie Chen Cao, Timothy Xueqian Lee, Guodong Zhao, Ray Wan, FeiYue Wang, Xiangen Hu, Robin Schmucker, Jinsheng Pan, Julian Quevedo, Yu Lu
This paper presents the Never Ending Open Learning Adaptive Framework (NEOLAF), an integrated neural-symbolic cognitive architecture that models and constructs intelligent agents.
no code implementations • 16 Jun 2023 • Yu Lu, Junwei Bao, Zichen Ma, Xiaoguang Han, Youzheng Wu, Shuguang Cui, Xiaodong He
High-quality data is essential for conversational recommendation systems and serves as the cornerstone of the network architecture development and training strategy design.
no code implementations • 2 Jun 2023 • Canjia Li, Xiaoyang Wang, Dongdong Li, Yiding Liu, Yu Lu, Shuaiqiang Wang, Zhicong Cheng, Simiu Gu, Dawei Yin
In this work, we focus on ranking user satisfaction rather than relevance in web search, and propose a PLM-based framework, namely SAT-Ranker, which comprehensively models different dimensions of user satisfaction in a unified manner.
no code implementations • 5 Oct 2022 • Yu Lu, Yi-Jia Wang, Ying Chen, Jia-Jun Wu
We present that by predicting the spectrum in discrete space from the phase shift in continuous space, the neural network can remarkably reproduce the numerical L\"uscher's formula to a high precision.
no code implementations • 12 Aug 2022 • Zichen Ma, Yu Lu, Wenye Li, Shuguang Cui
This dynamically personalized FL technique incentivizes clients to participate in personalizing local models while allowing the adoption of the global model when it performs better.
no code implementations • 20 May 2022 • Qingzhong Wang, Haifang Li, Haoyi Xiong, Wen Wang, Jiang Bian, Yu Lu, Shuaiqiang Wang, Zhicong Cheng, Dejing Dou, Dawei Yin
To handle the diverse query requests from users at web-scale, Baidu has done tremendous efforts in understanding users' queries, retrieve relevant contents from a pool of trillions of webpages, and rank the most relevant webpages on the top of results.
1 code implementation • 15 Apr 2022 • Gang Chen, Yu Lu, Rong Su, Zhaodan Kong
Machine learning-based methods have achieved successful applications in machinery fault diagnosis.
1 code implementation • ACL 2022 • Yu Lu, Jiali Zeng, Jiajun Zhang, Shuangzhi Wu, Mu Li
Confidence estimation aims to quantify the confidence of the model prediction, providing an expectation of success.
no code implementations • 10 Dec 2021 • Zichen Ma, Zihan Lu, Yu Lu, Wenye Li, JinFeng Yi, Shuguang Cui
In this paper, we design a federated two-stage learning framework that augments prototypical federated learning with a cut layer on devices and uses sign-based stochastic gradient descent with the majority vote method on model updates.
1 code implementation • CVPR 2022 • Zhaoqing Wang, Yu Lu, Qiang Li, Xunqiang Tao, Yandong Guo, Mingming Gong, Tongliang Liu
In addition, we present text-to-pixel contrastive learning to explicitly enforce the text feature similar to the related pixel-level features and dissimilar to the irrelevances.
no code implementations • 31 Oct 2021 • DeLiang Wang, Yu Lu, Qinggang Meng, Penghe Chen
With more deep learning techniques being introduced into the knowledge tracing domain, the interpretability issue of the knowledge tracing models has aroused researchers' attention.
no code implementations • ACL 2021 • Yu Lu, Jiali Zeng, Jiajun Zhang, Shuangzhi Wu, Mu Li
Attention mechanisms have achieved substantial improvements in neural machine translation by dynamically selecting relevant inputs for different predictions.
no code implementations • 17 Jun 2021 • Zichen Ma, Yu Lu, Zihan Lu, Wenye Li, JinFeng Yi, Shuguang Cui
Training in heterogeneous and potentially massive networks introduces bias into the system, which is originated from the non-IID data and the low participation rate in reality.
1 code implementation • 8 Jun 2021 • Zixuan Peng, Yu Lu, Shengfeng Pan, Yunfeng Liu
Emotion recognition from speech is a challenging task.
1 code implementation • Findings (ACL) 2021 • Yu Lu, Junwei Bao, Yan Song, Zichen Ma, Shuguang Cui, Youzheng Wu, Xiaodong He
Existing conversational recommendation (CR) systems usually suffer from insufficient item information when conducted on short dialogue history and unfamiliar items.
21 code implementations • 20 Apr 2021 • Jianlin Su, Yu Lu, Shengfeng Pan, Ahmed Murtadha, Bo Wen, Yunfeng Liu
Then, we propose a novel method named Rotary Position Embedding(RoPE) to effectively leverage the positional information.
Ranked #1 on
Semantic Text Matching
on CAIL2019-SCM - test
1 code implementation • ECCV 2020 • Tianyi Wu, Yu Lu, Yu Zhu, Chuang Zhang, Ming Wu, Zhanyu Ma, Guodong Guo
GI unit is further improved by the SC-loss to enhance the semantic representations over the exemplar-based semantic graph.
no code implementations • WS 2020 • Qian Wang, Yuchen Liu, Cong Ma, Yu Lu, Yining Wang, Long Zhou, Yang Zhao, Jiajun Zhang, Cheng-qing Zong
This paper describes the CASIA{'}s system for the IWSLT 2020 open domain translation task.
no code implementations • 13 May 2020 • Yu Lu, DeLiang Wang, Qinggang Meng, Penghe Chen
We thus propose to adopt the post-hoc method to tackle the interpretability issue for deep learning based knowledge tracing (DLKT) models.
1 code implementation • 4 Feb 2020 • Chenhao Lin, Siwen Wang, Dongqi Xu, Yu Lu, Wayne Zhang
Weakly supervised object detection (WSOD) using only image-level annotations has attracted growing attention over the past few years.
Ranked #16 on
Weakly Supervised Object Detection
on PASCAL VOC 2007
1 code implementation • 31 Jan 2020 • Yu Lu, Muyan Feng, Ming Wu, Chuang Zhang
Human parsing is an essential branch of semantic segmentation, which is a fine-grained semantic segmentation task to identify the constituent parts of human.
no code implementations • 26 Nov 2019 • Pingchuan Ma, Yao Zhou, Yu Lu, Wei zhang
To this end, we propose the video shuffle, a parameter-free plug-in component that efficiently reallocates the inputs of 2D convolution so that its receptive field can be extended to the temporal dimension.
1 code implementation • 25 Apr 2018 • Huan Wang, Qiming Zhang, Yuehai Wang, Yu Lu, Haoji Hu
Parameter pruning is a promising approach for CNN compression and acceleration by eliminating redundant model parameters with tolerable performance degrade.
no code implementations • ICML 2017 • Lihong Li, Yu Lu, Dengyong Zhou
Contextual bandits are widely used in Internet services from news recommendation to advertising, and to Web search.
no code implementations • 7 Dec 2016 • Yu Lu, Harrison H. Zhou
Lloyd's algorithm, proposed in 1957, is still possibly the most widely used clustering algorithm in practice due to its simplicity and empirical performance.
no code implementations • 25 May 2016 • Chao Gao, Yu Lu, Dengyong Zhou
In many machine learning applications, crowdsourcing has become the primary means for label collection.
no code implementations • 1 Dec 2015 • Chao Gao, Yu Lu, Zongming Ma, Harrison H. Zhou
Biclustering structures in data matrices were first formalized in a seminal paper by John Hartigan (1972) where one seeks to cluster cases and variables simultaneously.
no code implementations • 18 May 2015 • Yu Lu, Liu Chuang, Zhang Zi-Ke
Recommender systems, which can significantly help users find their interested items from the information era, has attracted an increasing attention from both the scientific and application society.
no code implementations • 9 Apr 2015 • Xiangru Li, Yu Lu, Georges Comte, Ali Luo, Yongheng Zhao, Yongjun Wang
On real spectra, we extracted 23 features to estimate $T_{eff}$, 62 features for log$~g$, and 68 features for [Fe/H].
no code implementations • 3 Oct 2014 • Yu Lu, Sahand N. Negahban
In recent years rank aggregation has received significant attention from the machine learning community.