no code implementations • 14 Mar 2024 • Kai Xiong, Xiao Ding, Ting Liu, Bing Qin, Dongliang Xu, Qing Yang, Hongtao Liu, Yixin Cao
Large language models (LLMs) have developed impressive performance and strong explainability across various reasoning scenarios, marking a significant stride towards mimicking human-like intelligence.
no code implementations • 28 Dec 2023 • Liang Zhao, Xiaocheng Feng, Xiachong Feng, Dongliang Xu, Qing Yang, Hongtao Liu, Bing Qin, Ting Liu
In this survey, we present these advances towards length extrapolation in a unified notation from the perspective of PE.
no code implementations • 19 Dec 2023 • Qiyao Peng, Hongtao Liu, Hongyan Xu, Yinghui Wang, Wenjun Wang
For alleviating this, we present a personalized pre-training and fine-tuning paradigm, which could effectively learn expert interest and expertise simultaneously.
1 code implementation • 25 Apr 2023 • Guangyuan Ma, Hongtao Liu, Xing Wu, Wanhui Qian, Zhepeng Lv, Qing Yang, Songlin Hu
Firstly, we introduce the user behavior masking pre-training task to recover the masked user behaviors based on their contextual behaviors.
1 code implementation • 30 Aug 2022 • Yong Zhong, Hongtao Liu, Xiaodong Liu, Fan Bao, Weiran Shen, Chongxuan Li
Deep generative models (DGMs) are data-eager because learning a complex model on limited data suffers from a large variance and easily overfits.
no code implementations • 14 Oct 2019 • Hongtao Liu
This research collects the data of the related political ads in the context of the U. S. midterm elections since August to study the overall pattern of political ads on social media and uses sets of machine learning methods to conduct sentiment analysis on these ads to classify the negative ads.
no code implementations • 15 Sep 2019 • Qianyu Guo, Sen Chen, Xiaofei Xie, Lei Ma, Qiang Hu, Hongtao Liu, Yang Liu, Jianjun Zhao, Xiaohong Li
However, the differences in architecture designs and implementations of existing frameworks and platforms bring new challenges for DL software development and deployment.
no code implementations • 29 May 2019 • Xianchen Wang, Hongtao Liu, Peiyi Wang, Fangzhao Wu, Hongyan Xu, Wenjun Wang, Xing Xie
In this paper, we propose a hierarchical attention model fusing latent factor model for rating prediction with reviews, which can focus on important words and informative reviews.
5 code implementations • 29 May 2019 • Hongtao Liu, Fangzhao Wu, Wenjun Wang, Xianchen Wang, Pengfei Jiao, Chuhan Wu, Xing Xie
In this paper we propose a neural recommendation approach with personalized attention to learn personalized representations of users and items from reviews.