no code implementations • 19 Feb 2025 • Feiyuan Zhang, Dezhi Zhu, James Ming, Yilun Jin, Di Chai, Liu Yang, Han Tian, Zhaoxin Fan, Kai Chen
Retrieval-Augmented Generation (RAG) systems have shown substantial benefits in applications such as question answering and multi-turn dialogue \citep{lewis2020retrieval}.
no code implementations • 1 Feb 2025 • Di Chai, Pengbo Li, Feiyuan Zhang, Yilun Jin, Han Tian, Junxue Zhang, Kai Chen
Utility assessments of training TinyLlama on 15B tokens indicate that Collider sustains the utility advancements of token filtering by relatively improving model utility by 16. 3% comparing to regular training, and reduces training time from 4. 7 days to 3. 5 days using 8 GPUs.
no code implementations • 7 Jan 2025 • Xudong Liao, Yijun Sun, Han Tian, Xinchen Wan, Yilun Jin, Zilong Wang, Zhenghang Ren, Xinyang Huang, Wenxue Li, Kin Fai Tse, Zhizhen Zhong, Guyue Liu, Ying Zhang, Xiaofeng Ye, Yiming Zhang, Kai Chen
Mixture-of-Expert (MoE) models outperform conventional models by selectively activating different subnets, named \emph{experts}, on a per-token basis.
1 code implementation • 28 Oct 2024 • Yilun Jin, Zheng Li, Chenwei Zhang, Tianyu Cao, Yifan Gao, Pratik Jayarao, Mao Li, Xin Liu, Ritesh Sarkhel, Xianfeng Tang, Haodong Wang, Zhengyang Wang, Wenju Xu, Jingfeng Yang, Qingyu Yin, Xian Li, Priyanka Nigam, Yi Xu, Kai Chen, Qiang Yang, Meng Jiang, Bing Yin
Shopping MMLU consists of 57 tasks covering 4 major shopping skills: concept understanding, knowledge reasoning, user behavior alignment, and multi-linguality, and can thus comprehensively evaluate the abilities of LLMs as general shop assistants.
no code implementations • 22 Aug 2024 • Weiyan Wang, Yilun Jin, Yiming Zhang, Victor Junqiu Wei, Han Tian, Li Chen, Kai Chen
In this paper, we present Academus for low-latency online inference of BERT-like models.
no code implementations • 1 May 2024 • Liu Yang, Shuowei Cai, Di Chai, Junxue Zhang, Han Tian, Yilun Jin, Kun Guo, Kai Chen, Qiang Yang
To this core, we propose PackVFL, an efficient VFL framework based on packed HE (PackedHE), to accelerate the existing HE-based VFL algorithms.
1 code implementation • KDD 2023 • Yilun Jin, Kai Chen, Qiang Yang
To address the problem, we propose TransGTR, a transferable structure learning framework for traffic forecasting that jointly learns and transfers the graph structures and forecasting models across cities.
no code implementations • 4 Apr 2023 • Liu Yang, Di Chai, Junxue Zhang, Yilun Jin, Leye Wang, Hao liu, Han Tian, Qian Xu, Kai Chen
From the hardware layer to the vertical federated system layer, researchers contribute to various aspects of VFL.
no code implementations • 25 Mar 2023 • Yilun Jin, Yang Liu, Kai Chen, Qiang Yang
Therefore, the problem of federated learning without full labels is important in real-world FL applications.
no code implementations • 28 Jun 2022 • Shuowei Cai, Di Chai, Liu Yang, Junxue Zhang, Yilun Jin, Leye Wang, Kun Guo, Kai Chen
In this paper, we focus on SplitNN, a well-known neural network framework in VFL, and identify a trade-off between data security and model performance in SplitNN.
1 code implementation • 7 Apr 2021 • Qingqing Long, Yilun Jin, Yi Wu, Guojie Song
However, the inability of GNNs to model substructures in graphs remains a significant drawback.
no code implementations • 16 Mar 2021 • Chang Liu, Lixin Fan, Kam Woh Ng, Yilun Jin, Ce Ju, Tianyu Zhang, Chee Seng Chan, Qiang Yang
This paper proposes a novel ternary hash encoding for learning to hash methods, which provides a principled more efficient coding scheme with performances better than those of the state-of-the-art binary hashing counterparts.
no code implementations • 27 Nov 2020 • Yilun Jin, Lixin Fan, Kam Woh Ng, Ce Ju, Qiang Yang
Deep neural networks (DNNs) are known to be prone to adversarial attacks, for which many remedies are proposed.
no code implementations • 24 Sep 2020 • Junshan Wang, Yilun Jin, Guojie Song, Xiaojun Ma
In this paper, we propose EPNE, a temporal network embedding model preserving evolutionary patterns of the local structure of nodes.
1 code implementation • 25 Jun 2020 • Qingqing Long, Yilun Jin, Guojie Song, Yi Li, Wei. Lin
Specifically, we build topic models upon graphs using anonymous walks and Graph Anchor LDA, an LDA variant that selects significant structural patterns first, so as to alleviate the complexity and generate structural topics efficiently.
no code implementations • 26 Feb 2020 • Yilun Jin, Xiguang Wei, Yang Liu, Qiang Yang
Federated Learning (FL) proposed in recent years has received significant attention from researchers in that it can bring separate data sources together and build machine learning models in a collaborative but private manner.
no code implementations • 18 Nov 2019 • Yilun Jin, Guojie Song, Chuan Shi
Specifically, we capture local graph structures via random anonymous walks, powerful and flexible tools that represent structural patterns.
2 code implementations • 3 Jun 2019 • Yizhou Zhang, Guojie Song, Lun Du, Shu-wen Yang, Yilun Jin
Recent works reveal that network embedding techniques enable many machine learning models to handle diverse downstream tasks on graph structured data.
1 code implementation • 25 Jan 2019 • Kewei Cheng, Tao Fan, Yilun Jin, Yang Liu, Tianjian Chen, Dimitrios Papadopoulos, Qiang Yang
This federated learning system allows the learning process to be jointly conducted over multiple parties with common user samples but different feature sets, which corresponds to a vertically partitioned data set.
no code implementations • 4 Jan 2018 • Yi Zhang, Houjun Huang, Haifeng Zhang, Liao Ni, Wei Xu, Nasir Uddin Ahmed, Md. Shakil Ahmed, Yilun Jin, Yingjie Chen, Jingxuan Wen, Wenxin Li
The development of finger vein recognition algorithms heavily depends on large-scale real-world data sets.