Search Results for author: Reynold Cheng

Found 15 papers, 10 papers with code

Unlocking Continual Learning Abilities in Language Models

1 code implementation25 Jun 2024 Wenyu Du, Shuang Cheng, Tongxu Luo, Zihan Qiu, Zeyu Huang, Ka Chun Cheung, Reynold Cheng, Jie Fu

To address this limitation, we introduce $\textbf{MIGU}$ ($\textbf{M}$agn$\textbf{I}$tude-based $\textbf{G}$radient $\textbf{U}$pdating for continual learning), a rehearsal-free and task-label-free method that only updates the model parameters with large magnitudes of output in LMs' linear layers.

Continual Learning Inductive Bias

Before Generation, Align it! A Novel and Effective Strategy for Mitigating Hallucinations in Text-to-SQL Generation

1 code implementation24 May 2024 Ge Qu, Jinyang Li, Bowen Li, Bowen Qin, Nan Huo, Chenhao Ma, Reynold Cheng

Despite these advancements, the inherent bad nature of the generalization of LLMs often results in hallucinations, which limits the full potential of LLMs.

In-Context Learning Text-To-SQL

Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training

no code implementations24 May 2024 Wenyu Du, Tongxu Luo, Zihan Qiu, Zeyu Huang, Yikang Shen, Reynold Cheng, Yike Guo, Jie Fu

For example, compared to a conventionally trained 7B model using 300B tokens, our $G_{\text{stack}}$ model converges to the same loss with 194B tokens, resulting in a 54. 6\% speedup.

Multi-order Graph Clustering with Adaptive Node-level Weight Learning

1 code implementation20 May 2024 Ye Liu, Xuelei Lin, Yejia Chen, Reynold Cheng

In this paper, we propose a multi-order graph clustering model (MOGC) to integrate multiple higher-order structures and edge connections at node level.

Clustering Graph Clustering

A Sampling-based Framework for Hypothesis Testing on Large Attributed Graphs

1 code implementation20 Mar 2024 Yun Wang, Chrysanthi Kosyfaki, Sihem Amer-Yahia, Reynold Cheng

Experiments on real datasets demonstrate the ability of our framework to leverage common graph sampling methods for hypothesis testing, and the superiority of hypothesis-aware sampling in terms of accuracy and time efficiency.

Graph Sampling

Tapilot-Crossing: Benchmarking and Evolving LLMs Towards Interactive Data Analysis Agents

1 code implementation8 Mar 2024 Jinyang Li, Nan Huo, Yan Gao, Jiayi Shi, Yingxiu Zhao, Ge Qu, Yurong Wu, Chenhao Ma, Jian-Guang Lou, Reynold Cheng

The challenges and costs of collecting realistic interactive logs for data analysis hinder the quantitative evaluation of Large Language Model (LLM) agents in this task.

Benchmarking Decision Making +2

A Survey on Knowledge Distillation of Large Language Models

1 code implementation20 Feb 2024 Xiaohan Xu, Ming Li, Chongyang Tao, Tao Shen, Reynold Cheng, Jinyang Li, Can Xu, DaCheng Tao, Tianyi Zhou

In the era of Large Language Models (LLMs), Knowledge Distillation (KD) emerges as a pivotal methodology for transferring advanced capabilities from leading proprietary LLMs, such as GPT-4, to their open-source counterparts like LLaMA and Mistral.

Data Augmentation Knowledge Distillation +2

Debiasing Recommendation with Personal Popularity

1 code implementation12 Feb 2024 Wentao Ning, Reynold Cheng, Xiao Yan, Ben Kao, Nan Huo, Nur AI Hasan Haldar, Bo Tang

Many methods have been proposed to reduce GP bias but they fail to notice the fundamental problem of GP, i. e., it considers popularity from a \textit{global} perspective of \textit{all users} and uses a single set of popular items, and thus cannot capture the interests of individual users.

counterfactual Counterfactual Inference

Multi-domain Recommendation with Embedding Disentangling and Domain Alignment

1 code implementation10 Aug 2023 Wentao Ning, Xiao Yan, Weiwen Liu, Reynold Cheng, Rui Zhang, Bo Tang

We propose a new MDR method named EDDA with two key components, i. e., embedding disentangling recommender and domain alignment, to tackle the two challenges respectively.

Transfer Learning

Can LLM Already Serve as A Database Interface? A BIg Bench for Large-Scale Database Grounded Text-to-SQLs

no code implementations NeurIPS 2023 Jinyang Li, Binyuan Hui, Ge Qu, Jiaxi Yang, Binhua Li, Bowen Li, Bailin Wang, Bowen Qin, Rongyu Cao, Ruiying Geng, Nan Huo, Xuanhe Zhou, Chenhao Ma, Guoliang Li, Kevin C. C. Chang, Fei Huang, Reynold Cheng, Yongbin Li

Our emphasis on database values highlights the new challenges of dirty database contents, external knowledge between NL questions and database contents, and SQL efficiency, particularly in the context of massive databases.

SQL Parsing Text-To-SQL

Graphix-T5: Mixing Pre-Trained Transformers with Graph-Aware Layers for Text-to-SQL Parsing

1 code implementation18 Jan 2023 Jinyang Li, Binyuan Hui, Reynold Cheng, Bowen Qin, Chenhao Ma, Nan Huo, Fei Huang, Wenyu Du, Luo Si, Yongbin Li

Recently, the pre-trained text-to-text transformer model, namely T5, though not specialized for text-to-SQL parsing, has achieved state-of-the-art performance on standard benchmarks targeting domain generalization.

Domain Generalization Inductive Bias +2

A Survey on Machine Learning Solutions for Graph Pattern Extraction

1 code implementation3 Apr 2022 Kai Siong Yow, Ningyi Liao, Siqiang Luo, Reynold Cheng, Chenhao Ma, Xiaolin Han

Many algorithms are proposed in studying subgraph problems, where one common approach is by extracting the patterns and structures of a given graph.

Community Detection Community Search +1

An End-to-End Deep RL Framework for Task Arrangement in Crowdsourcing Platforms

no code implementations4 Nov 2019 Caihua Shan, Nikos Mamoulis, Reynold Cheng, Guoliang Li, Xiang Li, Yuqiu Qian

In this paper, we propose a Deep Reinforcement Learning (RL) framework for task arrangement, which is a critical problem for the success of crowdsourcing platforms.

Deep Reinforcement Learning Reinforcement Learning (RL)

A General Early-Stopping Module for Crowdsourced Ranking

no code implementations4 Nov 2019 Caihua Shan, Leong Hou U, Nikos Mamoulis, Reynold Cheng, Xiang Li

The number of microtasks depends on the budget allocated for the problem.

Cannot find the paper you are looking for? You can Submit a new open access paper.