Search Results for author: Bin Cui

Found 109 papers, 63 papers with code

Don't Waste Your Bits! Squeeze Activations and Gradients for Deep Neural Networks via TinyScript

no code implementations ICML 2020 Fangcheng Fu, Yuzheng Hu, Yihan He, Jiawei Jiang, Yingxia Shao, Ce Zhang, Bin Cui

Recent years have witnessed intensive research interests on training deep neural networks (DNNs) more efficiently by quantization-based compression methods, which facilitate DNNs training in two ways: (1) activations are quantized to shrink the memory consumption, and (2) gradients are quantized to decrease the communication cost.

Quantization

VersaTune: An Efficient Data Composition Framework for Training Multi-Capability LLMs

no code implementations18 Nov 2024 Keer Lu, Keshi Zhao, Zheng Liang, Da Pan, Shusen Zhang, Xin Wu, WeiPeng Chen, Zenan Zhou, Guosheng Dong, Bin Cui, Wentao Zhang

Despite their potential, existing work mainly focuses on domain-specific enhancements during fine-tuning, the challenge of which lies in catastrophic forgetting of knowledge across other domains.

Nova: A Practical and Advanced Alignment

no code implementations19 Oct 2024 MingAn Lin, Fan Yang, Yanjun Shen, Haoze Sun, Tianpeng Li, Chenzheng Zhu, Tao Zhang, Miao Zheng, Xu Li, Yijie Zhou, Mingyang Chen, Yanzhao Qin, Youquan Li, Hao Liang, Fei Li, Yadong Li, Mang Wang, Guosheng Dong, Kun Fang, Jianhua Xu, Bin Cui, Wentao Zhang, Zenan Zhou, WeiPeng Chen

Importantly, Qwen2-Nova-72B and Llama3-PBM-Nova-70B are instruct versions of the Qwen2-72B and Llama-3-70B base models, optimized through Nova.

Facilitating Multi-turn Function Calling for LLMs via Compositional Instruction Tuning

1 code implementation16 Oct 2024 Mingyang Chen, Haoze Sun, Tianpeng Li, Fan Yang, Hao Liang, Keer Lu, Bin Cui, Wentao Zhang, Zenan Zhou, WeiPeng Chen

While current research on function calling by LLMs primarily focuses on single-turn interactions, this paper addresses the overlooked necessity for LLMs to engage in multi-turn function calling--critical for handling compositional, real-world queries that require planning with functions but not only use functions.

8k

SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights

1 code implementation11 Oct 2024 Ling Yang, Zhaochen Yu, Tianjun Zhang, Minkai Xu, Joseph E. Gonzalez, Bin Cui, Shuicheng Yan

Large language models (LLMs) like GPT-4, PaLM, and LLaMA have shown significant improvements in various reasoning tasks.

GSM8K Math +1

Universal Medical Image Representation Learning with Compositional Decoders

no code implementations30 Sep 2024 Kaini Wang, Ling Yang, Siping Zhou, Guangquan Zhou, Wentao Zhang, Bin Cui, Shuo Li

To address these challenges, we have developed a decomposed-composed universal medical imaging paradigm (UniMed) that supports tasks at all levels.

Decoder Representation Learning

Data Proportion Detection for Optimized Data Management for Large Language Models

1 code implementation26 Sep 2024 Hao Liang, Keshi Zhao, Yajie Yang, Bin Cui, Guosheng Dong, Zenan Zhou, Wentao Zhang

Large language models (LLMs) have demonstrated exceptional performance across a wide range of tasks and domains, with data preparation playing a critical role in achieving these results.

Management

Efficient Multi-Task Large Model Training via Data Heterogeneity-aware Model Management

no code implementations5 Sep 2024 Yujie Wang, Shenhan Zhu, Fangcheng Fu, Xupeng Miao, Jie Zhang, Juan Zhu, Fan Hong, Yong Li, Bin Cui

In this paper, we investigate how to achieve high-performance training of large-scale MT MM models through data heterogeneity-aware model management optimization.

Management Scheduling

DataSculpt: Crafting Data Landscapes for Long-Context LLMs through Multi-Objective Partitioning

1 code implementation2 Sep 2024 Keer Lu, Xiaonan Nie, Zheng Liang, Da Pan, Shusen Zhang, Keshi Zhao, WeiPeng Chen, Zenan Zhou, Guosheng Dong, Bin Cui, Wentao Zhang

Through extensive experimental analysis, we identified three key challenges in designing effective data management strategies that enable the model to achieve long-context capability without sacrificing performance in other tasks: (1) a shortage of long documents across multiple domains, (2) effective construction of context windows, and (3) efficient organization of large-scale datasets.

Code Completion Combinatorial Optimization +5

SysBench: Can Large Language Models Follow System Messages?

1 code implementation20 Aug 2024 Yanzhao Qin, Tao Zhang, Yanjun Shen, Wenjing Luo, Haoze Sun, Yan Zhang, Yujing Qiao, WeiPeng Chen, Zenan Zhou, Wentao Zhang, Bin Cui

Finally, we conduct extensive evaluation across various existing LLMs, measuring their ability to follow specified constraints given in system messages.

CFBench: A Comprehensive Constraints-Following Benchmark for LLMs

1 code implementation2 Aug 2024 Yanjun Shen, Wenjing Luo, Yan Zhang, Hao Liang, Tao Zhang, Fan Yang, MingAn Lin, Yujing Qiao, WeiPeng Chen, Bin Cui, Wentao Zhang, Zenan Zhou

The adeptness of Large Language Models (LLMs) in comprehending and following natural language instructions is critical for their deployment in sophisticated real-world applications.

SynthVLM: High-Efficiency and High-Quality Synthetic Data for Vision Language Models

1 code implementation30 Jul 2024 Zheng Liu, Hao Liang, Xijie Huang, Wentao Xiong, Qinhan Yu, Linzhuang Sun, Chong Chen, Conghui He, Bin Cui, Wentao Zhang

Crucially, our method's reliance on purely generated data ensures the preservation of privacy, achieving SoTA performance with just 100k data points (only 18% of the official dataset size).

Caption Generation Question Answering

Efficiently Training 7B LLM with 1 Million Sequence Length on 8 GPUs

no code implementations16 Jul 2024 Pinxue Zhao, Hailin Zhang, Fangcheng Fu, Xiaonan Nie, Qibin Liu, Fang Yang, Yuanbo Peng, Dian Jiao, Shuaipeng Li, Jinbao Xue, Yangyu Tao, Bin Cui

By leveraging fine-grained activation memory management, MEMO facilitates efficient training of 7B LLM with 1 million sequence length on just 8 A800 GPUs, achieving an MFU of 52. 30%.

Management

PAS: Data-Efficient Plug-and-Play Prompt Augmentation System

no code implementations8 Jul 2024 Miao Zheng, Hao Liang, Fan Yang, Haoze Sun, Tianpeng Li, Lingchu Xiong, Yan Zhang, Youzhen Wu, Kun Li, Yanjun Shen, MingAn Lin, Tao Zhang, Guosheng Dong, Yujing Qiao, Kun Fang, WeiPeng Chen, Bin Cui, Wentao Zhang, Zenan Zhou

This combination of high performance, efficiency, and flexibility makes PAS a valuable system for enhancing the usability and effectiveness of LLMs through improved prompt engineering.

Prompt Engineering

KeyVideoLLM: Towards Large-scale Video Keyframe Selection

no code implementations3 Jul 2024 Hao Liang, Jiapeng Li, Tianyi Bai, Xijie Huang, Linzhuang Sun, Zhengren Wang, Conghui He, Bin Cui, Chong Chen, Wentao Zhang

Recently, with the rise of web videos, managing and understanding large-scale video datasets has become increasingly important.

Data Compression Management +3

Efficient-Empathy: Towards Efficient and Effective Selection of Empathy Data

no code implementations2 Jul 2024 Linzhuang Sun, Hao Liang, Jingxuan Wei, Linkun Sun, Bihui Yu, Bin Cui, Wentao Zhang

By integrating sensibility and rationality data with a MoE structure, we achieve even higher performance, demonstrating the effectiveness of our Efficient-Empathy algorithm.

Consistency Flow Matching: Defining Straight Flows with Velocity Consistency

1 code implementation2 Jul 2024 Ling Yang, Zixiang Zhang, Zhilong Zhang, Xingchao Liu, Minkai Xu, Wentao Zhang, Chenlin Meng, Stefano Ermon, Bin Cui

Additionally, we propose a multi-segment training approach for Consistency-FM to enhance expressiveness, achieving a better trade-off between sampling quality and speed.

Image Generation

RobGC: Towards Robust Graph Condensation

no code implementations19 Jun 2024 Xinyi Gao, Hongzhi Yin, Tong Chen, Guanhua Ye, Wentao Zhang, Bin Cui

Graph neural networks (GNNs) have attracted widespread attention for their impressive capability of graph representation learning.

Denoising Graph Representation Learning

Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models

1 code implementation6 Jun 2024 Ling Yang, Zhaochen Yu, Tianjun Zhang, Shiyi Cao, Minkai Xu, Wentao Zhang, Joseph E. Gonzalez, Bin Cui

We introduce Buffer of Thoughts (BoT), a novel and versatile thought-augmented reasoning approach for enhancing accuracy, efficiency and robustness of large language models (LLMs).

Arithmetic Reasoning Code Generation +2

A Survey of Multimodal Large Language Model from A Data-centric Perspective

1 code implementation26 May 2024 Tianyi Bai, Hao Liang, Binwang Wan, Yanran Xu, Xi Li, Shiyu Li, Ling Yang, Bozhou Li, Yifan Wang, Bin Cui, Ping Huang, Jiulong Shan, Conghui He, Binhang Yuan, Wentao Zhang

Multimodal large language models (MLLMs) enhance the capabilities of standard large language models by integrating and processing data from multiple modalities, including text, vision, audio, video, and 3D environments.

Language Modelling Large Language Model +2

Surge Phenomenon in Optimal Learning Rate and Batch Size Scaling

no code implementations23 May 2024 Shuaipeng Li, Penghao Zhao, Hailin Zhang, Xingwu Sun, Hao Wu, Dian Jiao, Weiyan Wang, Chengjun Liu, Zheng Fang, Jinbao Xue, Yangyu Tao, Bin Cui, Di Wang

First, we raise the scaling law between batch sizes and optimal learning rates in the sign of gradient case, in which we prove that the optimal learning rate first rises and then falls as the batch size increases.

Acceleration Algorithms in GNNs: A Survey

1 code implementation7 May 2024 Lu Ma, Zeang Sheng, Xunkai Li, Xinyi Gao, Zhezheng Hao, Ling Yang, Wentao Zhang, Bin Cui

To address the critical challenges, a range of algorithms have been proposed to accelerate training and inference of GNNs, attracting increasing attention from the research community.

Graph Learning Survey

Clover: Regressive Lightweight Speculative Decoding with Sequential Knowledge

1 code implementation1 May 2024 Bin Xiao, Chunan Shi, Xiaonan Nie, Fan Yang, Xiangwei Deng, Lei Su, WeiPeng Chen, Bin Cui

Consequently, the GPU spends most of its time on memory transfer instead of computation.

SqueezeAttention: 2D Management of KV-Cache in LLM Inference via Layer-wise Optimal Budget

1 code implementation7 Apr 2024 ZiHao Wang, Bin Cui, Shaoduo Gan

In this work, we found that by identifying the importance of attention layers, we could optimize the KV-cache jointly from two dimensions, i. e., sequence-wise and layer-wise.

Language Modelling Large Language Model +1

LIST: Learning to Index Spatio-Textual Data for Embedding based Spatial Keyword Queries

no code implementations12 Mar 2024 Ziqi Yin, Shanshan Feng, Shang Liu, Gao Cong, Yew Soon Ong, Bin Cui

With the proliferation of spatio-textual data, Top-k KNN spatial keyword queries (TkQs), which return a list of objects based on a ranking function that considers both spatial and textual relevance, have found many real-life applications.

Pseudo Label

Retrieval-Augmented Generation for AI-Generated Content: A Survey

3 code implementations29 Feb 2024 Penghao Zhao, Hailin Zhang, Qinhan Yu, Zhengren Wang, Yunteng Geng, Fangcheng Fu, Ling Yang, Wentao Zhang, Jie Jiang, Bin Cui

We first classify RAG foundations according to how the retriever augments the generator, distilling the fundamental abstractions of the augmentation methodologies for various retrievers and generators.

Information Retrieval multimodal generation +3

Structure-Guided Adversarial Training of Diffusion Models

no code implementations CVPR 2024 Ling Yang, Haotian Qian, Zhilong Zhang, Jingwei Liu, Bin Cui

In this pioneering approach, we compel the model to learn manifold structures between samples in each training batch.

Conditional Image Generation Denoising

Contextualized Diffusion Models for Text-Guided Image and Video Generation

1 code implementation26 Feb 2024 Ling Yang, Zhilong Zhang, Zhaochen Yu, Jingwei Liu, Minkai Xu, Stefano Ermon, Bin Cui

To address this issue, we propose a novel and general contextualized diffusion model (ContextDiff) by incorporating the cross-modal context encompassing interactions and alignments between text condition and visual sample into forward and reverse processes.

Text-to-Image Generation Text-to-Video Editing +2

RealCompo: Balancing Realism and Compositionality Improves Text-to-Image Diffusion Models

2 code implementations20 Feb 2024 Xinchen Zhang, Ling Yang, Yaqi Cai, Zhaochen Yu, Kai-Ni Wang, Jiake Xie, Ye Tian, Minkai Xu, Yong Tang, Yujiu Yang, Bin Cui

In this paper, we propose RealCompo, a new training-free and transferred-friendly text-to-image generation framework, which aims to leverage the respective advantages of text-to-image models and spatial-aware image diffusion models (e. g., layout, keypoints and segmentation maps) to enhance both realism and compositionality of the generated images.

Denoising Text-to-Image Generation

Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs

1 code implementation22 Jan 2024 Ling Yang, Zhaochen Yu, Chenlin Meng, Minkai Xu, Stefano Ermon, Bin Cui

In this paper, we propose a brand new training-free text-to-image generation/editing framework, namely Recaption, Plan and Generate (RPG), harnessing the powerful chain-of-thought reasoning ability of multimodal LLMs to enhance the compositionality of text-to-image diffusion models.

Diffusion Personalization Tuning Free Large Language Model

On-Device Recommender Systems: A Tutorial on The New-Generation Recommendation Paradigm

no code implementations18 Dec 2023 Hongzhi Yin, Tong Chen, Liang Qu, Bin Cui

Given the sheer volume of contemporary e-commerce applications, recommender systems (RSs) have gained significant attention in both academia and industry.

Recommendation Systems

CAFE: Towards Compact, Adaptive, and Fast Embedding for Large-scale Recommendation Models

1 code implementation6 Dec 2023 Hailin Zhang, Zirui Liu, Boxuan Chen, Yikai Zhao, Tong Zhao, Tong Yang, Bin Cui

Guided by our design philosophy, we further propose a multi-level hash embedding framework to optimize the embedding tables of non-hot features.

Feature Importance Philosophy

Unveiling Vulnerabilities of Contrastive Recommender Systems to Poisoning Attacks

1 code implementation30 Nov 2023 Zongwei Wang, Junliang Yu, Min Gao, Hongzhi Yin, Bin Cui, Shazia Sadiq

Our analysis indicates that this vulnerability is attributed to the uniform spread of representations caused by the InfoNCE loss.

Contrastive Learning Recommendation Systems

Experimental Analysis of Large-scale Learnable Vector Storage Compression

1 code implementation27 Nov 2023 Hailin Zhang, Penghao Zhao, Xupeng Miao, Yingxia Shao, Zirui Liu, Tong Yang, Bin Cui

Learnable embedding vector is one of the most important applications in machine learning, and is widely used in various database-related domains.

Benchmarking

SpotServe: Serving Generative Large Language Models on Preemptible Instances

1 code implementation27 Nov 2023 Xupeng Miao, Chunan Shi, Jiangfei Duan, Xiaoli Xi, Dahua Lin, Bin Cui, Zhihao Jia

This paper aims to reduce the monetary cost for serving LLMs by leveraging preemptible GPU instances on modern clouds, which offer accesses to spare GPUs at a much cheaper price than regular instances but may be preempted by the cloud at any time.

Graph Matching

Accelerating Scalable Graph Neural Network Inference with Node-Adaptive Propagation

no code implementations17 Oct 2023 Xinyi Gao, Wentao Zhang, Junliang Yu, Yingxia Shao, Quoc Viet Hung Nguyen, Bin Cui, Hongzhi Yin

To further accelerate Scalable GNNs inference in this inductive setting, we propose an online propagation framework and two novel node-adaptive propagation methods that can customize the optimal propagation depth for each node based on its topological information and thereby avoid redundant feature propagation.

Graph Neural Network

Model-enhanced Vector Index

1 code implementation NeurIPS 2023 Hailin Zhang, Yujing Wang, Qi Chen, Ruiheng Chang, Ting Zhang, Ziming Miao, Yingyan Hou, Yang Ding, Xupeng Miao, Haonan Wang, Bochen Pang, Yuefeng Zhan, Hao Sun, Weiwei Deng, Qi Zhang, Fan Yang, Xing Xie, Mao Yang, Bin Cui

We empirically show that our model achieves better performance on the commonly used academic benchmarks MSMARCO Passage and Natural Questions, with comparable serving latency to dense retrieval solutions.

Natural Questions Quantization +1

Towards General and Efficient Online Tuning for Spark

no code implementations5 Sep 2023 Yang Li, Huaijun Jiang, Yu Shen, Yide Fang, Xiaofeng Yang, Danqing Huang, Xinyi Zhang, Wentao Zhang, Ce Zhang, Peng Chen, Bin Cui

The distributed data analytic system -- Spark is a common choice for processing massive volumes of heterogeneous data, while it is challenging to tune its parameters to achieve high performance.

Bayesian Optimization Meta-Learning

Mitigating Semantic Confusion from Hostile Neighborhood for Graph Active Learning

1 code implementation17 Aug 2023 Tianmeng Yang, Min Zhou, Yujing Wang, Zhengjie Lin, Lujia Pan, Bin Cui, Yunhai Tong

Graph Active Learning (GAL), which aims to find the most informative nodes in graphs for annotation to maximize the Graph Neural Networks (GNNs) performance, has attracted many research efforts but remains non-trivial challenges.

Active Learning Diversity +1

VQGraph: Rethinking Graph Representation Space for Bridging GNNs and MLPs

1 code implementation4 Aug 2023 Ling Yang, Ye Tian, Minkai Xu, Zhongyi Liu, Shenda Hong, Wei Qu, Wentao Zhang, Bin Cui, Muhan Zhang, Jure Leskovec

To address this issue, we propose to learn a new powerful graph representation space by directly labeling nodes' diverse local structures for GNN-to-MLP distillation.

Knowledge Distillation Quantization +1

BOURNE: Bootstrapped Self-supervised Learning Framework for Unified Graph Anomaly Detection

no code implementations28 Jul 2023 Jie Liu, Mengting He, Xuequn Shang, Jieming Shi, Bin Cui, Hongzhi Yin

By swapping the context embeddings between nodes and edges and measuring the agreement in the embedding space, we enable the mutual detection of node and edge anomalies.

CoLA Contrastive Learning +2

Improving Automatic Parallel Training via Balanced Memory Workload Optimization

1 code implementation5 Jul 2023 Yujie Wang, Youhe Jiang, Xupeng Miao, Fangcheng Fu, Shenhan Zhu, Xiaonan Nie, Yaofeng Tu, Bin Cui

Transformer models have emerged as the leading approach for achieving state-of-the-art performance across various application domains, serving as the foundation for advanced large-scale deep learning (DL) models.

Navigate

StreamE: Learning to Update Representations for Temporal Knowledge Graphs in Streaming Scenarios

2 code implementations journal 2023 Jiasheng Zhang, Jie Shao, Bin Cui

To reduce the parameter size, entity representations in StreamE are decoupled from the model training to serve as the memory module to store the historical information of entities.

Knowledge Graphs

Individual and Structural Graph Information Bottlenecks for Out-of-Distribution Generalization

1 code implementation28 Jun 2023 Ling Yang, Jiayi Zheng, Heyuan Wang, Zhongyi Liu, Zhilin Huang, Shenda Hong, Wentao Zhang, Bin Cui

To remove class spurious feature caused by distribution shifts, we propose Individual Graph Information Bottleneck (I-GIB) which discards irrelevant information by minimizing the mutual information between the input graph and its embeddings.

Graph Learning Out-of-Distribution Generalization

Accelerating Text-to-Image Editing via Cache-Enabled Sparse Diffusion Inference

2 code implementations27 May 2023 Zihao Yu, Haoyang Li, Fangcheng Fu, Xupeng Miao, Bin Cui

The key intuition behind our approach is to utilize the semantic mapping between the minor modifications on the input text and the affected regions on the output image.

Text-to-Image Generation

OpenBox: A Python Toolkit for Generalized Black-box Optimization

1 code implementation26 Apr 2023 Huaijun Jiang, Yu Shen, Yang Li, Beicheng Xu, Sixian Du, Wentao Zhang, Ce Zhang, Bin Cui

Black-box optimization (BBO) has a broad range of applications, including automatic machine learning, experimental design, and database knob tuning.

Experimental Design

FlexMoE: Scaling Large-scale Sparse Pre-trained Model Training via Dynamic Device Placement

no code implementations8 Apr 2023 Xiaonan Nie, Xupeng Miao, Zilong Wang, Zichao Yang, Jilong Xue, Lingxiao Ma, Gang Cao, Bin Cui

We first present an empirical analysis on the problems and opportunities of training MoE models, which motivates us to overcome the routing imbalance and fluctuation problems by a dynamic expert management and device placement mechanism.

Scheduling

A Unified and Efficient Coordinating Framework for Autonomous DBMS Tuning

no code implementations10 Mar 2023 Xinyi Zhang, Zhuo Chang, Hong Wu, Yang Li, Jia Chen, Jian Tan, Feifei Li, Bin Cui

To tune different components for DBMS, a coordinating mechanism is needed to make the multiple agents cognizant of each other.

Thompson Sampling

Angel-PTM: A Scalable and Economical Large-scale Pre-training System in Tencent

no code implementations6 Mar 2023 Xiaonan Nie, Yi Liu, Fangcheng Fu, Jinbao Xue, Dian Jiao, Xupeng Miao, Yangyu Tao, Bin Cui

Recent years have witnessed the unprecedented achievements of large-scale pre-trained models, especially the Transformer models.

Management Scheduling

Transfer Learning for Bayesian Optimization: A Survey

no code implementations12 Feb 2023 Tianyi Bai, Yang Li, Yu Shen, Xinyi Zhang, Wentao Zhang, Bin Cui

A wide spectrum of design and decision problems, including parameter tuning, A/B testing and drug design, intrinsically are instances of black-box optimization.

Bayesian Optimization Survey +1

Rover: An online Spark SQL tuning service via generalized transfer learning

no code implementations8 Feb 2023 Yu Shen, Xinyuyang Ren, Yupeng Lu, Huaijun Jiang, Huanyong Xu, Di Peng, Yang Li, Wentao Zhang, Bin Cui

When applying transfer learning to accelerate the tuning process, we notice two domain-specific challenges: 1) most previous work focus on transferring tuning history, while expert knowledge from Spark engineers is of great potential to improve the tuning performance but is not well studied so far; 2) history tasks should be carefully utilized, where using dissimilar ones lead to a deteriorated performance in production.

Bayesian Optimization Transfer Learning

DivBO: Diversity-aware CASH for Ensemble Learning

no code implementations7 Feb 2023 Yu Shen, Yupeng Lu, Yang Li, Yaofeng Tu, Wentao Zhang, Bin Cui

To tackle this issue and further enhance the ensemble performance, we propose DivBO, a diversity-aware framework to inject explicit search of diversity into the CASH problems.

AutoML Bayesian Optimization +2

Galvatron: Efficient Transformer Training over Multiple GPUs Using Automatic Parallelism

2 code implementations25 Nov 2022 Xupeng Miao, Yujie Wang, Youhe Jiang, Chunan Shi, Xiaonan Nie, Hailin Zhang, Bin Cui

Transformer models have achieved state-of-the-art performance on various domains of applications and gradually becomes the foundations of the advanced large deep learning (DL) models.

Diffusion-Based Scene Graph to Image Generation with Masked Contrastive Pre-Training

1 code implementation21 Nov 2022 Ling Yang, Zhilin Huang, Yang song, Shenda Hong, Guohao Li, Wentao Zhang, Bin Cui, Bernard Ghanem, Ming-Hsuan Yang

Generating images from graph-structured inputs, such as scene graphs, is uniquely challenging due to the difficulty of aligning nodes and connections in graphs with objects and their relations in images.

Image Generation

Efficient Graph Neural Network Inference at Large Scale

no code implementations1 Nov 2022 Xinyi Gao, Wentao Zhang, Yingxia Shao, Quoc Viet Hung Nguyen, Bin Cui, Hongzhi Yin

Graph neural networks (GNNs) have demonstrated excellent performance in a wide range of applications.

Graph Neural Network

Distributed Graph Neural Network Training: A Survey

no code implementations1 Nov 2022 Yingxia Shao, Hongzheng Li, Xizhi Gu, Hongbo Yin, Yawen Li, Xupeng Miao, Wentao Zhang, Bin Cui, Lei Chen

In recent years, many efforts have been made on distributed GNN training, and an array of training algorithms and systems have been proposed.

Distributed Computing Graph Neural Network +1

CALIP: Zero-Shot Enhancement of CLIP with Parameter-free Attention

1 code implementation28 Sep 2022 Ziyu Guo, Renrui Zhang, Longtian Qiu, Xianzheng Ma, Xupeng Miao, Xuming He, Bin Cui

Contrastive Language-Image Pre-training (CLIP) has been shown to learn visual representations with great transferability, which achieves promising accuracy for zero-shot classification.

Training-free 3D Point Cloud Classification Transfer Learning +1

Diffusion Models: A Comprehensive Survey of Methods and Applications

2 code implementations2 Sep 2022 Ling Yang, Zhilong Zhang, Yang song, Shenda Hong, Runsheng Xu, Yue Zhao, Yingxia Shao, Wentao Zhang, Bin Cui, Ming-Hsuan Yang

This survey aims to provide a contextualized, in-depth look at the state of diffusion models, identifying the key areas of focus and pointing to potential areas for further exploration.

Image Super-Resolution Survey +2

Towards Communication-efficient Vertical Federated Learning Training via Cache-enabled Local Updates

no code implementations29 Jul 2022 Fangcheng Fu, Xupeng Miao, Jiawei Jiang, Huanran Xue, Bin Cui

Vertical federated learning (VFL) is an emerging paradigm that allows different parties (e. g., organizations or enterprises) to collaboratively build machine learning models with privacy protection.

Vertical Federated Learning

Efficient End-to-End AutoML via Scalable Search Space Decomposition

1 code implementation19 Jun 2022 Yang Li, Yu Shen, Wentao Zhang, Ce Zhang, Bin Cui

End-to-end AutoML has attracted intensive interests from both academia and industry which automatically searches for ML pipelines in a space induced by feature engineering, algorithm/model selection, and hyper-parameter tuning.

AutoML Feature Engineering +1

NAFS: A Simple yet Tough-to-beat Baseline for Graph Representation Learning

2 code implementations17 Jun 2022 Wentao Zhang, Zeang Sheng, Mingyu Yang, Yang Li, Yu Shen, Zhi Yang, Bin Cui

First, GNNs can learn higher-order structural information by stacking more layers but can not deal with large depth due to the over-smoothing issue.

Graph Representation Learning Link Prediction +1

BlindFL: Vertical Federated Machine Learning without Peeking into Your Data

no code implementations16 Jun 2022 Fangcheng Fu, Huanran Xue, Yong Cheng, Yangyu Tao, Bin Cui

First, to address the functionality of VFL models, we propose the federated source layers to unite the data from different parties.

BIG-bench Machine Learning Vertical Federated Learning

Model Degradation Hinders Deep Graph Neural Networks

1 code implementation9 Jun 2022 Wentao Zhang, Zeang Sheng, Ziqi Yin, Yuezihan Jiang, Yikuan Xia, Jun Gao, Zhi Yang, Bin Cui

Graph Neural Networks (GNNs) have achieved great success in various graph mining tasks. However, drastic performance degradation is always observed when a GNN is stacked with many layers.

Attribute Graph Mining

TransBO: Hyperparameter Optimization via Two-Phase Transfer Learning

no code implementations6 Jun 2022 Yang Li, Yu Shen, Huaijun Jiang, Wentao Zhang, Zhi Yang, Ce Zhang, Bin Cui

With the extensive applications of machine learning models, automatic hyperparameter optimization (HPO) has become increasingly important.

Hyperparameter Optimization Neural Architecture Search +2

Transfer Learning based Search Space Design for Hyperparameter Tuning

no code implementations6 Jun 2022 Yang Li, Yu Shen, Huaijun Jiang, Tianyi Bai, Wentao Zhang, Ce Zhang, Bin Cui

The extensive experiments show that our approach considerably boosts BO by designing a promising and compact search space instead of using the entire space, and outperforms the state-of-the-arts on a wide range of benchmarks, including machine learning and deep learning tuning tasks, and neural architecture search.

Bayesian Optimization BIG-bench Machine Learning +2

Instance-wise Prompt Tuning for Pretrained Language Models

no code implementations4 Jun 2022 Yuezihan Jiang, Hao Yang, Junyang Lin, Hanyu Zhao, An Yang, Chang Zhou, Hongxia Yang, Zhi Yang, Bin Cui

Prompt Learning has recently gained great popularity in bridging the gap between pretraining tasks and various downstream tasks.

AutoML for Deep Recommender Systems: A Survey

no code implementations25 Mar 2022 Ruiqi Zheng, Liang Qu, Bin Cui, Yuhui Shi, Hongzhi Yin

To tackle this problem, Automated Machine Learning (AutoML) is introduced to automatically search for the proper candidates for different parts of deep recommender systems.

AutoML feature selection +2

ZOOMER: Boosting Retrieval on Web-scale Graphs by Regions of Interest

1 code implementation20 Mar 2022 Yuezihan Jiang, Yu Cheng, Hanyu Zhao, Wentao Zhang, Xupeng Miao, Yu He, Liang Wang, Zhi Yang, Bin Cui

We introduce ZOOMER, a system deployed at Taobao, the largest e-commerce platform in China, for training and serving GNN-based recommendations over web-scale graphs.

Retrieval

Information Gain Propagation: a new way to Graph Active Learning with Soft Labels

1 code implementation ICLR 2022 Wentao Zhang, Yexin Wang, Zhenbang You, Meng Cao, Ping Huang, Jiulong Shan, Zhi Yang, Bin Cui

Graph Neural Networks (GNNs) have achieved great success in various tasks, but their performance highly relies on a large number of labeled nodes, which typically requires considerable human effort.

Active Learning

PaSca: a Graph Neural Architecture Search System under the Scalable Paradigm

1 code implementation1 Mar 2022 Wentao Zhang, Yu Shen, Zheyu Lin, Yang Li, Xiaosen Li, Wen Ouyang, Yangyu Tao, Zhi Yang, Bin Cui

Through deconstructing the message passing mechanism, PasCa presents a novel Scalable Graph Neural Architecture Paradigm (SGAP), together with a general architecture design space consisting of 150k different designs.

Neural Architecture Search

Hyper-Tune: Towards Efficient Hyper-parameter Tuning at Scale

no code implementations18 Jan 2022 Yang Li, Yu Shen, Huaijun Jiang, Wentao Zhang, Jixiang Li, Ji Liu, Ce Zhang, Bin Cui

The ever-growing demand and complexity of machine learning are putting pressure on hyper-parameter tuning systems: while the evaluation cost of models continues to increase, the scalability of state-of-the-arts starts to become a crucial bottleneck.

Scheduling

K-Core Decomposition on Super Large Graphs with Limited Resources

no code implementations26 Dec 2021 Shicheng Gao, Jie Xu, Xiaosen Li, Fangcheng Fu, Wentao Zhang, Wen Ouyang, Yangyu Tao, Bin Cui

For example, the distributed K-core decomposition algorithm can scale to a large graph with 136 billion edges without losing correctness with our divide-and-conquer technique.

PointCLIP: Point Cloud Understanding by CLIP

2 code implementations CVPR 2022 Renrui Zhang, Ziyu Guo, Wei zhang, Kunchang Li, Xupeng Miao, Bin Cui, Yu Qiao, Peng Gao, Hongsheng Li

On top of that, we design an inter-view adapter to better extract the global feature and adaptively fuse the few-shot knowledge learned from 3D into CLIP pre-trained in 2D.

3D Open-Vocabulary Instance Segmentation Few-Shot Learning +7

RIM: Reliable Influence-based Active Learning on Graphs

1 code implementation NeurIPS 2021 Wentao Zhang, Yexin Wang, Zhenbang You, Meng Cao, Ping Huang, Jiulong Shan, Zhi Yang, Bin Cui

Message passing is the core of most graph models such as Graph Convolutional Network (GCN) and Label Propagation (LP), which usually require a large number of clean labeled data to smooth out the neighborhood over the graph.

Active Learning

Cardinality Estimation in DBMS: A Comprehensive Benchmark Evaluation

1 code implementation13 Sep 2021 Yuxing Han, Ziniu Wu, Peizhi Wu, Rong Zhu, Jingyi Yang, Liang Wei Tan, Kai Zeng, Gao Cong, Yanzhao Qin, Andreas Pfadler, Zhengping Qian, Jingren Zhou, Jiangneng Li, Bin Cui

Therefore, we propose a new metric P-Error to evaluate the performance of CardEst methods, which overcomes the limitation of Q-Error and is able to reflect the overall end-to-end performance of CardEst methods.

Graph Attention MLP with Reliable Label Utilization

no code implementations23 Aug 2021 Wentao Zhang, Ziqi Yin, Zeang Sheng, Wen Ouyang, Xiaosen Li, Yangyu Tao, Zhi Yang, Bin Cui

Graph neural networks (GNNs) have recently achieved state-of-the-art performance in many graph-based applications.

Graph Attention

Evaluating Deep Graph Neural Networks

1 code implementation2 Aug 2021 Wentao Zhang, Zeang Sheng, Yuezihan Jiang, Yikuan Xia, Jun Gao, Zhi Yang, Bin Cui

Based on the experimental results, we answer the following two essential questions: (1) what actually leads to the compromised performance of deep GNNs; (2) when we need and how to build deep GNNs.

Graph Mining Node Classification

Grain: Improving Data Efficiency of Graph Neural Networks via Diversified Influence Maximization

1 code implementation31 Jul 2021 Wentao Zhang, Zhi Yang, Yexin Wang, Yu Shen, Yang Li, Liang Wang, Bin Cui

Data selection methods, such as active learning and core-set selection, are useful tools for improving the data efficiency of deep learning models on large-scale datasets.

Active Learning Knowledge Graphs

ROD: Reception-aware Online Distillation for Sparse Graphs

1 code implementation25 Jul 2021 Wentao Zhang, Yuezihan Jiang, Yang Li, Zeang Sheng, Yu Shen, Xupeng Miao, Liang Wang, Zhi Yang, Bin Cui

Unfortunately, many real-world networks are sparse in terms of both edges and labels, leading to sub-optimal performance of GNNs.

Clustering Graph Learning +5

VolcanoML: Speeding up End-to-End AutoML via Scalable Search Space Decomposition

3 code implementations19 Jul 2021 Yang Li, Yu Shen, Wentao Zhang, Jiawei Jiang, Bolin Ding, Yaliang Li, Jingren Zhou, Zhi Yang, Wentao Wu, Ce Zhang, Bin Cui

End-to-end AutoML has attracted intensive interests from both academia and industry, which automatically searches for ML pipelines in a space induced by feature engineering, algorithm/model selection, and hyper-parameter tuning.

AutoML Feature Engineering +1

OpenBox: A Generalized Black-box Optimization Service

6 code implementations1 Jun 2021 Yang Li, Yu Shen, Wentao Zhang, Yuanwei Chen, Huaijun Jiang, Mingchao Liu, Jiawei Jiang, Jinyang Gao, Wentao Wu, Zhi Yang, Ce Zhang, Bin Cui

Black-box optimization (BBO) has a broad range of applications, including automatic machine learning, engineering, physics, and experimental design.

Experimental Design Transfer Learning

CausCF: Causal Collaborative Filtering for RecommendationEffect Estimation

no code implementations28 May 2021 Xu Xie, Zhaoyang Liu, Shiwen Wu, Fei Sun, Cihang Liu, Jiawei Chen, Jinyang Gao, Bin Cui, Bolin Ding

It is based on the idea that similar users not only have a similar taste on items, but also have similar treatment effect under recommendations.

Collaborative Filtering Recommendation Systems

GMLP: Building Scalable and Flexible Graph Neural Networks with Feature-Message Passing

no code implementations20 Apr 2021 Wentao Zhang, Yu Shen, Zheyu Lin, Yang Li, Xiaosen Li, Wen Ouyang, Yangyu Tao, Zhi Yang, Bin Cui

In recent studies, neural message passing has proved to be an effective way to design graph neural networks (GNNs), which have achieved state-of-the-art performance in many graph-based tasks.

Explore User Neighborhood for Real-time E-commerce Recommendation

no code implementations28 Feb 2021 Xu Xie, Fei Sun, Xiaoyong Yang, Zhao Yang, Jinyang Gao, Wenwu Ou, Bin Cui

On the one hand, it utilizes UI relations and user neighborhood to capture both global and local information.

Collaborative Filtering Recommendation Systems

Efficient Automatic CASH via Rising Bandits

no code implementations8 Dec 2020 Yang Li, Jiawei Jiang, Jinyang Gao, Yingxia Shao, Ce Zhang, Bin Cui

In this framework, the BO methods are used to solve the HPO problem for each ML algorithm separately, incorporating a much smaller hyperparameter space for BO methods.

Bayesian Optimization BIG-bench Machine Learning +2

Efficient and Scalable Structure Learning for Bayesian Networks: Algorithms and Applications

no code implementations7 Dec 2020 Rong Zhu, Andreas Pfadler, Ziniu Wu, Yuxing Han, Xiaoke Yang, Feng Ye, Zhenping Qian, Jingren Zhou, Bin Cui

To resolve this, we propose a new structure learning algorithm LEAST, which comprehensively fulfills our business requirements as it attains high accuracy, efficiency and scalability at the same time.

Anomaly Detection Explainable Recommendation

MFES-HB: Efficient Hyperband with Multi-Fidelity Quality Measurements

5 code implementations5 Dec 2020 Yang Li, Yu Shen, Jiawei Jiang, Jinyang Gao, Ce Zhang, Bin Cui

Instead of sampling configurations randomly in HB, BOHB samples configurations based on a BO surrogate model, which is constructed with the high-fidelity measurements only.

Bayesian Optimization Hyperparameter Optimization

FLAT: Fast, Lightweight and Accurate Method for Cardinality Estimation

1 code implementation18 Nov 2020 Rong Zhu, Ziniu Wu, Yuxing Han, Kai Zeng, Andreas Pfadler, Zhengping Qian, Jingren Zhou, Bin Cui

Despite decades of research, existing methods either over simplify the models only using independent factorization which leads to inaccurate estimates, or over complicate them by lossless conditional factorization without any independent assumption which results in slow probability computation.

Graph Neural Networks in Recommender Systems: A Survey

1 code implementation4 Nov 2020 Shiwen Wu, Fei Sun, Wentao Zhang, Xu Xie, Bin Cui

With the explosive growth of online information, recommender systems play a key role to alleviate such information overload.

Graph Neural Network Graph Representation Learning +1

Contrastive Learning for Sequential Recommendation

1 code implementation27 Oct 2020 Xu Xie, Fei Sun, Zhaoyang Liu, Shiwen Wu, Jinyang Gao, Bolin Ding, Bin Cui

Sequential recommendation methods play a crucial role in modern recommender systems because of their ability to capture a user's dynamic interest from her/his historical interactions.

Contrastive Learning Data Augmentation +1

UniNet: Scalable Network Representation Learning with Metropolis-Hastings Sampling

1 code implementation10 Oct 2020 Xingyu Yao, Yingxia Shao, Bin Cui, Lei Chen

Finally, with the new edge sampler and random walk model abstraction, we carefully implement a scalable NRL framework called UniNet.

Representation Learning

DeGNN: Characterizing and Improving Graph Neural Networks with Graph Decomposition

no code implementations10 Oct 2019 Xupeng Miao, Nezihe Merve Gürel, Wentao Zhang, Zhichao Han, Bo Li, Wei Min, Xi Rao, Hansheng Ren, Yinan Shan, Yingxia Shao, Yujie Wang, Fan Wu, Hui Xue, Yaming Yang, Zitao Zhang, Yang Zhao, Shuai Zhang, Yujing Wang, Bin Cui, Ce Zhang

Despite the wide application of Graph Convolutional Network (GCN), one major limitation is that it does not benefit from the increasing depth and suffers from the oversmoothing problem.

Graph Neural Network

X-Forest: Approximate Random Projection Trees for Similarity Measurement

1 code implementation25 Sep 2019 Yikai Zhao, Peiqing Chen, Zidong Zhao, Tong Yang, Jie Jiang, Bin Cui, Gong Zhang, Steve Uhlig

First, we introduced RP Trees into the tasks of similarity measurement such that accuracy is improved.

AHash: A Load-Balanced One Permutation Hash

1 code implementation25 Sep 2019 Chenxingyu Zhao, Jie Gui, Yixiao Guo, Jie Jiang, Tong Yang, Bin Cui, Gong Zhang

Unlike the densification to fill the empty bins after they undesirably occur, our design goal is to balance the load so as to reduce the empty bins in advance.

An Experimental Evaluation of Large Scale GBDT Systems

no code implementations3 Jul 2019 Fangcheng Fu, Jiawei Jiang, Yingxia Shao, Bin Cui

Gradient boosting decision tree (GBDT) is a widely-used machine learning algorithm in both data analytic competitions and real-world industrial applications.

Management

Cannot find the paper you are looking for? You can Submit a new open access paper.